1/*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a copy
4 * of this software and associated documentation files (the "Software"), to deal
5 * in the Software without restriction, including without limitation the rights
6 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7 * copies of the Software, and to permit persons to whom the Software is
8 * furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19 * THE SOFTWARE.
20 *
21 *===-----------------------------------------------------------------------===
22 */
23
24#ifndef __ARM_NEON_H
25#define __ARM_NEON_H
26
27#if !defined(__ARM_NEON)
28#error "NEON support not enabled"
29#endif
30
31#include <stdint.h>
32
33typedef float float32_t;
34typedef __fp16 float16_t;
35#ifdef __aarch64__
36typedef double float64_t;
37#endif
38
39#ifdef __aarch64__
40typedef uint8_t poly8_t;
41typedef uint16_t poly16_t;
42typedef uint64_t poly64_t;
43typedef __uint128_t poly128_t;
44#else
45typedef int8_t poly8_t;
46typedef int16_t poly16_t;
47#endif
48typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
49typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
50typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
51typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
52typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
53typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
54typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
55typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
56typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
57typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
58typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
59typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
60typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
61typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
62typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
63typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
64typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
65typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
66typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
67typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
68#ifdef __aarch64__
69typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
70typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
71#endif
72typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
73typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
74typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
75typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
76#ifdef __aarch64__
77typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
78typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
79#endif
80
81typedef struct int8x8x2_t {
82  int8x8_t val[2];
83} int8x8x2_t;
84
85typedef struct int8x16x2_t {
86  int8x16_t val[2];
87} int8x16x2_t;
88
89typedef struct int16x4x2_t {
90  int16x4_t val[2];
91} int16x4x2_t;
92
93typedef struct int16x8x2_t {
94  int16x8_t val[2];
95} int16x8x2_t;
96
97typedef struct int32x2x2_t {
98  int32x2_t val[2];
99} int32x2x2_t;
100
101typedef struct int32x4x2_t {
102  int32x4_t val[2];
103} int32x4x2_t;
104
105typedef struct int64x1x2_t {
106  int64x1_t val[2];
107} int64x1x2_t;
108
109typedef struct int64x2x2_t {
110  int64x2_t val[2];
111} int64x2x2_t;
112
113typedef struct uint8x8x2_t {
114  uint8x8_t val[2];
115} uint8x8x2_t;
116
117typedef struct uint8x16x2_t {
118  uint8x16_t val[2];
119} uint8x16x2_t;
120
121typedef struct uint16x4x2_t {
122  uint16x4_t val[2];
123} uint16x4x2_t;
124
125typedef struct uint16x8x2_t {
126  uint16x8_t val[2];
127} uint16x8x2_t;
128
129typedef struct uint32x2x2_t {
130  uint32x2_t val[2];
131} uint32x2x2_t;
132
133typedef struct uint32x4x2_t {
134  uint32x4_t val[2];
135} uint32x4x2_t;
136
137typedef struct uint64x1x2_t {
138  uint64x1_t val[2];
139} uint64x1x2_t;
140
141typedef struct uint64x2x2_t {
142  uint64x2_t val[2];
143} uint64x2x2_t;
144
145typedef struct float16x4x2_t {
146  float16x4_t val[2];
147} float16x4x2_t;
148
149typedef struct float16x8x2_t {
150  float16x8_t val[2];
151} float16x8x2_t;
152
153typedef struct float32x2x2_t {
154  float32x2_t val[2];
155} float32x2x2_t;
156
157typedef struct float32x4x2_t {
158  float32x4_t val[2];
159} float32x4x2_t;
160
161#ifdef __aarch64__
162typedef struct float64x1x2_t {
163  float64x1_t val[2];
164} float64x1x2_t;
165
166typedef struct float64x2x2_t {
167  float64x2_t val[2];
168} float64x2x2_t;
169
170#endif
171typedef struct poly8x8x2_t {
172  poly8x8_t val[2];
173} poly8x8x2_t;
174
175typedef struct poly8x16x2_t {
176  poly8x16_t val[2];
177} poly8x16x2_t;
178
179typedef struct poly16x4x2_t {
180  poly16x4_t val[2];
181} poly16x4x2_t;
182
183typedef struct poly16x8x2_t {
184  poly16x8_t val[2];
185} poly16x8x2_t;
186
187#ifdef __aarch64__
188typedef struct poly64x1x2_t {
189  poly64x1_t val[2];
190} poly64x1x2_t;
191
192typedef struct poly64x2x2_t {
193  poly64x2_t val[2];
194} poly64x2x2_t;
195
196#endif
197typedef struct int8x8x3_t {
198  int8x8_t val[3];
199} int8x8x3_t;
200
201typedef struct int8x16x3_t {
202  int8x16_t val[3];
203} int8x16x3_t;
204
205typedef struct int16x4x3_t {
206  int16x4_t val[3];
207} int16x4x3_t;
208
209typedef struct int16x8x3_t {
210  int16x8_t val[3];
211} int16x8x3_t;
212
213typedef struct int32x2x3_t {
214  int32x2_t val[3];
215} int32x2x3_t;
216
217typedef struct int32x4x3_t {
218  int32x4_t val[3];
219} int32x4x3_t;
220
221typedef struct int64x1x3_t {
222  int64x1_t val[3];
223} int64x1x3_t;
224
225typedef struct int64x2x3_t {
226  int64x2_t val[3];
227} int64x2x3_t;
228
229typedef struct uint8x8x3_t {
230  uint8x8_t val[3];
231} uint8x8x3_t;
232
233typedef struct uint8x16x3_t {
234  uint8x16_t val[3];
235} uint8x16x3_t;
236
237typedef struct uint16x4x3_t {
238  uint16x4_t val[3];
239} uint16x4x3_t;
240
241typedef struct uint16x8x3_t {
242  uint16x8_t val[3];
243} uint16x8x3_t;
244
245typedef struct uint32x2x3_t {
246  uint32x2_t val[3];
247} uint32x2x3_t;
248
249typedef struct uint32x4x3_t {
250  uint32x4_t val[3];
251} uint32x4x3_t;
252
253typedef struct uint64x1x3_t {
254  uint64x1_t val[3];
255} uint64x1x3_t;
256
257typedef struct uint64x2x3_t {
258  uint64x2_t val[3];
259} uint64x2x3_t;
260
261typedef struct float16x4x3_t {
262  float16x4_t val[3];
263} float16x4x3_t;
264
265typedef struct float16x8x3_t {
266  float16x8_t val[3];
267} float16x8x3_t;
268
269typedef struct float32x2x3_t {
270  float32x2_t val[3];
271} float32x2x3_t;
272
273typedef struct float32x4x3_t {
274  float32x4_t val[3];
275} float32x4x3_t;
276
277#ifdef __aarch64__
278typedef struct float64x1x3_t {
279  float64x1_t val[3];
280} float64x1x3_t;
281
282typedef struct float64x2x3_t {
283  float64x2_t val[3];
284} float64x2x3_t;
285
286#endif
287typedef struct poly8x8x3_t {
288  poly8x8_t val[3];
289} poly8x8x3_t;
290
291typedef struct poly8x16x3_t {
292  poly8x16_t val[3];
293} poly8x16x3_t;
294
295typedef struct poly16x4x3_t {
296  poly16x4_t val[3];
297} poly16x4x3_t;
298
299typedef struct poly16x8x3_t {
300  poly16x8_t val[3];
301} poly16x8x3_t;
302
303#ifdef __aarch64__
304typedef struct poly64x1x3_t {
305  poly64x1_t val[3];
306} poly64x1x3_t;
307
308typedef struct poly64x2x3_t {
309  poly64x2_t val[3];
310} poly64x2x3_t;
311
312#endif
313typedef struct int8x8x4_t {
314  int8x8_t val[4];
315} int8x8x4_t;
316
317typedef struct int8x16x4_t {
318  int8x16_t val[4];
319} int8x16x4_t;
320
321typedef struct int16x4x4_t {
322  int16x4_t val[4];
323} int16x4x4_t;
324
325typedef struct int16x8x4_t {
326  int16x8_t val[4];
327} int16x8x4_t;
328
329typedef struct int32x2x4_t {
330  int32x2_t val[4];
331} int32x2x4_t;
332
333typedef struct int32x4x4_t {
334  int32x4_t val[4];
335} int32x4x4_t;
336
337typedef struct int64x1x4_t {
338  int64x1_t val[4];
339} int64x1x4_t;
340
341typedef struct int64x2x4_t {
342  int64x2_t val[4];
343} int64x2x4_t;
344
345typedef struct uint8x8x4_t {
346  uint8x8_t val[4];
347} uint8x8x4_t;
348
349typedef struct uint8x16x4_t {
350  uint8x16_t val[4];
351} uint8x16x4_t;
352
353typedef struct uint16x4x4_t {
354  uint16x4_t val[4];
355} uint16x4x4_t;
356
357typedef struct uint16x8x4_t {
358  uint16x8_t val[4];
359} uint16x8x4_t;
360
361typedef struct uint32x2x4_t {
362  uint32x2_t val[4];
363} uint32x2x4_t;
364
365typedef struct uint32x4x4_t {
366  uint32x4_t val[4];
367} uint32x4x4_t;
368
369typedef struct uint64x1x4_t {
370  uint64x1_t val[4];
371} uint64x1x4_t;
372
373typedef struct uint64x2x4_t {
374  uint64x2_t val[4];
375} uint64x2x4_t;
376
377typedef struct float16x4x4_t {
378  float16x4_t val[4];
379} float16x4x4_t;
380
381typedef struct float16x8x4_t {
382  float16x8_t val[4];
383} float16x8x4_t;
384
385typedef struct float32x2x4_t {
386  float32x2_t val[4];
387} float32x2x4_t;
388
389typedef struct float32x4x4_t {
390  float32x4_t val[4];
391} float32x4x4_t;
392
393#ifdef __aarch64__
394typedef struct float64x1x4_t {
395  float64x1_t val[4];
396} float64x1x4_t;
397
398typedef struct float64x2x4_t {
399  float64x2_t val[4];
400} float64x2x4_t;
401
402#endif
403typedef struct poly8x8x4_t {
404  poly8x8_t val[4];
405} poly8x8x4_t;
406
407typedef struct poly8x16x4_t {
408  poly8x16_t val[4];
409} poly8x16x4_t;
410
411typedef struct poly16x4x4_t {
412  poly16x4_t val[4];
413} poly16x4x4_t;
414
415typedef struct poly16x8x4_t {
416  poly16x8_t val[4];
417} poly16x8x4_t;
418
419#ifdef __aarch64__
420typedef struct poly64x1x4_t {
421  poly64x1_t val[4];
422} poly64x1x4_t;
423
424typedef struct poly64x2x4_t {
425  poly64x2_t val[4];
426} poly64x2x4_t;
427
428#endif
429
430#define __ai static inline __attribute__((__always_inline__, __nodebug__))
431
432#ifdef __LITTLE_ENDIAN__
433__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
434  uint8x16_t __ret;
435  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
436  return __ret;
437}
438#else
439__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
440  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
441  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
442  uint8x16_t __ret;
443  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
444  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
445  return __ret;
446}
447__ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
448  uint8x16_t __ret;
449  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
450  return __ret;
451}
452#endif
453
454#ifdef __LITTLE_ENDIAN__
455__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
456  uint32x4_t __ret;
457  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
458  return __ret;
459}
460#else
461__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
462  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
463  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
464  uint32x4_t __ret;
465  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
467  return __ret;
468}
469__ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
470  uint32x4_t __ret;
471  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
472  return __ret;
473}
474#endif
475
476#ifdef __LITTLE_ENDIAN__
477__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
478  uint16x8_t __ret;
479  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
480  return __ret;
481}
482#else
483__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
484  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
485  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
486  uint16x8_t __ret;
487  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
488  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
489  return __ret;
490}
491__ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
492  uint16x8_t __ret;
493  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
494  return __ret;
495}
496#endif
497
498#ifdef __LITTLE_ENDIAN__
499__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
500  int8x16_t __ret;
501  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
502  return __ret;
503}
504#else
505__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
506  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
507  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
508  int8x16_t __ret;
509  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
511  return __ret;
512}
513__ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
514  int8x16_t __ret;
515  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
516  return __ret;
517}
518#endif
519
520#ifdef __LITTLE_ENDIAN__
521__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
522  float32x4_t __ret;
523  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
524  return __ret;
525}
526#else
527__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
528  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
529  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
530  float32x4_t __ret;
531  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
533  return __ret;
534}
535#endif
536
537#ifdef __LITTLE_ENDIAN__
538__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
539  int32x4_t __ret;
540  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
541  return __ret;
542}
543#else
544__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
545  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
546  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
547  int32x4_t __ret;
548  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
549  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
550  return __ret;
551}
552__ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
553  int32x4_t __ret;
554  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
555  return __ret;
556}
557#endif
558
559#ifdef __LITTLE_ENDIAN__
560__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
561  int16x8_t __ret;
562  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
563  return __ret;
564}
565#else
566__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
568  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
569  int16x8_t __ret;
570  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
571  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
572  return __ret;
573}
574__ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
575  int16x8_t __ret;
576  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
577  return __ret;
578}
579#endif
580
581#ifdef __LITTLE_ENDIAN__
582__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
583  uint8x8_t __ret;
584  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
585  return __ret;
586}
587#else
588__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
589  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
590  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
591  uint8x8_t __ret;
592  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
593  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
594  return __ret;
595}
596__ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
597  uint8x8_t __ret;
598  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
599  return __ret;
600}
601#endif
602
603#ifdef __LITTLE_ENDIAN__
604__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
605  uint32x2_t __ret;
606  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
607  return __ret;
608}
609#else
610__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
611  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
612  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
613  uint32x2_t __ret;
614  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
616  return __ret;
617}
618__ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
619  uint32x2_t __ret;
620  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
621  return __ret;
622}
623#endif
624
625#ifdef __LITTLE_ENDIAN__
626__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
627  uint16x4_t __ret;
628  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
629  return __ret;
630}
631#else
632__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
633  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
634  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
635  uint16x4_t __ret;
636  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
638  return __ret;
639}
640__ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
641  uint16x4_t __ret;
642  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
643  return __ret;
644}
645#endif
646
647#ifdef __LITTLE_ENDIAN__
648__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
649  int8x8_t __ret;
650  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
651  return __ret;
652}
653#else
654__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
655  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
656  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
657  int8x8_t __ret;
658  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
659  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
660  return __ret;
661}
662__ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
663  int8x8_t __ret;
664  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
665  return __ret;
666}
667#endif
668
669#ifdef __LITTLE_ENDIAN__
670__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
671  float32x2_t __ret;
672  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
673  return __ret;
674}
675#else
676__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
677  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
678  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
679  float32x2_t __ret;
680  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
682  return __ret;
683}
684#endif
685
686#ifdef __LITTLE_ENDIAN__
687__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
688  int32x2_t __ret;
689  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
690  return __ret;
691}
692#else
693__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
694  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
695  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
696  int32x2_t __ret;
697  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
699  return __ret;
700}
701__ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
702  int32x2_t __ret;
703  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
704  return __ret;
705}
706#endif
707
708#ifdef __LITTLE_ENDIAN__
709__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
710  int16x4_t __ret;
711  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
712  return __ret;
713}
714#else
715__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
716  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
717  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
718  int16x4_t __ret;
719  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
721  return __ret;
722}
723__ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
724  int16x4_t __ret;
725  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
726  return __ret;
727}
728#endif
729
730#ifdef __LITTLE_ENDIAN__
731__ai int8x16_t vabsq_s8(int8x16_t __p0) {
732  int8x16_t __ret;
733  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
734  return __ret;
735}
736#else
737__ai int8x16_t vabsq_s8(int8x16_t __p0) {
738  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
739  int8x16_t __ret;
740  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
742  return __ret;
743}
744#endif
745
746#ifdef __LITTLE_ENDIAN__
747__ai float32x4_t vabsq_f32(float32x4_t __p0) {
748  float32x4_t __ret;
749  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
750  return __ret;
751}
752#else
753__ai float32x4_t vabsq_f32(float32x4_t __p0) {
754  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
755  float32x4_t __ret;
756  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
757  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
758  return __ret;
759}
760#endif
761
762#ifdef __LITTLE_ENDIAN__
763__ai int32x4_t vabsq_s32(int32x4_t __p0) {
764  int32x4_t __ret;
765  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
766  return __ret;
767}
768#else
769__ai int32x4_t vabsq_s32(int32x4_t __p0) {
770  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
771  int32x4_t __ret;
772  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
774  return __ret;
775}
776#endif
777
778#ifdef __LITTLE_ENDIAN__
779__ai int16x8_t vabsq_s16(int16x8_t __p0) {
780  int16x8_t __ret;
781  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
782  return __ret;
783}
784#else
785__ai int16x8_t vabsq_s16(int16x8_t __p0) {
786  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
787  int16x8_t __ret;
788  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
790  return __ret;
791}
792#endif
793
794#ifdef __LITTLE_ENDIAN__
795__ai int8x8_t vabs_s8(int8x8_t __p0) {
796  int8x8_t __ret;
797  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
798  return __ret;
799}
800#else
801__ai int8x8_t vabs_s8(int8x8_t __p0) {
802  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
803  int8x8_t __ret;
804  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
806  return __ret;
807}
808#endif
809
810#ifdef __LITTLE_ENDIAN__
811__ai float32x2_t vabs_f32(float32x2_t __p0) {
812  float32x2_t __ret;
813  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
814  return __ret;
815}
816#else
817__ai float32x2_t vabs_f32(float32x2_t __p0) {
818  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
819  float32x2_t __ret;
820  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
822  return __ret;
823}
824#endif
825
826#ifdef __LITTLE_ENDIAN__
827__ai int32x2_t vabs_s32(int32x2_t __p0) {
828  int32x2_t __ret;
829  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
830  return __ret;
831}
832#else
833__ai int32x2_t vabs_s32(int32x2_t __p0) {
834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
835  int32x2_t __ret;
836  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
838  return __ret;
839}
840#endif
841
842#ifdef __LITTLE_ENDIAN__
843__ai int16x4_t vabs_s16(int16x4_t __p0) {
844  int16x4_t __ret;
845  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
846  return __ret;
847}
848#else
849__ai int16x4_t vabs_s16(int16x4_t __p0) {
850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
851  int16x4_t __ret;
852  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
854  return __ret;
855}
856#endif
857
858#ifdef __LITTLE_ENDIAN__
859__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
860  uint8x16_t __ret;
861  __ret = __p0 + __p1;
862  return __ret;
863}
864#else
865__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
867  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
868  uint8x16_t __ret;
869  __ret = __rev0 + __rev1;
870  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
871  return __ret;
872}
873#endif
874
875#ifdef __LITTLE_ENDIAN__
876__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
877  uint32x4_t __ret;
878  __ret = __p0 + __p1;
879  return __ret;
880}
881#else
882__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
883  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
884  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
885  uint32x4_t __ret;
886  __ret = __rev0 + __rev1;
887  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
888  return __ret;
889}
890#endif
891
892#ifdef __LITTLE_ENDIAN__
893__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
894  uint64x2_t __ret;
895  __ret = __p0 + __p1;
896  return __ret;
897}
898#else
899__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
900  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
901  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
902  uint64x2_t __ret;
903  __ret = __rev0 + __rev1;
904  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
905  return __ret;
906}
907#endif
908
909#ifdef __LITTLE_ENDIAN__
910__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
911  uint16x8_t __ret;
912  __ret = __p0 + __p1;
913  return __ret;
914}
915#else
916__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
917  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
918  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
919  uint16x8_t __ret;
920  __ret = __rev0 + __rev1;
921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
922  return __ret;
923}
924#endif
925
926#ifdef __LITTLE_ENDIAN__
927__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
928  int8x16_t __ret;
929  __ret = __p0 + __p1;
930  return __ret;
931}
932#else
933__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
934  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
935  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
936  int8x16_t __ret;
937  __ret = __rev0 + __rev1;
938  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
939  return __ret;
940}
941#endif
942
943#ifdef __LITTLE_ENDIAN__
944__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
945  float32x4_t __ret;
946  __ret = __p0 + __p1;
947  return __ret;
948}
949#else
950__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
951  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
952  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
953  float32x4_t __ret;
954  __ret = __rev0 + __rev1;
955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
956  return __ret;
957}
958#endif
959
960#ifdef __LITTLE_ENDIAN__
961__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
962  int32x4_t __ret;
963  __ret = __p0 + __p1;
964  return __ret;
965}
966#else
967__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
968  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
969  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
970  int32x4_t __ret;
971  __ret = __rev0 + __rev1;
972  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
973  return __ret;
974}
975#endif
976
977#ifdef __LITTLE_ENDIAN__
978__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
979  int64x2_t __ret;
980  __ret = __p0 + __p1;
981  return __ret;
982}
983#else
984__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
985  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
986  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
987  int64x2_t __ret;
988  __ret = __rev0 + __rev1;
989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
990  return __ret;
991}
992#endif
993
994#ifdef __LITTLE_ENDIAN__
995__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
996  int16x8_t __ret;
997  __ret = __p0 + __p1;
998  return __ret;
999}
1000#else
1001__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1002  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1003  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1004  int16x8_t __ret;
1005  __ret = __rev0 + __rev1;
1006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1007  return __ret;
1008}
1009#endif
1010
1011#ifdef __LITTLE_ENDIAN__
1012__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1013  uint8x8_t __ret;
1014  __ret = __p0 + __p1;
1015  return __ret;
1016}
1017#else
1018__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1019  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1020  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1021  uint8x8_t __ret;
1022  __ret = __rev0 + __rev1;
1023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1024  return __ret;
1025}
1026#endif
1027
1028#ifdef __LITTLE_ENDIAN__
1029__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1030  uint32x2_t __ret;
1031  __ret = __p0 + __p1;
1032  return __ret;
1033}
1034#else
1035__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1036  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1037  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1038  uint32x2_t __ret;
1039  __ret = __rev0 + __rev1;
1040  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1041  return __ret;
1042}
1043#endif
1044
1045#ifdef __LITTLE_ENDIAN__
1046__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1047  uint64x1_t __ret;
1048  __ret = __p0 + __p1;
1049  return __ret;
1050}
1051#else
1052__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1053  uint64x1_t __ret;
1054  __ret = __p0 + __p1;
1055  return __ret;
1056}
1057#endif
1058
1059#ifdef __LITTLE_ENDIAN__
1060__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1061  uint16x4_t __ret;
1062  __ret = __p0 + __p1;
1063  return __ret;
1064}
1065#else
1066__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1067  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1068  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1069  uint16x4_t __ret;
1070  __ret = __rev0 + __rev1;
1071  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1072  return __ret;
1073}
1074#endif
1075
1076#ifdef __LITTLE_ENDIAN__
1077__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1078  int8x8_t __ret;
1079  __ret = __p0 + __p1;
1080  return __ret;
1081}
1082#else
1083__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1084  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1085  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1086  int8x8_t __ret;
1087  __ret = __rev0 + __rev1;
1088  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1089  return __ret;
1090}
1091#endif
1092
1093#ifdef __LITTLE_ENDIAN__
1094__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1095  float32x2_t __ret;
1096  __ret = __p0 + __p1;
1097  return __ret;
1098}
1099#else
1100__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1101  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1102  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1103  float32x2_t __ret;
1104  __ret = __rev0 + __rev1;
1105  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1106  return __ret;
1107}
1108#endif
1109
1110#ifdef __LITTLE_ENDIAN__
1111__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1112  int32x2_t __ret;
1113  __ret = __p0 + __p1;
1114  return __ret;
1115}
1116#else
1117__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1118  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1119  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1120  int32x2_t __ret;
1121  __ret = __rev0 + __rev1;
1122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1123  return __ret;
1124}
1125#endif
1126
1127#ifdef __LITTLE_ENDIAN__
1128__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1129  int64x1_t __ret;
1130  __ret = __p0 + __p1;
1131  return __ret;
1132}
1133#else
1134__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1135  int64x1_t __ret;
1136  __ret = __p0 + __p1;
1137  return __ret;
1138}
1139#endif
1140
1141#ifdef __LITTLE_ENDIAN__
1142__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1143  int16x4_t __ret;
1144  __ret = __p0 + __p1;
1145  return __ret;
1146}
1147#else
1148__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1149  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1150  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1151  int16x4_t __ret;
1152  __ret = __rev0 + __rev1;
1153  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1154  return __ret;
1155}
1156#endif
1157
1158#ifdef __LITTLE_ENDIAN__
1159__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1160  uint16x4_t __ret;
1161  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1162  return __ret;
1163}
1164#else
1165__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1166  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1167  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1168  uint16x4_t __ret;
1169  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
1170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1171  return __ret;
1172}
1173__ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1174  uint16x4_t __ret;
1175  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1176  return __ret;
1177}
1178#endif
1179
1180#ifdef __LITTLE_ENDIAN__
1181__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1182  uint32x2_t __ret;
1183  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1184  return __ret;
1185}
1186#else
1187__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1188  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1189  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1190  uint32x2_t __ret;
1191  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
1192  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1193  return __ret;
1194}
1195__ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1196  uint32x2_t __ret;
1197  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1198  return __ret;
1199}
1200#endif
1201
1202#ifdef __LITTLE_ENDIAN__
1203__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1204  uint8x8_t __ret;
1205  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1206  return __ret;
1207}
1208#else
1209__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1210  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1211  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1212  uint8x8_t __ret;
1213  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
1214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1215  return __ret;
1216}
1217__ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1218  uint8x8_t __ret;
1219  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1220  return __ret;
1221}
1222#endif
1223
1224#ifdef __LITTLE_ENDIAN__
1225__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1226  int16x4_t __ret;
1227  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1228  return __ret;
1229}
1230#else
1231__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1232  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1233  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1234  int16x4_t __ret;
1235  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
1236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1237  return __ret;
1238}
1239__ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1240  int16x4_t __ret;
1241  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1242  return __ret;
1243}
1244#endif
1245
1246#ifdef __LITTLE_ENDIAN__
1247__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1248  int32x2_t __ret;
1249  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1250  return __ret;
1251}
1252#else
1253__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1254  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1255  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1256  int32x2_t __ret;
1257  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
1258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1259  return __ret;
1260}
1261__ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1262  int32x2_t __ret;
1263  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1264  return __ret;
1265}
1266#endif
1267
1268#ifdef __LITTLE_ENDIAN__
1269__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1270  int8x8_t __ret;
1271  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1272  return __ret;
1273}
1274#else
1275__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1276  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1277  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1278  int8x8_t __ret;
1279  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
1280  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1281  return __ret;
1282}
1283__ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1284  int8x8_t __ret;
1285  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1286  return __ret;
1287}
1288#endif
1289
1290#ifdef __LITTLE_ENDIAN__
1291__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1292  uint8x16_t __ret;
1293  __ret = __p0 & __p1;
1294  return __ret;
1295}
1296#else
1297__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1298  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1299  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1300  uint8x16_t __ret;
1301  __ret = __rev0 & __rev1;
1302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1303  return __ret;
1304}
1305#endif
1306
1307#ifdef __LITTLE_ENDIAN__
1308__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1309  uint32x4_t __ret;
1310  __ret = __p0 & __p1;
1311  return __ret;
1312}
1313#else
1314__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1315  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1316  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1317  uint32x4_t __ret;
1318  __ret = __rev0 & __rev1;
1319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1320  return __ret;
1321}
1322#endif
1323
1324#ifdef __LITTLE_ENDIAN__
1325__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1326  uint64x2_t __ret;
1327  __ret = __p0 & __p1;
1328  return __ret;
1329}
1330#else
1331__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1332  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1333  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1334  uint64x2_t __ret;
1335  __ret = __rev0 & __rev1;
1336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1337  return __ret;
1338}
1339#endif
1340
1341#ifdef __LITTLE_ENDIAN__
1342__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1343  uint16x8_t __ret;
1344  __ret = __p0 & __p1;
1345  return __ret;
1346}
1347#else
1348__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1349  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1350  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1351  uint16x8_t __ret;
1352  __ret = __rev0 & __rev1;
1353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1354  return __ret;
1355}
1356#endif
1357
1358#ifdef __LITTLE_ENDIAN__
1359__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1360  int8x16_t __ret;
1361  __ret = __p0 & __p1;
1362  return __ret;
1363}
1364#else
1365__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1366  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1367  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1368  int8x16_t __ret;
1369  __ret = __rev0 & __rev1;
1370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1371  return __ret;
1372}
1373#endif
1374
1375#ifdef __LITTLE_ENDIAN__
1376__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1377  int32x4_t __ret;
1378  __ret = __p0 & __p1;
1379  return __ret;
1380}
1381#else
1382__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1383  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1384  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1385  int32x4_t __ret;
1386  __ret = __rev0 & __rev1;
1387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1388  return __ret;
1389}
1390#endif
1391
1392#ifdef __LITTLE_ENDIAN__
1393__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1394  int64x2_t __ret;
1395  __ret = __p0 & __p1;
1396  return __ret;
1397}
1398#else
1399__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1400  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1401  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1402  int64x2_t __ret;
1403  __ret = __rev0 & __rev1;
1404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1405  return __ret;
1406}
1407#endif
1408
1409#ifdef __LITTLE_ENDIAN__
1410__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1411  int16x8_t __ret;
1412  __ret = __p0 & __p1;
1413  return __ret;
1414}
1415#else
1416__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1417  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1418  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1419  int16x8_t __ret;
1420  __ret = __rev0 & __rev1;
1421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1422  return __ret;
1423}
1424#endif
1425
1426#ifdef __LITTLE_ENDIAN__
1427__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1428  uint8x8_t __ret;
1429  __ret = __p0 & __p1;
1430  return __ret;
1431}
1432#else
1433__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1434  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1435  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1436  uint8x8_t __ret;
1437  __ret = __rev0 & __rev1;
1438  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1439  return __ret;
1440}
1441#endif
1442
1443#ifdef __LITTLE_ENDIAN__
1444__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1445  uint32x2_t __ret;
1446  __ret = __p0 & __p1;
1447  return __ret;
1448}
1449#else
1450__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1451  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1452  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1453  uint32x2_t __ret;
1454  __ret = __rev0 & __rev1;
1455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1456  return __ret;
1457}
1458#endif
1459
1460#ifdef __LITTLE_ENDIAN__
1461__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1462  uint64x1_t __ret;
1463  __ret = __p0 & __p1;
1464  return __ret;
1465}
1466#else
1467__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1468  uint64x1_t __ret;
1469  __ret = __p0 & __p1;
1470  return __ret;
1471}
1472#endif
1473
1474#ifdef __LITTLE_ENDIAN__
1475__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1476  uint16x4_t __ret;
1477  __ret = __p0 & __p1;
1478  return __ret;
1479}
1480#else
1481__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1482  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1483  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1484  uint16x4_t __ret;
1485  __ret = __rev0 & __rev1;
1486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1487  return __ret;
1488}
1489#endif
1490
1491#ifdef __LITTLE_ENDIAN__
1492__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1493  int8x8_t __ret;
1494  __ret = __p0 & __p1;
1495  return __ret;
1496}
1497#else
1498__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1499  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1500  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1501  int8x8_t __ret;
1502  __ret = __rev0 & __rev1;
1503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1504  return __ret;
1505}
1506#endif
1507
1508#ifdef __LITTLE_ENDIAN__
1509__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1510  int32x2_t __ret;
1511  __ret = __p0 & __p1;
1512  return __ret;
1513}
1514#else
1515__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1516  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1517  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1518  int32x2_t __ret;
1519  __ret = __rev0 & __rev1;
1520  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1521  return __ret;
1522}
1523#endif
1524
1525#ifdef __LITTLE_ENDIAN__
1526__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1527  int64x1_t __ret;
1528  __ret = __p0 & __p1;
1529  return __ret;
1530}
1531#else
1532__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1533  int64x1_t __ret;
1534  __ret = __p0 & __p1;
1535  return __ret;
1536}
1537#endif
1538
1539#ifdef __LITTLE_ENDIAN__
1540__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1541  int16x4_t __ret;
1542  __ret = __p0 & __p1;
1543  return __ret;
1544}
1545#else
1546__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1547  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1548  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1549  int16x4_t __ret;
1550  __ret = __rev0 & __rev1;
1551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1552  return __ret;
1553}
1554#endif
1555
1556#ifdef __LITTLE_ENDIAN__
1557__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1558  uint8x16_t __ret;
1559  __ret = __p0 & ~__p1;
1560  return __ret;
1561}
1562#else
1563__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1564  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1565  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1566  uint8x16_t __ret;
1567  __ret = __rev0 & ~__rev1;
1568  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1569  return __ret;
1570}
1571#endif
1572
1573#ifdef __LITTLE_ENDIAN__
1574__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1575  uint32x4_t __ret;
1576  __ret = __p0 & ~__p1;
1577  return __ret;
1578}
1579#else
1580__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1581  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1582  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1583  uint32x4_t __ret;
1584  __ret = __rev0 & ~__rev1;
1585  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1586  return __ret;
1587}
1588#endif
1589
1590#ifdef __LITTLE_ENDIAN__
1591__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1592  uint64x2_t __ret;
1593  __ret = __p0 & ~__p1;
1594  return __ret;
1595}
1596#else
1597__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1598  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1599  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1600  uint64x2_t __ret;
1601  __ret = __rev0 & ~__rev1;
1602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1603  return __ret;
1604}
1605#endif
1606
1607#ifdef __LITTLE_ENDIAN__
1608__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1609  uint16x8_t __ret;
1610  __ret = __p0 & ~__p1;
1611  return __ret;
1612}
1613#else
1614__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1615  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1616  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1617  uint16x8_t __ret;
1618  __ret = __rev0 & ~__rev1;
1619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1620  return __ret;
1621}
1622#endif
1623
1624#ifdef __LITTLE_ENDIAN__
1625__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1626  int8x16_t __ret;
1627  __ret = __p0 & ~__p1;
1628  return __ret;
1629}
1630#else
1631__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1632  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1633  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1634  int8x16_t __ret;
1635  __ret = __rev0 & ~__rev1;
1636  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1637  return __ret;
1638}
1639#endif
1640
1641#ifdef __LITTLE_ENDIAN__
1642__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1643  int32x4_t __ret;
1644  __ret = __p0 & ~__p1;
1645  return __ret;
1646}
1647#else
1648__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1649  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1650  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1651  int32x4_t __ret;
1652  __ret = __rev0 & ~__rev1;
1653  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1654  return __ret;
1655}
1656#endif
1657
1658#ifdef __LITTLE_ENDIAN__
1659__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1660  int64x2_t __ret;
1661  __ret = __p0 & ~__p1;
1662  return __ret;
1663}
1664#else
1665__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1666  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1667  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1668  int64x2_t __ret;
1669  __ret = __rev0 & ~__rev1;
1670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1671  return __ret;
1672}
1673#endif
1674
1675#ifdef __LITTLE_ENDIAN__
1676__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1677  int16x8_t __ret;
1678  __ret = __p0 & ~__p1;
1679  return __ret;
1680}
1681#else
1682__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1683  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1684  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1685  int16x8_t __ret;
1686  __ret = __rev0 & ~__rev1;
1687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1688  return __ret;
1689}
1690#endif
1691
1692#ifdef __LITTLE_ENDIAN__
1693__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1694  uint8x8_t __ret;
1695  __ret = __p0 & ~__p1;
1696  return __ret;
1697}
1698#else
1699__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1700  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1701  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1702  uint8x8_t __ret;
1703  __ret = __rev0 & ~__rev1;
1704  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1705  return __ret;
1706}
1707#endif
1708
1709#ifdef __LITTLE_ENDIAN__
1710__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1711  uint32x2_t __ret;
1712  __ret = __p0 & ~__p1;
1713  return __ret;
1714}
1715#else
1716__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1717  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1718  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1719  uint32x2_t __ret;
1720  __ret = __rev0 & ~__rev1;
1721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1722  return __ret;
1723}
1724#endif
1725
1726#ifdef __LITTLE_ENDIAN__
1727__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1728  uint64x1_t __ret;
1729  __ret = __p0 & ~__p1;
1730  return __ret;
1731}
1732#else
1733__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1734  uint64x1_t __ret;
1735  __ret = __p0 & ~__p1;
1736  return __ret;
1737}
1738#endif
1739
1740#ifdef __LITTLE_ENDIAN__
1741__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1742  uint16x4_t __ret;
1743  __ret = __p0 & ~__p1;
1744  return __ret;
1745}
1746#else
1747__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1748  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1749  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1750  uint16x4_t __ret;
1751  __ret = __rev0 & ~__rev1;
1752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1753  return __ret;
1754}
1755#endif
1756
1757#ifdef __LITTLE_ENDIAN__
1758__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1759  int8x8_t __ret;
1760  __ret = __p0 & ~__p1;
1761  return __ret;
1762}
1763#else
1764__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1765  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1766  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1767  int8x8_t __ret;
1768  __ret = __rev0 & ~__rev1;
1769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1770  return __ret;
1771}
1772#endif
1773
1774#ifdef __LITTLE_ENDIAN__
1775__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1776  int32x2_t __ret;
1777  __ret = __p0 & ~__p1;
1778  return __ret;
1779}
1780#else
1781__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1782  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1783  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1784  int32x2_t __ret;
1785  __ret = __rev0 & ~__rev1;
1786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1787  return __ret;
1788}
1789#endif
1790
1791#ifdef __LITTLE_ENDIAN__
1792__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1793  int64x1_t __ret;
1794  __ret = __p0 & ~__p1;
1795  return __ret;
1796}
1797#else
1798__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1799  int64x1_t __ret;
1800  __ret = __p0 & ~__p1;
1801  return __ret;
1802}
1803#endif
1804
1805#ifdef __LITTLE_ENDIAN__
1806__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1807  int16x4_t __ret;
1808  __ret = __p0 & ~__p1;
1809  return __ret;
1810}
1811#else
1812__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1813  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1814  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1815  int16x4_t __ret;
1816  __ret = __rev0 & ~__rev1;
1817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1818  return __ret;
1819}
1820#endif
1821
1822#ifdef __LITTLE_ENDIAN__
1823__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1824  poly8x8_t __ret;
1825  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
1826  return __ret;
1827}
1828#else
1829__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1830  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1831  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1832  poly8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1833  poly8x8_t __ret;
1834  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
1835  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1836  return __ret;
1837}
1838#endif
1839
1840#ifdef __LITTLE_ENDIAN__
1841__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1842  poly16x4_t __ret;
1843  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
1844  return __ret;
1845}
1846#else
1847__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1848  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1849  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1850  poly16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1851  poly16x4_t __ret;
1852  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
1853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1854  return __ret;
1855}
1856#endif
1857
1858#ifdef __LITTLE_ENDIAN__
1859__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1860  poly8x16_t __ret;
1861  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
1862  return __ret;
1863}
1864#else
1865__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1867  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1868  poly8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1869  poly8x16_t __ret;
1870  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
1871  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1872  return __ret;
1873}
1874#endif
1875
1876#ifdef __LITTLE_ENDIAN__
1877__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1878  poly16x8_t __ret;
1879  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
1880  return __ret;
1881}
1882#else
1883__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1884  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1885  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1886  poly16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1887  poly16x8_t __ret;
1888  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
1889  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1890  return __ret;
1891}
1892#endif
1893
1894#ifdef __LITTLE_ENDIAN__
1895__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1896  uint8x16_t __ret;
1897  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
1898  return __ret;
1899}
1900#else
1901__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1902  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1903  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1904  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1905  uint8x16_t __ret;
1906  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
1907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1908  return __ret;
1909}
1910#endif
1911
1912#ifdef __LITTLE_ENDIAN__
1913__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1914  uint32x4_t __ret;
1915  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
1916  return __ret;
1917}
1918#else
1919__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1920  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1921  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1922  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1923  uint32x4_t __ret;
1924  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
1925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1926  return __ret;
1927}
1928#endif
1929
1930#ifdef __LITTLE_ENDIAN__
1931__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1932  uint64x2_t __ret;
1933  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
1934  return __ret;
1935}
1936#else
1937__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1938  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1939  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1940  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
1941  uint64x2_t __ret;
1942  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
1943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1944  return __ret;
1945}
1946#endif
1947
1948#ifdef __LITTLE_ENDIAN__
1949__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1950  uint16x8_t __ret;
1951  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
1952  return __ret;
1953}
1954#else
1955__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1956  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1957  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1958  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1959  uint16x8_t __ret;
1960  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
1961  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1962  return __ret;
1963}
1964#endif
1965
1966#ifdef __LITTLE_ENDIAN__
1967__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1968  int8x16_t __ret;
1969  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
1970  return __ret;
1971}
1972#else
1973__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1974  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1975  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1976  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1977  int8x16_t __ret;
1978  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
1979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1980  return __ret;
1981}
1982#endif
1983
1984#ifdef __LITTLE_ENDIAN__
1985__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1986  float32x4_t __ret;
1987  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
1988  return __ret;
1989}
1990#else
1991__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1992  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1993  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1994  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1995  float32x4_t __ret;
1996  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
1997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1998  return __ret;
1999}
2000#endif
2001
2002#ifdef __LITTLE_ENDIAN__
2003__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2004  int32x4_t __ret;
2005  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
2006  return __ret;
2007}
2008#else
2009__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2010  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2011  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2012  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2013  int32x4_t __ret;
2014  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
2015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2016  return __ret;
2017}
2018#endif
2019
2020#ifdef __LITTLE_ENDIAN__
2021__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2022  int64x2_t __ret;
2023  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
2024  return __ret;
2025}
2026#else
2027__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2028  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2029  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2030  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2031  int64x2_t __ret;
2032  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
2033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2034  return __ret;
2035}
2036#endif
2037
2038#ifdef __LITTLE_ENDIAN__
2039__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2040  int16x8_t __ret;
2041  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
2042  return __ret;
2043}
2044#else
2045__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2046  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2047  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2048  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2049  int16x8_t __ret;
2050  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
2051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2052  return __ret;
2053}
2054#endif
2055
2056#ifdef __LITTLE_ENDIAN__
2057__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2058  uint8x8_t __ret;
2059  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
2060  return __ret;
2061}
2062#else
2063__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2064  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2065  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2066  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2067  uint8x8_t __ret;
2068  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
2069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2070  return __ret;
2071}
2072#endif
2073
2074#ifdef __LITTLE_ENDIAN__
2075__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2076  uint32x2_t __ret;
2077  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
2078  return __ret;
2079}
2080#else
2081__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2082  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2083  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2084  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2085  uint32x2_t __ret;
2086  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
2087  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2088  return __ret;
2089}
2090#endif
2091
2092#ifdef __LITTLE_ENDIAN__
2093__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2094  uint64x1_t __ret;
2095  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2096  return __ret;
2097}
2098#else
2099__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2100  uint64x1_t __ret;
2101  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2102  return __ret;
2103}
2104#endif
2105
2106#ifdef __LITTLE_ENDIAN__
2107__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2108  uint16x4_t __ret;
2109  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
2110  return __ret;
2111}
2112#else
2113__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2114  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2115  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2116  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2117  uint16x4_t __ret;
2118  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
2119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2120  return __ret;
2121}
2122#endif
2123
2124#ifdef __LITTLE_ENDIAN__
2125__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2126  int8x8_t __ret;
2127  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
2128  return __ret;
2129}
2130#else
2131__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2132  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2133  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2134  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2135  int8x8_t __ret;
2136  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
2137  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2138  return __ret;
2139}
2140#endif
2141
2142#ifdef __LITTLE_ENDIAN__
2143__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2144  float32x2_t __ret;
2145  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
2146  return __ret;
2147}
2148#else
2149__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2150  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2151  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2152  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2153  float32x2_t __ret;
2154  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
2155  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2156  return __ret;
2157}
2158#endif
2159
2160#ifdef __LITTLE_ENDIAN__
2161__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2162  int32x2_t __ret;
2163  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
2164  return __ret;
2165}
2166#else
2167__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2168  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2169  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2170  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2171  int32x2_t __ret;
2172  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
2173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2174  return __ret;
2175}
2176#endif
2177
2178#ifdef __LITTLE_ENDIAN__
2179__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2180  int64x1_t __ret;
2181  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2182  return __ret;
2183}
2184#else
2185__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2186  int64x1_t __ret;
2187  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2188  return __ret;
2189}
2190#endif
2191
2192#ifdef __LITTLE_ENDIAN__
2193__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2194  int16x4_t __ret;
2195  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
2196  return __ret;
2197}
2198#else
2199__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2200  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2201  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2202  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2203  int16x4_t __ret;
2204  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
2205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2206  return __ret;
2207}
2208#endif
2209
2210#ifdef __LITTLE_ENDIAN__
2211__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2212  uint32x4_t __ret;
2213  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2214  return __ret;
2215}
2216#else
2217__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2218  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2219  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2220  uint32x4_t __ret;
2221  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2223  return __ret;
2224}
2225#endif
2226
2227#ifdef __LITTLE_ENDIAN__
2228__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2229  uint32x2_t __ret;
2230  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2231  return __ret;
2232}
2233#else
2234__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2235  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2236  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2237  uint32x2_t __ret;
2238  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2240  return __ret;
2241}
2242#endif
2243
2244#ifdef __LITTLE_ENDIAN__
2245__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2246  uint32x4_t __ret;
2247  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2248  return __ret;
2249}
2250#else
2251__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2252  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2253  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2254  uint32x4_t __ret;
2255  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2257  return __ret;
2258}
2259#endif
2260
2261#ifdef __LITTLE_ENDIAN__
2262__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2263  uint32x2_t __ret;
2264  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2265  return __ret;
2266}
2267#else
2268__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2269  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2270  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2271  uint32x2_t __ret;
2272  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2273  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2274  return __ret;
2275}
2276#endif
2277
2278#ifdef __LITTLE_ENDIAN__
2279__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2280  uint32x4_t __ret;
2281  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2282  return __ret;
2283}
2284#else
2285__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2286  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2287  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2288  uint32x4_t __ret;
2289  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2290  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2291  return __ret;
2292}
2293#endif
2294
2295#ifdef __LITTLE_ENDIAN__
2296__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2297  uint32x2_t __ret;
2298  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2299  return __ret;
2300}
2301#else
2302__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2303  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2304  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2305  uint32x2_t __ret;
2306  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2308  return __ret;
2309}
2310#endif
2311
2312#ifdef __LITTLE_ENDIAN__
2313__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2314  uint32x4_t __ret;
2315  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2316  return __ret;
2317}
2318#else
2319__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2320  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2321  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2322  uint32x4_t __ret;
2323  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2325  return __ret;
2326}
2327#endif
2328
2329#ifdef __LITTLE_ENDIAN__
2330__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2331  uint32x2_t __ret;
2332  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2333  return __ret;
2334}
2335#else
2336__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2337  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2338  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2339  uint32x2_t __ret;
2340  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2342  return __ret;
2343}
2344#endif
2345
2346#ifdef __LITTLE_ENDIAN__
2347__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2348  uint8x8_t __ret;
2349  __ret = (uint8x8_t)(__p0 == __p1);
2350  return __ret;
2351}
2352#else
2353__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2354  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2355  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2356  uint8x8_t __ret;
2357  __ret = (uint8x8_t)(__rev0 == __rev1);
2358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2359  return __ret;
2360}
2361#endif
2362
2363#ifdef __LITTLE_ENDIAN__
2364__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2365  uint8x16_t __ret;
2366  __ret = (uint8x16_t)(__p0 == __p1);
2367  return __ret;
2368}
2369#else
2370__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2371  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2372  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2373  uint8x16_t __ret;
2374  __ret = (uint8x16_t)(__rev0 == __rev1);
2375  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2376  return __ret;
2377}
2378#endif
2379
2380#ifdef __LITTLE_ENDIAN__
2381__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2382  uint8x16_t __ret;
2383  __ret = (uint8x16_t)(__p0 == __p1);
2384  return __ret;
2385}
2386#else
2387__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2388  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2389  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2390  uint8x16_t __ret;
2391  __ret = (uint8x16_t)(__rev0 == __rev1);
2392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2393  return __ret;
2394}
2395#endif
2396
2397#ifdef __LITTLE_ENDIAN__
2398__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2399  uint32x4_t __ret;
2400  __ret = (uint32x4_t)(__p0 == __p1);
2401  return __ret;
2402}
2403#else
2404__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2405  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2406  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2407  uint32x4_t __ret;
2408  __ret = (uint32x4_t)(__rev0 == __rev1);
2409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2410  return __ret;
2411}
2412#endif
2413
2414#ifdef __LITTLE_ENDIAN__
2415__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2416  uint16x8_t __ret;
2417  __ret = (uint16x8_t)(__p0 == __p1);
2418  return __ret;
2419}
2420#else
2421__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2422  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2423  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2424  uint16x8_t __ret;
2425  __ret = (uint16x8_t)(__rev0 == __rev1);
2426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2427  return __ret;
2428}
2429#endif
2430
2431#ifdef __LITTLE_ENDIAN__
2432__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2433  uint8x16_t __ret;
2434  __ret = (uint8x16_t)(__p0 == __p1);
2435  return __ret;
2436}
2437#else
2438__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2439  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2440  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2441  uint8x16_t __ret;
2442  __ret = (uint8x16_t)(__rev0 == __rev1);
2443  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2444  return __ret;
2445}
2446#endif
2447
2448#ifdef __LITTLE_ENDIAN__
2449__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2450  uint32x4_t __ret;
2451  __ret = (uint32x4_t)(__p0 == __p1);
2452  return __ret;
2453}
2454#else
2455__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2456  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2457  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2458  uint32x4_t __ret;
2459  __ret = (uint32x4_t)(__rev0 == __rev1);
2460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2461  return __ret;
2462}
2463#endif
2464
2465#ifdef __LITTLE_ENDIAN__
2466__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2467  uint32x4_t __ret;
2468  __ret = (uint32x4_t)(__p0 == __p1);
2469  return __ret;
2470}
2471#else
2472__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2473  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2474  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2475  uint32x4_t __ret;
2476  __ret = (uint32x4_t)(__rev0 == __rev1);
2477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2478  return __ret;
2479}
2480#endif
2481
2482#ifdef __LITTLE_ENDIAN__
2483__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2484  uint16x8_t __ret;
2485  __ret = (uint16x8_t)(__p0 == __p1);
2486  return __ret;
2487}
2488#else
2489__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2490  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2491  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2492  uint16x8_t __ret;
2493  __ret = (uint16x8_t)(__rev0 == __rev1);
2494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2495  return __ret;
2496}
2497#endif
2498
2499#ifdef __LITTLE_ENDIAN__
2500__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2501  uint8x8_t __ret;
2502  __ret = (uint8x8_t)(__p0 == __p1);
2503  return __ret;
2504}
2505#else
2506__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2507  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2508  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2509  uint8x8_t __ret;
2510  __ret = (uint8x8_t)(__rev0 == __rev1);
2511  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2512  return __ret;
2513}
2514#endif
2515
2516#ifdef __LITTLE_ENDIAN__
2517__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2518  uint32x2_t __ret;
2519  __ret = (uint32x2_t)(__p0 == __p1);
2520  return __ret;
2521}
2522#else
2523__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2524  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2525  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2526  uint32x2_t __ret;
2527  __ret = (uint32x2_t)(__rev0 == __rev1);
2528  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2529  return __ret;
2530}
2531#endif
2532
2533#ifdef __LITTLE_ENDIAN__
2534__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2535  uint16x4_t __ret;
2536  __ret = (uint16x4_t)(__p0 == __p1);
2537  return __ret;
2538}
2539#else
2540__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2541  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2542  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2543  uint16x4_t __ret;
2544  __ret = (uint16x4_t)(__rev0 == __rev1);
2545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2546  return __ret;
2547}
2548#endif
2549
2550#ifdef __LITTLE_ENDIAN__
2551__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2552  uint8x8_t __ret;
2553  __ret = (uint8x8_t)(__p0 == __p1);
2554  return __ret;
2555}
2556#else
2557__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2558  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2559  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2560  uint8x8_t __ret;
2561  __ret = (uint8x8_t)(__rev0 == __rev1);
2562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2563  return __ret;
2564}
2565#endif
2566
2567#ifdef __LITTLE_ENDIAN__
2568__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2569  uint32x2_t __ret;
2570  __ret = (uint32x2_t)(__p0 == __p1);
2571  return __ret;
2572}
2573#else
2574__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2575  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2576  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2577  uint32x2_t __ret;
2578  __ret = (uint32x2_t)(__rev0 == __rev1);
2579  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2580  return __ret;
2581}
2582#endif
2583
2584#ifdef __LITTLE_ENDIAN__
2585__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2586  uint32x2_t __ret;
2587  __ret = (uint32x2_t)(__p0 == __p1);
2588  return __ret;
2589}
2590#else
2591__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2592  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2593  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2594  uint32x2_t __ret;
2595  __ret = (uint32x2_t)(__rev0 == __rev1);
2596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2597  return __ret;
2598}
2599#endif
2600
2601#ifdef __LITTLE_ENDIAN__
2602__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2603  uint16x4_t __ret;
2604  __ret = (uint16x4_t)(__p0 == __p1);
2605  return __ret;
2606}
2607#else
2608__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2609  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2610  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2611  uint16x4_t __ret;
2612  __ret = (uint16x4_t)(__rev0 == __rev1);
2613  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2614  return __ret;
2615}
2616#endif
2617
2618#ifdef __LITTLE_ENDIAN__
2619__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2620  uint8x16_t __ret;
2621  __ret = (uint8x16_t)(__p0 >= __p1);
2622  return __ret;
2623}
2624#else
2625__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2626  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2627  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2628  uint8x16_t __ret;
2629  __ret = (uint8x16_t)(__rev0 >= __rev1);
2630  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2631  return __ret;
2632}
2633#endif
2634
2635#ifdef __LITTLE_ENDIAN__
2636__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2637  uint32x4_t __ret;
2638  __ret = (uint32x4_t)(__p0 >= __p1);
2639  return __ret;
2640}
2641#else
2642__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2643  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2644  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2645  uint32x4_t __ret;
2646  __ret = (uint32x4_t)(__rev0 >= __rev1);
2647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2648  return __ret;
2649}
2650#endif
2651
2652#ifdef __LITTLE_ENDIAN__
2653__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2654  uint16x8_t __ret;
2655  __ret = (uint16x8_t)(__p0 >= __p1);
2656  return __ret;
2657}
2658#else
2659__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2660  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2661  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2662  uint16x8_t __ret;
2663  __ret = (uint16x8_t)(__rev0 >= __rev1);
2664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2665  return __ret;
2666}
2667#endif
2668
2669#ifdef __LITTLE_ENDIAN__
2670__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2671  uint8x16_t __ret;
2672  __ret = (uint8x16_t)(__p0 >= __p1);
2673  return __ret;
2674}
2675#else
2676__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2677  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2678  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2679  uint8x16_t __ret;
2680  __ret = (uint8x16_t)(__rev0 >= __rev1);
2681  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2682  return __ret;
2683}
2684#endif
2685
2686#ifdef __LITTLE_ENDIAN__
2687__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2688  uint32x4_t __ret;
2689  __ret = (uint32x4_t)(__p0 >= __p1);
2690  return __ret;
2691}
2692#else
2693__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2694  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2695  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2696  uint32x4_t __ret;
2697  __ret = (uint32x4_t)(__rev0 >= __rev1);
2698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2699  return __ret;
2700}
2701#endif
2702
2703#ifdef __LITTLE_ENDIAN__
2704__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2705  uint32x4_t __ret;
2706  __ret = (uint32x4_t)(__p0 >= __p1);
2707  return __ret;
2708}
2709#else
2710__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2711  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2712  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2713  uint32x4_t __ret;
2714  __ret = (uint32x4_t)(__rev0 >= __rev1);
2715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2716  return __ret;
2717}
2718#endif
2719
2720#ifdef __LITTLE_ENDIAN__
2721__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2722  uint16x8_t __ret;
2723  __ret = (uint16x8_t)(__p0 >= __p1);
2724  return __ret;
2725}
2726#else
2727__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2728  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2729  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2730  uint16x8_t __ret;
2731  __ret = (uint16x8_t)(__rev0 >= __rev1);
2732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2733  return __ret;
2734}
2735#endif
2736
2737#ifdef __LITTLE_ENDIAN__
2738__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2739  uint8x8_t __ret;
2740  __ret = (uint8x8_t)(__p0 >= __p1);
2741  return __ret;
2742}
2743#else
2744__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2745  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2746  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2747  uint8x8_t __ret;
2748  __ret = (uint8x8_t)(__rev0 >= __rev1);
2749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2750  return __ret;
2751}
2752#endif
2753
2754#ifdef __LITTLE_ENDIAN__
2755__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2756  uint32x2_t __ret;
2757  __ret = (uint32x2_t)(__p0 >= __p1);
2758  return __ret;
2759}
2760#else
2761__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2762  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2763  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2764  uint32x2_t __ret;
2765  __ret = (uint32x2_t)(__rev0 >= __rev1);
2766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2767  return __ret;
2768}
2769#endif
2770
2771#ifdef __LITTLE_ENDIAN__
2772__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2773  uint16x4_t __ret;
2774  __ret = (uint16x4_t)(__p0 >= __p1);
2775  return __ret;
2776}
2777#else
2778__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2779  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2781  uint16x4_t __ret;
2782  __ret = (uint16x4_t)(__rev0 >= __rev1);
2783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2784  return __ret;
2785}
2786#endif
2787
2788#ifdef __LITTLE_ENDIAN__
2789__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2790  uint8x8_t __ret;
2791  __ret = (uint8x8_t)(__p0 >= __p1);
2792  return __ret;
2793}
2794#else
2795__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2796  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2797  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2798  uint8x8_t __ret;
2799  __ret = (uint8x8_t)(__rev0 >= __rev1);
2800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2801  return __ret;
2802}
2803#endif
2804
2805#ifdef __LITTLE_ENDIAN__
2806__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2807  uint32x2_t __ret;
2808  __ret = (uint32x2_t)(__p0 >= __p1);
2809  return __ret;
2810}
2811#else
2812__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2813  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2814  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2815  uint32x2_t __ret;
2816  __ret = (uint32x2_t)(__rev0 >= __rev1);
2817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2818  return __ret;
2819}
2820#endif
2821
2822#ifdef __LITTLE_ENDIAN__
2823__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2824  uint32x2_t __ret;
2825  __ret = (uint32x2_t)(__p0 >= __p1);
2826  return __ret;
2827}
2828#else
2829__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2830  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2831  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2832  uint32x2_t __ret;
2833  __ret = (uint32x2_t)(__rev0 >= __rev1);
2834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2835  return __ret;
2836}
2837#endif
2838
2839#ifdef __LITTLE_ENDIAN__
2840__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2841  uint16x4_t __ret;
2842  __ret = (uint16x4_t)(__p0 >= __p1);
2843  return __ret;
2844}
2845#else
2846__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2847  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2848  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2849  uint16x4_t __ret;
2850  __ret = (uint16x4_t)(__rev0 >= __rev1);
2851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2852  return __ret;
2853}
2854#endif
2855
2856#ifdef __LITTLE_ENDIAN__
2857__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2858  uint8x16_t __ret;
2859  __ret = (uint8x16_t)(__p0 > __p1);
2860  return __ret;
2861}
2862#else
2863__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2864  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2865  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2866  uint8x16_t __ret;
2867  __ret = (uint8x16_t)(__rev0 > __rev1);
2868  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2869  return __ret;
2870}
2871#endif
2872
2873#ifdef __LITTLE_ENDIAN__
2874__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2875  uint32x4_t __ret;
2876  __ret = (uint32x4_t)(__p0 > __p1);
2877  return __ret;
2878}
2879#else
2880__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2881  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2882  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2883  uint32x4_t __ret;
2884  __ret = (uint32x4_t)(__rev0 > __rev1);
2885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2886  return __ret;
2887}
2888#endif
2889
2890#ifdef __LITTLE_ENDIAN__
2891__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2892  uint16x8_t __ret;
2893  __ret = (uint16x8_t)(__p0 > __p1);
2894  return __ret;
2895}
2896#else
2897__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2898  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2899  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2900  uint16x8_t __ret;
2901  __ret = (uint16x8_t)(__rev0 > __rev1);
2902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2903  return __ret;
2904}
2905#endif
2906
2907#ifdef __LITTLE_ENDIAN__
2908__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2909  uint8x16_t __ret;
2910  __ret = (uint8x16_t)(__p0 > __p1);
2911  return __ret;
2912}
2913#else
2914__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2915  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2916  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2917  uint8x16_t __ret;
2918  __ret = (uint8x16_t)(__rev0 > __rev1);
2919  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2920  return __ret;
2921}
2922#endif
2923
2924#ifdef __LITTLE_ENDIAN__
2925__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2926  uint32x4_t __ret;
2927  __ret = (uint32x4_t)(__p0 > __p1);
2928  return __ret;
2929}
2930#else
2931__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2932  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2933  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2934  uint32x4_t __ret;
2935  __ret = (uint32x4_t)(__rev0 > __rev1);
2936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2937  return __ret;
2938}
2939#endif
2940
2941#ifdef __LITTLE_ENDIAN__
2942__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2943  uint32x4_t __ret;
2944  __ret = (uint32x4_t)(__p0 > __p1);
2945  return __ret;
2946}
2947#else
2948__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2949  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2950  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2951  uint32x4_t __ret;
2952  __ret = (uint32x4_t)(__rev0 > __rev1);
2953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2954  return __ret;
2955}
2956#endif
2957
2958#ifdef __LITTLE_ENDIAN__
2959__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2960  uint16x8_t __ret;
2961  __ret = (uint16x8_t)(__p0 > __p1);
2962  return __ret;
2963}
2964#else
2965__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2966  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2967  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2968  uint16x8_t __ret;
2969  __ret = (uint16x8_t)(__rev0 > __rev1);
2970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2971  return __ret;
2972}
2973#endif
2974
2975#ifdef __LITTLE_ENDIAN__
2976__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2977  uint8x8_t __ret;
2978  __ret = (uint8x8_t)(__p0 > __p1);
2979  return __ret;
2980}
2981#else
2982__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2983  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2984  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2985  uint8x8_t __ret;
2986  __ret = (uint8x8_t)(__rev0 > __rev1);
2987  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2988  return __ret;
2989}
2990#endif
2991
2992#ifdef __LITTLE_ENDIAN__
2993__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
2994  uint32x2_t __ret;
2995  __ret = (uint32x2_t)(__p0 > __p1);
2996  return __ret;
2997}
2998#else
2999__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3001  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3002  uint32x2_t __ret;
3003  __ret = (uint32x2_t)(__rev0 > __rev1);
3004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3005  return __ret;
3006}
3007#endif
3008
3009#ifdef __LITTLE_ENDIAN__
3010__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3011  uint16x4_t __ret;
3012  __ret = (uint16x4_t)(__p0 > __p1);
3013  return __ret;
3014}
3015#else
3016__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3017  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3018  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3019  uint16x4_t __ret;
3020  __ret = (uint16x4_t)(__rev0 > __rev1);
3021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3022  return __ret;
3023}
3024#endif
3025
3026#ifdef __LITTLE_ENDIAN__
3027__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3028  uint8x8_t __ret;
3029  __ret = (uint8x8_t)(__p0 > __p1);
3030  return __ret;
3031}
3032#else
3033__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3034  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3035  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3036  uint8x8_t __ret;
3037  __ret = (uint8x8_t)(__rev0 > __rev1);
3038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3039  return __ret;
3040}
3041#endif
3042
3043#ifdef __LITTLE_ENDIAN__
3044__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3045  uint32x2_t __ret;
3046  __ret = (uint32x2_t)(__p0 > __p1);
3047  return __ret;
3048}
3049#else
3050__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3051  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3052  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3053  uint32x2_t __ret;
3054  __ret = (uint32x2_t)(__rev0 > __rev1);
3055  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3056  return __ret;
3057}
3058#endif
3059
3060#ifdef __LITTLE_ENDIAN__
3061__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3062  uint32x2_t __ret;
3063  __ret = (uint32x2_t)(__p0 > __p1);
3064  return __ret;
3065}
3066#else
3067__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3068  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3069  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3070  uint32x2_t __ret;
3071  __ret = (uint32x2_t)(__rev0 > __rev1);
3072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3073  return __ret;
3074}
3075#endif
3076
3077#ifdef __LITTLE_ENDIAN__
3078__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3079  uint16x4_t __ret;
3080  __ret = (uint16x4_t)(__p0 > __p1);
3081  return __ret;
3082}
3083#else
3084__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3085  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3086  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3087  uint16x4_t __ret;
3088  __ret = (uint16x4_t)(__rev0 > __rev1);
3089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3090  return __ret;
3091}
3092#endif
3093
3094#ifdef __LITTLE_ENDIAN__
3095__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3096  uint8x16_t __ret;
3097  __ret = (uint8x16_t)(__p0 <= __p1);
3098  return __ret;
3099}
3100#else
3101__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3102  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3103  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3104  uint8x16_t __ret;
3105  __ret = (uint8x16_t)(__rev0 <= __rev1);
3106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3107  return __ret;
3108}
3109#endif
3110
3111#ifdef __LITTLE_ENDIAN__
3112__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3113  uint32x4_t __ret;
3114  __ret = (uint32x4_t)(__p0 <= __p1);
3115  return __ret;
3116}
3117#else
3118__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3119  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3120  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3121  uint32x4_t __ret;
3122  __ret = (uint32x4_t)(__rev0 <= __rev1);
3123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3124  return __ret;
3125}
3126#endif
3127
3128#ifdef __LITTLE_ENDIAN__
3129__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3130  uint16x8_t __ret;
3131  __ret = (uint16x8_t)(__p0 <= __p1);
3132  return __ret;
3133}
3134#else
3135__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3136  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3137  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3138  uint16x8_t __ret;
3139  __ret = (uint16x8_t)(__rev0 <= __rev1);
3140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3141  return __ret;
3142}
3143#endif
3144
3145#ifdef __LITTLE_ENDIAN__
3146__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3147  uint8x16_t __ret;
3148  __ret = (uint8x16_t)(__p0 <= __p1);
3149  return __ret;
3150}
3151#else
3152__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3153  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3154  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3155  uint8x16_t __ret;
3156  __ret = (uint8x16_t)(__rev0 <= __rev1);
3157  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3158  return __ret;
3159}
3160#endif
3161
3162#ifdef __LITTLE_ENDIAN__
3163__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3164  uint32x4_t __ret;
3165  __ret = (uint32x4_t)(__p0 <= __p1);
3166  return __ret;
3167}
3168#else
3169__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3170  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3171  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3172  uint32x4_t __ret;
3173  __ret = (uint32x4_t)(__rev0 <= __rev1);
3174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3175  return __ret;
3176}
3177#endif
3178
3179#ifdef __LITTLE_ENDIAN__
3180__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3181  uint32x4_t __ret;
3182  __ret = (uint32x4_t)(__p0 <= __p1);
3183  return __ret;
3184}
3185#else
3186__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3187  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3188  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3189  uint32x4_t __ret;
3190  __ret = (uint32x4_t)(__rev0 <= __rev1);
3191  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3192  return __ret;
3193}
3194#endif
3195
3196#ifdef __LITTLE_ENDIAN__
3197__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3198  uint16x8_t __ret;
3199  __ret = (uint16x8_t)(__p0 <= __p1);
3200  return __ret;
3201}
3202#else
3203__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3204  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3205  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3206  uint16x8_t __ret;
3207  __ret = (uint16x8_t)(__rev0 <= __rev1);
3208  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3209  return __ret;
3210}
3211#endif
3212
3213#ifdef __LITTLE_ENDIAN__
3214__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3215  uint8x8_t __ret;
3216  __ret = (uint8x8_t)(__p0 <= __p1);
3217  return __ret;
3218}
3219#else
3220__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3221  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3222  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3223  uint8x8_t __ret;
3224  __ret = (uint8x8_t)(__rev0 <= __rev1);
3225  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3226  return __ret;
3227}
3228#endif
3229
3230#ifdef __LITTLE_ENDIAN__
3231__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3232  uint32x2_t __ret;
3233  __ret = (uint32x2_t)(__p0 <= __p1);
3234  return __ret;
3235}
3236#else
3237__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3238  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3239  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3240  uint32x2_t __ret;
3241  __ret = (uint32x2_t)(__rev0 <= __rev1);
3242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3243  return __ret;
3244}
3245#endif
3246
3247#ifdef __LITTLE_ENDIAN__
3248__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3249  uint16x4_t __ret;
3250  __ret = (uint16x4_t)(__p0 <= __p1);
3251  return __ret;
3252}
3253#else
3254__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3255  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3256  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3257  uint16x4_t __ret;
3258  __ret = (uint16x4_t)(__rev0 <= __rev1);
3259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3260  return __ret;
3261}
3262#endif
3263
3264#ifdef __LITTLE_ENDIAN__
3265__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3266  uint8x8_t __ret;
3267  __ret = (uint8x8_t)(__p0 <= __p1);
3268  return __ret;
3269}
3270#else
3271__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3272  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3273  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3274  uint8x8_t __ret;
3275  __ret = (uint8x8_t)(__rev0 <= __rev1);
3276  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3277  return __ret;
3278}
3279#endif
3280
3281#ifdef __LITTLE_ENDIAN__
3282__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3283  uint32x2_t __ret;
3284  __ret = (uint32x2_t)(__p0 <= __p1);
3285  return __ret;
3286}
3287#else
3288__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3289  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3290  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3291  uint32x2_t __ret;
3292  __ret = (uint32x2_t)(__rev0 <= __rev1);
3293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3294  return __ret;
3295}
3296#endif
3297
3298#ifdef __LITTLE_ENDIAN__
3299__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3300  uint32x2_t __ret;
3301  __ret = (uint32x2_t)(__p0 <= __p1);
3302  return __ret;
3303}
3304#else
3305__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3306  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3307  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3308  uint32x2_t __ret;
3309  __ret = (uint32x2_t)(__rev0 <= __rev1);
3310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3311  return __ret;
3312}
3313#endif
3314
3315#ifdef __LITTLE_ENDIAN__
3316__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3317  uint16x4_t __ret;
3318  __ret = (uint16x4_t)(__p0 <= __p1);
3319  return __ret;
3320}
3321#else
3322__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3323  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3324  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3325  uint16x4_t __ret;
3326  __ret = (uint16x4_t)(__rev0 <= __rev1);
3327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3328  return __ret;
3329}
3330#endif
3331
3332#ifdef __LITTLE_ENDIAN__
3333__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3334  int8x16_t __ret;
3335  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
3336  return __ret;
3337}
3338#else
3339__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3340  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3341  int8x16_t __ret;
3342  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
3343  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3344  return __ret;
3345}
3346#endif
3347
3348#ifdef __LITTLE_ENDIAN__
3349__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3350  int32x4_t __ret;
3351  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
3352  return __ret;
3353}
3354#else
3355__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3356  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3357  int32x4_t __ret;
3358  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
3359  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3360  return __ret;
3361}
3362#endif
3363
3364#ifdef __LITTLE_ENDIAN__
3365__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3366  int16x8_t __ret;
3367  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
3368  return __ret;
3369}
3370#else
3371__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3372  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3373  int16x8_t __ret;
3374  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
3375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3376  return __ret;
3377}
3378#endif
3379
3380#ifdef __LITTLE_ENDIAN__
3381__ai int8x8_t vcls_s8(int8x8_t __p0) {
3382  int8x8_t __ret;
3383  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
3384  return __ret;
3385}
3386#else
3387__ai int8x8_t vcls_s8(int8x8_t __p0) {
3388  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3389  int8x8_t __ret;
3390  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
3391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3392  return __ret;
3393}
3394#endif
3395
3396#ifdef __LITTLE_ENDIAN__
3397__ai int32x2_t vcls_s32(int32x2_t __p0) {
3398  int32x2_t __ret;
3399  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
3400  return __ret;
3401}
3402#else
3403__ai int32x2_t vcls_s32(int32x2_t __p0) {
3404  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3405  int32x2_t __ret;
3406  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
3407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3408  return __ret;
3409}
3410#endif
3411
3412#ifdef __LITTLE_ENDIAN__
3413__ai int16x4_t vcls_s16(int16x4_t __p0) {
3414  int16x4_t __ret;
3415  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
3416  return __ret;
3417}
3418#else
3419__ai int16x4_t vcls_s16(int16x4_t __p0) {
3420  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3421  int16x4_t __ret;
3422  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
3423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3424  return __ret;
3425}
3426#endif
3427
3428#ifdef __LITTLE_ENDIAN__
3429__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3430  uint8x16_t __ret;
3431  __ret = (uint8x16_t)(__p0 < __p1);
3432  return __ret;
3433}
3434#else
3435__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3436  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3437  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3438  uint8x16_t __ret;
3439  __ret = (uint8x16_t)(__rev0 < __rev1);
3440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3441  return __ret;
3442}
3443#endif
3444
3445#ifdef __LITTLE_ENDIAN__
3446__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3447  uint32x4_t __ret;
3448  __ret = (uint32x4_t)(__p0 < __p1);
3449  return __ret;
3450}
3451#else
3452__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3454  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3455  uint32x4_t __ret;
3456  __ret = (uint32x4_t)(__rev0 < __rev1);
3457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3458  return __ret;
3459}
3460#endif
3461
3462#ifdef __LITTLE_ENDIAN__
3463__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3464  uint16x8_t __ret;
3465  __ret = (uint16x8_t)(__p0 < __p1);
3466  return __ret;
3467}
3468#else
3469__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3470  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3471  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3472  uint16x8_t __ret;
3473  __ret = (uint16x8_t)(__rev0 < __rev1);
3474  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3475  return __ret;
3476}
3477#endif
3478
3479#ifdef __LITTLE_ENDIAN__
3480__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3481  uint8x16_t __ret;
3482  __ret = (uint8x16_t)(__p0 < __p1);
3483  return __ret;
3484}
3485#else
3486__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3487  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3488  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3489  uint8x16_t __ret;
3490  __ret = (uint8x16_t)(__rev0 < __rev1);
3491  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3492  return __ret;
3493}
3494#endif
3495
3496#ifdef __LITTLE_ENDIAN__
3497__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3498  uint32x4_t __ret;
3499  __ret = (uint32x4_t)(__p0 < __p1);
3500  return __ret;
3501}
3502#else
3503__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3504  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3505  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3506  uint32x4_t __ret;
3507  __ret = (uint32x4_t)(__rev0 < __rev1);
3508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3509  return __ret;
3510}
3511#endif
3512
3513#ifdef __LITTLE_ENDIAN__
3514__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3515  uint32x4_t __ret;
3516  __ret = (uint32x4_t)(__p0 < __p1);
3517  return __ret;
3518}
3519#else
3520__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3521  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3522  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3523  uint32x4_t __ret;
3524  __ret = (uint32x4_t)(__rev0 < __rev1);
3525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3526  return __ret;
3527}
3528#endif
3529
3530#ifdef __LITTLE_ENDIAN__
3531__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3532  uint16x8_t __ret;
3533  __ret = (uint16x8_t)(__p0 < __p1);
3534  return __ret;
3535}
3536#else
3537__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3538  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3539  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3540  uint16x8_t __ret;
3541  __ret = (uint16x8_t)(__rev0 < __rev1);
3542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3543  return __ret;
3544}
3545#endif
3546
3547#ifdef __LITTLE_ENDIAN__
3548__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3549  uint8x8_t __ret;
3550  __ret = (uint8x8_t)(__p0 < __p1);
3551  return __ret;
3552}
3553#else
3554__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3555  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3556  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3557  uint8x8_t __ret;
3558  __ret = (uint8x8_t)(__rev0 < __rev1);
3559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3560  return __ret;
3561}
3562#endif
3563
3564#ifdef __LITTLE_ENDIAN__
3565__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3566  uint32x2_t __ret;
3567  __ret = (uint32x2_t)(__p0 < __p1);
3568  return __ret;
3569}
3570#else
3571__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3572  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3573  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3574  uint32x2_t __ret;
3575  __ret = (uint32x2_t)(__rev0 < __rev1);
3576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3577  return __ret;
3578}
3579#endif
3580
3581#ifdef __LITTLE_ENDIAN__
3582__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3583  uint16x4_t __ret;
3584  __ret = (uint16x4_t)(__p0 < __p1);
3585  return __ret;
3586}
3587#else
3588__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3589  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3590  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3591  uint16x4_t __ret;
3592  __ret = (uint16x4_t)(__rev0 < __rev1);
3593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3594  return __ret;
3595}
3596#endif
3597
3598#ifdef __LITTLE_ENDIAN__
3599__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3600  uint8x8_t __ret;
3601  __ret = (uint8x8_t)(__p0 < __p1);
3602  return __ret;
3603}
3604#else
3605__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3606  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3607  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3608  uint8x8_t __ret;
3609  __ret = (uint8x8_t)(__rev0 < __rev1);
3610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3611  return __ret;
3612}
3613#endif
3614
3615#ifdef __LITTLE_ENDIAN__
3616__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3617  uint32x2_t __ret;
3618  __ret = (uint32x2_t)(__p0 < __p1);
3619  return __ret;
3620}
3621#else
3622__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3623  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3624  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3625  uint32x2_t __ret;
3626  __ret = (uint32x2_t)(__rev0 < __rev1);
3627  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3628  return __ret;
3629}
3630#endif
3631
3632#ifdef __LITTLE_ENDIAN__
3633__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3634  uint32x2_t __ret;
3635  __ret = (uint32x2_t)(__p0 < __p1);
3636  return __ret;
3637}
3638#else
3639__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3640  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3641  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3642  uint32x2_t __ret;
3643  __ret = (uint32x2_t)(__rev0 < __rev1);
3644  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3645  return __ret;
3646}
3647#endif
3648
3649#ifdef __LITTLE_ENDIAN__
3650__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3651  uint16x4_t __ret;
3652  __ret = (uint16x4_t)(__p0 < __p1);
3653  return __ret;
3654}
3655#else
3656__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3657  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3658  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3659  uint16x4_t __ret;
3660  __ret = (uint16x4_t)(__rev0 < __rev1);
3661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3662  return __ret;
3663}
3664#endif
3665
3666#ifdef __LITTLE_ENDIAN__
3667__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3668  uint8x16_t __ret;
3669  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
3670  return __ret;
3671}
3672#else
3673__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3674  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3675  uint8x16_t __ret;
3676  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
3677  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3678  return __ret;
3679}
3680#endif
3681
3682#ifdef __LITTLE_ENDIAN__
3683__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3684  uint32x4_t __ret;
3685  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
3686  return __ret;
3687}
3688#else
3689__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3690  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3691  uint32x4_t __ret;
3692  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
3693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3694  return __ret;
3695}
3696#endif
3697
3698#ifdef __LITTLE_ENDIAN__
3699__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3700  uint16x8_t __ret;
3701  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
3702  return __ret;
3703}
3704#else
3705__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3706  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3707  uint16x8_t __ret;
3708  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
3709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3710  return __ret;
3711}
3712#endif
3713
3714#ifdef __LITTLE_ENDIAN__
3715__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3716  int8x16_t __ret;
3717  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
3718  return __ret;
3719}
3720#else
3721__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3722  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3723  int8x16_t __ret;
3724  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
3725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3726  return __ret;
3727}
3728#endif
3729
3730#ifdef __LITTLE_ENDIAN__
3731__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3732  int32x4_t __ret;
3733  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
3734  return __ret;
3735}
3736#else
3737__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3738  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3739  int32x4_t __ret;
3740  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
3741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3742  return __ret;
3743}
3744#endif
3745
3746#ifdef __LITTLE_ENDIAN__
3747__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3748  int16x8_t __ret;
3749  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
3750  return __ret;
3751}
3752#else
3753__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3754  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3755  int16x8_t __ret;
3756  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
3757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3758  return __ret;
3759}
3760#endif
3761
3762#ifdef __LITTLE_ENDIAN__
3763__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3764  uint8x8_t __ret;
3765  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
3766  return __ret;
3767}
3768#else
3769__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3770  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3771  uint8x8_t __ret;
3772  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
3773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3774  return __ret;
3775}
3776#endif
3777
3778#ifdef __LITTLE_ENDIAN__
3779__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3780  uint32x2_t __ret;
3781  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
3782  return __ret;
3783}
3784#else
3785__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3786  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3787  uint32x2_t __ret;
3788  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
3789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3790  return __ret;
3791}
3792#endif
3793
3794#ifdef __LITTLE_ENDIAN__
3795__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3796  uint16x4_t __ret;
3797  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
3798  return __ret;
3799}
3800#else
3801__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3802  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3803  uint16x4_t __ret;
3804  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
3805  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3806  return __ret;
3807}
3808#endif
3809
3810#ifdef __LITTLE_ENDIAN__
3811__ai int8x8_t vclz_s8(int8x8_t __p0) {
3812  int8x8_t __ret;
3813  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
3814  return __ret;
3815}
3816#else
3817__ai int8x8_t vclz_s8(int8x8_t __p0) {
3818  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3819  int8x8_t __ret;
3820  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
3821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3822  return __ret;
3823}
3824#endif
3825
3826#ifdef __LITTLE_ENDIAN__
3827__ai int32x2_t vclz_s32(int32x2_t __p0) {
3828  int32x2_t __ret;
3829  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
3830  return __ret;
3831}
3832#else
3833__ai int32x2_t vclz_s32(int32x2_t __p0) {
3834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3835  int32x2_t __ret;
3836  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
3837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3838  return __ret;
3839}
3840#endif
3841
3842#ifdef __LITTLE_ENDIAN__
3843__ai int16x4_t vclz_s16(int16x4_t __p0) {
3844  int16x4_t __ret;
3845  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
3846  return __ret;
3847}
3848#else
3849__ai int16x4_t vclz_s16(int16x4_t __p0) {
3850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3851  int16x4_t __ret;
3852  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
3853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3854  return __ret;
3855}
3856#endif
3857
3858#ifdef __LITTLE_ENDIAN__
3859__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3860  poly8x8_t __ret;
3861  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
3862  return __ret;
3863}
3864#else
3865__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3866  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3867  poly8x8_t __ret;
3868  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
3869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3870  return __ret;
3871}
3872#endif
3873
3874#ifdef __LITTLE_ENDIAN__
3875__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3876  poly8x16_t __ret;
3877  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
3878  return __ret;
3879}
3880#else
3881__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3882  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3883  poly8x16_t __ret;
3884  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
3885  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3886  return __ret;
3887}
3888#endif
3889
3890#ifdef __LITTLE_ENDIAN__
3891__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3892  uint8x16_t __ret;
3893  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
3894  return __ret;
3895}
3896#else
3897__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3898  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3899  uint8x16_t __ret;
3900  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
3901  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3902  return __ret;
3903}
3904#endif
3905
3906#ifdef __LITTLE_ENDIAN__
3907__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3908  int8x16_t __ret;
3909  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
3910  return __ret;
3911}
3912#else
3913__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3914  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3915  int8x16_t __ret;
3916  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
3917  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3918  return __ret;
3919}
3920#endif
3921
3922#ifdef __LITTLE_ENDIAN__
3923__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3924  uint8x8_t __ret;
3925  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
3926  return __ret;
3927}
3928#else
3929__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3930  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3931  uint8x8_t __ret;
3932  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
3933  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3934  return __ret;
3935}
3936#endif
3937
3938#ifdef __LITTLE_ENDIAN__
3939__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3940  int8x8_t __ret;
3941  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
3942  return __ret;
3943}
3944#else
3945__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3946  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3947  int8x8_t __ret;
3948  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
3949  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3950  return __ret;
3951}
3952#endif
3953
3954#ifdef __LITTLE_ENDIAN__
3955__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3956  poly8x16_t __ret;
3957  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3958  return __ret;
3959}
3960#else
3961__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3962  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3963  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3964  poly8x16_t __ret;
3965  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3966  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3967  return __ret;
3968}
3969#endif
3970
3971#ifdef __LITTLE_ENDIAN__
3972__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3973  poly16x8_t __ret;
3974  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
3975  return __ret;
3976}
3977#else
3978__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3979  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3980  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3981  poly16x8_t __ret;
3982  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
3983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3984  return __ret;
3985}
3986#endif
3987
3988#ifdef __LITTLE_ENDIAN__
3989__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3990  uint8x16_t __ret;
3991  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3992  return __ret;
3993}
3994#else
3995__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3996  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3997  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3998  uint8x16_t __ret;
3999  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4000  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4001  return __ret;
4002}
4003__ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
4004  uint8x16_t __ret;
4005  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4006  return __ret;
4007}
4008#endif
4009
4010#ifdef __LITTLE_ENDIAN__
4011__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4012  uint32x4_t __ret;
4013  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4014  return __ret;
4015}
4016#else
4017__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4018  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4019  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4020  uint32x4_t __ret;
4021  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4022  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4023  return __ret;
4024}
4025__ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4026  uint32x4_t __ret;
4027  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4028  return __ret;
4029}
4030#endif
4031
4032#ifdef __LITTLE_ENDIAN__
4033__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4034  uint64x2_t __ret;
4035  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4036  return __ret;
4037}
4038#else
4039__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4040  uint64x2_t __ret;
4041  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4043  return __ret;
4044}
4045#endif
4046
4047#ifdef __LITTLE_ENDIAN__
4048__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4049  uint16x8_t __ret;
4050  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4051  return __ret;
4052}
4053#else
4054__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4055  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4056  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4057  uint16x8_t __ret;
4058  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4060  return __ret;
4061}
4062__ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4063  uint16x8_t __ret;
4064  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4065  return __ret;
4066}
4067#endif
4068
4069#ifdef __LITTLE_ENDIAN__
4070__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4071  int8x16_t __ret;
4072  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4073  return __ret;
4074}
4075#else
4076__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4077  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4078  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4079  int8x16_t __ret;
4080  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4081  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4082  return __ret;
4083}
4084__ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4085  int8x16_t __ret;
4086  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4087  return __ret;
4088}
4089#endif
4090
4091#ifdef __LITTLE_ENDIAN__
4092__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4093  float32x4_t __ret;
4094  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4095  return __ret;
4096}
4097#else
4098__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4099  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4100  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4101  float32x4_t __ret;
4102  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4104  return __ret;
4105}
4106__ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4107  float32x4_t __ret;
4108  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4109  return __ret;
4110}
4111#endif
4112
4113#ifdef __LITTLE_ENDIAN__
4114__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4115  float16x8_t __ret;
4116  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4117  return __ret;
4118}
4119#else
4120__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4121  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4122  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4123  float16x8_t __ret;
4124  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4126  return __ret;
4127}
4128__ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4129  float16x8_t __ret;
4130  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4131  return __ret;
4132}
4133#endif
4134
4135#ifdef __LITTLE_ENDIAN__
4136__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4137  int32x4_t __ret;
4138  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4139  return __ret;
4140}
4141#else
4142__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4143  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4144  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4145  int32x4_t __ret;
4146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4148  return __ret;
4149}
4150__ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4151  int32x4_t __ret;
4152  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4153  return __ret;
4154}
4155#endif
4156
4157#ifdef __LITTLE_ENDIAN__
4158__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4159  int64x2_t __ret;
4160  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4161  return __ret;
4162}
4163#else
4164__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4165  int64x2_t __ret;
4166  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4168  return __ret;
4169}
4170#endif
4171
4172#ifdef __LITTLE_ENDIAN__
4173__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4174  int16x8_t __ret;
4175  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4176  return __ret;
4177}
4178#else
4179__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4180  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4181  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4182  int16x8_t __ret;
4183  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4185  return __ret;
4186}
4187__ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4188  int16x8_t __ret;
4189  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4190  return __ret;
4191}
4192#endif
4193
4194#ifdef __LITTLE_ENDIAN__
4195__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4196  poly8x8_t __ret;
4197  __ret = (poly8x8_t)(__p0);
4198  return __ret;
4199}
4200#else
4201__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4202  poly8x8_t __ret;
4203  __ret = (poly8x8_t)(__p0);
4204  return __ret;
4205}
4206#endif
4207
4208#ifdef __LITTLE_ENDIAN__
4209__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4210  poly16x4_t __ret;
4211  __ret = (poly16x4_t)(__p0);
4212  return __ret;
4213}
4214#else
4215__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4216  poly16x4_t __ret;
4217  __ret = (poly16x4_t)(__p0);
4218  return __ret;
4219}
4220#endif
4221
4222#ifdef __LITTLE_ENDIAN__
4223__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4224  uint8x8_t __ret;
4225  __ret = (uint8x8_t)(__p0);
4226  return __ret;
4227}
4228#else
4229__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4230  uint8x8_t __ret;
4231  __ret = (uint8x8_t)(__p0);
4232  return __ret;
4233}
4234#endif
4235
4236#ifdef __LITTLE_ENDIAN__
4237__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4238  uint32x2_t __ret;
4239  __ret = (uint32x2_t)(__p0);
4240  return __ret;
4241}
4242#else
4243__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4244  uint32x2_t __ret;
4245  __ret = (uint32x2_t)(__p0);
4246  return __ret;
4247}
4248#endif
4249
4250#ifdef __LITTLE_ENDIAN__
4251__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4252  uint64x1_t __ret;
4253  __ret = (uint64x1_t)(__p0);
4254  return __ret;
4255}
4256#else
4257__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4258  uint64x1_t __ret;
4259  __ret = (uint64x1_t)(__p0);
4260  return __ret;
4261}
4262#endif
4263
4264#ifdef __LITTLE_ENDIAN__
4265__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4266  uint16x4_t __ret;
4267  __ret = (uint16x4_t)(__p0);
4268  return __ret;
4269}
4270#else
4271__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4272  uint16x4_t __ret;
4273  __ret = (uint16x4_t)(__p0);
4274  return __ret;
4275}
4276#endif
4277
4278#ifdef __LITTLE_ENDIAN__
4279__ai int8x8_t vcreate_s8(uint64_t __p0) {
4280  int8x8_t __ret;
4281  __ret = (int8x8_t)(__p0);
4282  return __ret;
4283}
4284#else
4285__ai int8x8_t vcreate_s8(uint64_t __p0) {
4286  int8x8_t __ret;
4287  __ret = (int8x8_t)(__p0);
4288  return __ret;
4289}
4290#endif
4291
4292#ifdef __LITTLE_ENDIAN__
4293__ai float32x2_t vcreate_f32(uint64_t __p0) {
4294  float32x2_t __ret;
4295  __ret = (float32x2_t)(__p0);
4296  return __ret;
4297}
4298#else
4299__ai float32x2_t vcreate_f32(uint64_t __p0) {
4300  float32x2_t __ret;
4301  __ret = (float32x2_t)(__p0);
4302  return __ret;
4303}
4304#endif
4305
4306#ifdef __LITTLE_ENDIAN__
4307__ai float16x4_t vcreate_f16(uint64_t __p0) {
4308  float16x4_t __ret;
4309  __ret = (float16x4_t)(__p0);
4310  return __ret;
4311}
4312#else
4313__ai float16x4_t vcreate_f16(uint64_t __p0) {
4314  float16x4_t __ret;
4315  __ret = (float16x4_t)(__p0);
4316  return __ret;
4317}
4318#endif
4319
4320#ifdef __LITTLE_ENDIAN__
4321__ai int32x2_t vcreate_s32(uint64_t __p0) {
4322  int32x2_t __ret;
4323  __ret = (int32x2_t)(__p0);
4324  return __ret;
4325}
4326#else
4327__ai int32x2_t vcreate_s32(uint64_t __p0) {
4328  int32x2_t __ret;
4329  __ret = (int32x2_t)(__p0);
4330  return __ret;
4331}
4332#endif
4333
4334#ifdef __LITTLE_ENDIAN__
4335__ai int64x1_t vcreate_s64(uint64_t __p0) {
4336  int64x1_t __ret;
4337  __ret = (int64x1_t)(__p0);
4338  return __ret;
4339}
4340#else
4341__ai int64x1_t vcreate_s64(uint64_t __p0) {
4342  int64x1_t __ret;
4343  __ret = (int64x1_t)(__p0);
4344  return __ret;
4345}
4346#endif
4347
4348#ifdef __LITTLE_ENDIAN__
4349__ai int16x4_t vcreate_s16(uint64_t __p0) {
4350  int16x4_t __ret;
4351  __ret = (int16x4_t)(__p0);
4352  return __ret;
4353}
4354#else
4355__ai int16x4_t vcreate_s16(uint64_t __p0) {
4356  int16x4_t __ret;
4357  __ret = (int16x4_t)(__p0);
4358  return __ret;
4359}
4360#endif
4361
4362#ifdef __LITTLE_ENDIAN__
4363__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4364  float32x4_t __ret;
4365  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
4366  return __ret;
4367}
4368#else
4369__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4370  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4371  float32x4_t __ret;
4372  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
4373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4374  return __ret;
4375}
4376#endif
4377
4378#ifdef __LITTLE_ENDIAN__
4379__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4380  float32x4_t __ret;
4381  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
4382  return __ret;
4383}
4384#else
4385__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4386  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4387  float32x4_t __ret;
4388  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
4389  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4390  return __ret;
4391}
4392#endif
4393
4394#ifdef __LITTLE_ENDIAN__
4395__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4396  float32x2_t __ret;
4397  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
4398  return __ret;
4399}
4400#else
4401__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4402  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4403  float32x2_t __ret;
4404  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
4405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4406  return __ret;
4407}
4408#endif
4409
4410#ifdef __LITTLE_ENDIAN__
4411__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4412  float32x2_t __ret;
4413  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
4414  return __ret;
4415}
4416#else
4417__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4418  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4419  float32x2_t __ret;
4420  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
4421  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4422  return __ret;
4423}
4424#endif
4425
4426#ifdef __LITTLE_ENDIAN__
4427#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4428  uint32x4_t __s0 = __p0; \
4429  float32x4_t __ret; \
4430  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
4431  __ret; \
4432})
4433#else
4434#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4435  uint32x4_t __s0 = __p0; \
4436  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4437  float32x4_t __ret; \
4438  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
4439  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4440  __ret; \
4441})
4442#endif
4443
4444#ifdef __LITTLE_ENDIAN__
4445#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4446  int32x4_t __s0 = __p0; \
4447  float32x4_t __ret; \
4448  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
4449  __ret; \
4450})
4451#else
4452#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4453  int32x4_t __s0 = __p0; \
4454  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4455  float32x4_t __ret; \
4456  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
4457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4458  __ret; \
4459})
4460#endif
4461
4462#ifdef __LITTLE_ENDIAN__
4463#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4464  uint32x2_t __s0 = __p0; \
4465  float32x2_t __ret; \
4466  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
4467  __ret; \
4468})
4469#else
4470#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4471  uint32x2_t __s0 = __p0; \
4472  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4473  float32x2_t __ret; \
4474  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
4475  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4476  __ret; \
4477})
4478#endif
4479
4480#ifdef __LITTLE_ENDIAN__
4481#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4482  int32x2_t __s0 = __p0; \
4483  float32x2_t __ret; \
4484  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
4485  __ret; \
4486})
4487#else
4488#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4489  int32x2_t __s0 = __p0; \
4490  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4491  float32x2_t __ret; \
4492  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
4493  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4494  __ret; \
4495})
4496#endif
4497
4498#ifdef __LITTLE_ENDIAN__
4499#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4500  float32x4_t __s0 = __p0; \
4501  int32x4_t __ret; \
4502  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
4503  __ret; \
4504})
4505#else
4506#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4507  float32x4_t __s0 = __p0; \
4508  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4509  int32x4_t __ret; \
4510  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
4511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4512  __ret; \
4513})
4514#endif
4515
4516#ifdef __LITTLE_ENDIAN__
4517#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4518  float32x2_t __s0 = __p0; \
4519  int32x2_t __ret; \
4520  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
4521  __ret; \
4522})
4523#else
4524#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4525  float32x2_t __s0 = __p0; \
4526  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4527  int32x2_t __ret; \
4528  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
4529  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4530  __ret; \
4531})
4532#endif
4533
4534#ifdef __LITTLE_ENDIAN__
4535#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4536  float32x4_t __s0 = __p0; \
4537  uint32x4_t __ret; \
4538  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
4539  __ret; \
4540})
4541#else
4542#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4543  float32x4_t __s0 = __p0; \
4544  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4545  uint32x4_t __ret; \
4546  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
4547  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4548  __ret; \
4549})
4550#endif
4551
4552#ifdef __LITTLE_ENDIAN__
4553#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4554  float32x2_t __s0 = __p0; \
4555  uint32x2_t __ret; \
4556  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
4557  __ret; \
4558})
4559#else
4560#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4561  float32x2_t __s0 = __p0; \
4562  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4563  uint32x2_t __ret; \
4564  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
4565  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4566  __ret; \
4567})
4568#endif
4569
4570#ifdef __LITTLE_ENDIAN__
4571__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4572  int32x4_t __ret;
4573  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
4574  return __ret;
4575}
4576#else
4577__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4578  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4579  int32x4_t __ret;
4580  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
4581  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4582  return __ret;
4583}
4584#endif
4585
4586#ifdef __LITTLE_ENDIAN__
4587__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4588  int32x2_t __ret;
4589  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
4590  return __ret;
4591}
4592#else
4593__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4594  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4595  int32x2_t __ret;
4596  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
4597  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4598  return __ret;
4599}
4600#endif
4601
4602#ifdef __LITTLE_ENDIAN__
4603__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4604  uint32x4_t __ret;
4605  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
4606  return __ret;
4607}
4608#else
4609__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4610  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4611  uint32x4_t __ret;
4612  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
4613  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4614  return __ret;
4615}
4616#endif
4617
4618#ifdef __LITTLE_ENDIAN__
4619__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4620  uint32x2_t __ret;
4621  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
4622  return __ret;
4623}
4624#else
4625__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4626  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4627  uint32x2_t __ret;
4628  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
4629  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4630  return __ret;
4631}
4632#endif
4633
4634#ifdef __LITTLE_ENDIAN__
4635#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4636  poly8x8_t __s0 = __p0; \
4637  poly8x8_t __ret; \
4638  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4639  __ret; \
4640})
4641#else
4642#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4643  poly8x8_t __s0 = __p0; \
4644  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4645  poly8x8_t __ret; \
4646  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4647  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4648  __ret; \
4649})
4650#endif
4651
4652#ifdef __LITTLE_ENDIAN__
4653#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4654  poly16x4_t __s0 = __p0; \
4655  poly16x4_t __ret; \
4656  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4657  __ret; \
4658})
4659#else
4660#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4661  poly16x4_t __s0 = __p0; \
4662  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4663  poly16x4_t __ret; \
4664  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4665  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4666  __ret; \
4667})
4668#endif
4669
4670#ifdef __LITTLE_ENDIAN__
4671#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4672  poly8x8_t __s0 = __p0; \
4673  poly8x16_t __ret; \
4674  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4675  __ret; \
4676})
4677#else
4678#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4679  poly8x8_t __s0 = __p0; \
4680  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4681  poly8x16_t __ret; \
4682  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4683  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4684  __ret; \
4685})
4686#endif
4687
4688#ifdef __LITTLE_ENDIAN__
4689#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4690  poly16x4_t __s0 = __p0; \
4691  poly16x8_t __ret; \
4692  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4693  __ret; \
4694})
4695#else
4696#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4697  poly16x4_t __s0 = __p0; \
4698  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4699  poly16x8_t __ret; \
4700  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4701  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4702  __ret; \
4703})
4704#endif
4705
4706#ifdef __LITTLE_ENDIAN__
4707#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4708  uint8x8_t __s0 = __p0; \
4709  uint8x16_t __ret; \
4710  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4711  __ret; \
4712})
4713#else
4714#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4715  uint8x8_t __s0 = __p0; \
4716  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4717  uint8x16_t __ret; \
4718  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4719  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4720  __ret; \
4721})
4722#endif
4723
4724#ifdef __LITTLE_ENDIAN__
4725#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4726  uint32x2_t __s0 = __p0; \
4727  uint32x4_t __ret; \
4728  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4729  __ret; \
4730})
4731#else
4732#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4733  uint32x2_t __s0 = __p0; \
4734  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4735  uint32x4_t __ret; \
4736  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4737  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4738  __ret; \
4739})
4740#endif
4741
4742#ifdef __LITTLE_ENDIAN__
4743#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4744  uint64x1_t __s0 = __p0; \
4745  uint64x2_t __ret; \
4746  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4747  __ret; \
4748})
4749#else
4750#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4751  uint64x1_t __s0 = __p0; \
4752  uint64x2_t __ret; \
4753  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4754  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4755  __ret; \
4756})
4757#endif
4758
4759#ifdef __LITTLE_ENDIAN__
4760#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4761  uint16x4_t __s0 = __p0; \
4762  uint16x8_t __ret; \
4763  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4764  __ret; \
4765})
4766#else
4767#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4768  uint16x4_t __s0 = __p0; \
4769  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4770  uint16x8_t __ret; \
4771  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4772  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4773  __ret; \
4774})
4775#endif
4776
4777#ifdef __LITTLE_ENDIAN__
4778#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4779  int8x8_t __s0 = __p0; \
4780  int8x16_t __ret; \
4781  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4782  __ret; \
4783})
4784#else
4785#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4786  int8x8_t __s0 = __p0; \
4787  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4788  int8x16_t __ret; \
4789  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4790  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4791  __ret; \
4792})
4793#endif
4794
4795#ifdef __LITTLE_ENDIAN__
4796#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4797  float32x2_t __s0 = __p0; \
4798  float32x4_t __ret; \
4799  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4800  __ret; \
4801})
4802#else
4803#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4804  float32x2_t __s0 = __p0; \
4805  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4806  float32x4_t __ret; \
4807  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4808  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4809  __ret; \
4810})
4811#endif
4812
4813#ifdef __LITTLE_ENDIAN__
4814#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4815  int32x2_t __s0 = __p0; \
4816  int32x4_t __ret; \
4817  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4818  __ret; \
4819})
4820#else
4821#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4822  int32x2_t __s0 = __p0; \
4823  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4824  int32x4_t __ret; \
4825  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4826  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4827  __ret; \
4828})
4829#endif
4830
4831#ifdef __LITTLE_ENDIAN__
4832#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4833  int64x1_t __s0 = __p0; \
4834  int64x2_t __ret; \
4835  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4836  __ret; \
4837})
4838#else
4839#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4840  int64x1_t __s0 = __p0; \
4841  int64x2_t __ret; \
4842  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4843  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4844  __ret; \
4845})
4846#endif
4847
4848#ifdef __LITTLE_ENDIAN__
4849#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4850  int16x4_t __s0 = __p0; \
4851  int16x8_t __ret; \
4852  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4853  __ret; \
4854})
4855#else
4856#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4857  int16x4_t __s0 = __p0; \
4858  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4859  int16x8_t __ret; \
4860  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4861  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4862  __ret; \
4863})
4864#endif
4865
4866#ifdef __LITTLE_ENDIAN__
4867#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4868  uint8x8_t __s0 = __p0; \
4869  uint8x8_t __ret; \
4870  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4871  __ret; \
4872})
4873#else
4874#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4875  uint8x8_t __s0 = __p0; \
4876  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4877  uint8x8_t __ret; \
4878  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4880  __ret; \
4881})
4882#endif
4883
4884#ifdef __LITTLE_ENDIAN__
4885#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4886  uint32x2_t __s0 = __p0; \
4887  uint32x2_t __ret; \
4888  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4889  __ret; \
4890})
4891#else
4892#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4893  uint32x2_t __s0 = __p0; \
4894  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4895  uint32x2_t __ret; \
4896  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4897  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4898  __ret; \
4899})
4900#endif
4901
4902#ifdef __LITTLE_ENDIAN__
4903#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4904  uint64x1_t __s0 = __p0; \
4905  uint64x1_t __ret; \
4906  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4907  __ret; \
4908})
4909#else
4910#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4911  uint64x1_t __s0 = __p0; \
4912  uint64x1_t __ret; \
4913  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4914  __ret; \
4915})
4916#endif
4917
4918#ifdef __LITTLE_ENDIAN__
4919#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4920  uint16x4_t __s0 = __p0; \
4921  uint16x4_t __ret; \
4922  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4923  __ret; \
4924})
4925#else
4926#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4927  uint16x4_t __s0 = __p0; \
4928  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4929  uint16x4_t __ret; \
4930  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4931  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4932  __ret; \
4933})
4934#endif
4935
4936#ifdef __LITTLE_ENDIAN__
4937#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4938  int8x8_t __s0 = __p0; \
4939  int8x8_t __ret; \
4940  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4941  __ret; \
4942})
4943#else
4944#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4945  int8x8_t __s0 = __p0; \
4946  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4947  int8x8_t __ret; \
4948  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4949  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4950  __ret; \
4951})
4952#endif
4953
4954#ifdef __LITTLE_ENDIAN__
4955#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
4956  float32x2_t __s0 = __p0; \
4957  float32x2_t __ret; \
4958  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4959  __ret; \
4960})
4961#else
4962#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
4963  float32x2_t __s0 = __p0; \
4964  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4965  float32x2_t __ret; \
4966  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4968  __ret; \
4969})
4970#endif
4971
4972#ifdef __LITTLE_ENDIAN__
4973#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
4974  int32x2_t __s0 = __p0; \
4975  int32x2_t __ret; \
4976  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4977  __ret; \
4978})
4979#else
4980#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
4981  int32x2_t __s0 = __p0; \
4982  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4983  int32x2_t __ret; \
4984  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4985  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4986  __ret; \
4987})
4988#endif
4989
4990#ifdef __LITTLE_ENDIAN__
4991#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
4992  int64x1_t __s0 = __p0; \
4993  int64x1_t __ret; \
4994  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4995  __ret; \
4996})
4997#else
4998#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
4999  int64x1_t __s0 = __p0; \
5000  int64x1_t __ret; \
5001  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5002  __ret; \
5003})
5004#endif
5005
5006#ifdef __LITTLE_ENDIAN__
5007#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5008  int16x4_t __s0 = __p0; \
5009  int16x4_t __ret; \
5010  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
5011  __ret; \
5012})
5013#else
5014#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5015  int16x4_t __s0 = __p0; \
5016  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5017  int16x4_t __ret; \
5018  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
5019  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5020  __ret; \
5021})
5022#endif
5023
5024#ifdef __LITTLE_ENDIAN__
5025__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5026  poly8x8_t __ret;
5027  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5028  return __ret;
5029}
5030#else
5031__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5032  poly8x8_t __ret;
5033  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5034  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5035  return __ret;
5036}
5037#endif
5038
5039#ifdef __LITTLE_ENDIAN__
5040__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5041  poly16x4_t __ret;
5042  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5043  return __ret;
5044}
5045#else
5046__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5047  poly16x4_t __ret;
5048  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5049  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5050  return __ret;
5051}
5052#endif
5053
5054#ifdef __LITTLE_ENDIAN__
5055__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5056  poly8x16_t __ret;
5057  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5058  return __ret;
5059}
5060#else
5061__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5062  poly8x16_t __ret;
5063  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5064  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5065  return __ret;
5066}
5067#endif
5068
5069#ifdef __LITTLE_ENDIAN__
5070__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5071  poly16x8_t __ret;
5072  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5073  return __ret;
5074}
5075#else
5076__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5077  poly16x8_t __ret;
5078  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5079  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5080  return __ret;
5081}
5082#endif
5083
5084#ifdef __LITTLE_ENDIAN__
5085__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5086  uint8x16_t __ret;
5087  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5088  return __ret;
5089}
5090#else
5091__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5092  uint8x16_t __ret;
5093  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5094  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5095  return __ret;
5096}
5097#endif
5098
5099#ifdef __LITTLE_ENDIAN__
5100__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5101  uint32x4_t __ret;
5102  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5103  return __ret;
5104}
5105#else
5106__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5107  uint32x4_t __ret;
5108  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5109  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5110  return __ret;
5111}
5112#endif
5113
5114#ifdef __LITTLE_ENDIAN__
5115__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5116  uint64x2_t __ret;
5117  __ret = (uint64x2_t) {__p0, __p0};
5118  return __ret;
5119}
5120#else
5121__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5122  uint64x2_t __ret;
5123  __ret = (uint64x2_t) {__p0, __p0};
5124  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5125  return __ret;
5126}
5127#endif
5128
5129#ifdef __LITTLE_ENDIAN__
5130__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5131  uint16x8_t __ret;
5132  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5133  return __ret;
5134}
5135#else
5136__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5137  uint16x8_t __ret;
5138  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5139  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5140  return __ret;
5141}
5142#endif
5143
5144#ifdef __LITTLE_ENDIAN__
5145__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5146  int8x16_t __ret;
5147  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5148  return __ret;
5149}
5150#else
5151__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5152  int8x16_t __ret;
5153  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5154  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5155  return __ret;
5156}
5157#endif
5158
5159#ifdef __LITTLE_ENDIAN__
5160__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5161  float32x4_t __ret;
5162  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5163  return __ret;
5164}
5165#else
5166__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5167  float32x4_t __ret;
5168  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5169  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5170  return __ret;
5171}
5172#endif
5173
5174#ifdef __LITTLE_ENDIAN__
5175#define vdupq_n_f16(__p0) __extension__ ({ \
5176  float16_t __s0 = __p0; \
5177  float16x8_t __ret; \
5178  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5179  __ret; \
5180})
5181#else
5182#define vdupq_n_f16(__p0) __extension__ ({ \
5183  float16_t __s0 = __p0; \
5184  float16x8_t __ret; \
5185  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5186  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5187  __ret; \
5188})
5189#endif
5190
5191#ifdef __LITTLE_ENDIAN__
5192__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5193  int32x4_t __ret;
5194  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5195  return __ret;
5196}
5197#else
5198__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5199  int32x4_t __ret;
5200  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5201  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5202  return __ret;
5203}
5204#endif
5205
5206#ifdef __LITTLE_ENDIAN__
5207__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5208  int64x2_t __ret;
5209  __ret = (int64x2_t) {__p0, __p0};
5210  return __ret;
5211}
5212#else
5213__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5214  int64x2_t __ret;
5215  __ret = (int64x2_t) {__p0, __p0};
5216  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5217  return __ret;
5218}
5219#endif
5220
5221#ifdef __LITTLE_ENDIAN__
5222__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5223  int16x8_t __ret;
5224  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5225  return __ret;
5226}
5227#else
5228__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5229  int16x8_t __ret;
5230  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5231  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5232  return __ret;
5233}
5234#endif
5235
5236#ifdef __LITTLE_ENDIAN__
5237__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5238  uint8x8_t __ret;
5239  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5240  return __ret;
5241}
5242#else
5243__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5244  uint8x8_t __ret;
5245  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5246  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5247  return __ret;
5248}
5249#endif
5250
5251#ifdef __LITTLE_ENDIAN__
5252__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5253  uint32x2_t __ret;
5254  __ret = (uint32x2_t) {__p0, __p0};
5255  return __ret;
5256}
5257#else
5258__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5259  uint32x2_t __ret;
5260  __ret = (uint32x2_t) {__p0, __p0};
5261  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5262  return __ret;
5263}
5264#endif
5265
5266#ifdef __LITTLE_ENDIAN__
5267__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5268  uint64x1_t __ret;
5269  __ret = (uint64x1_t) {__p0};
5270  return __ret;
5271}
5272#else
5273__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5274  uint64x1_t __ret;
5275  __ret = (uint64x1_t) {__p0};
5276  return __ret;
5277}
5278#endif
5279
5280#ifdef __LITTLE_ENDIAN__
5281__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5282  uint16x4_t __ret;
5283  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5284  return __ret;
5285}
5286#else
5287__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5288  uint16x4_t __ret;
5289  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5290  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5291  return __ret;
5292}
5293#endif
5294
5295#ifdef __LITTLE_ENDIAN__
5296__ai int8x8_t vdup_n_s8(int8_t __p0) {
5297  int8x8_t __ret;
5298  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5299  return __ret;
5300}
5301#else
5302__ai int8x8_t vdup_n_s8(int8_t __p0) {
5303  int8x8_t __ret;
5304  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5305  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5306  return __ret;
5307}
5308#endif
5309
5310#ifdef __LITTLE_ENDIAN__
5311__ai float32x2_t vdup_n_f32(float32_t __p0) {
5312  float32x2_t __ret;
5313  __ret = (float32x2_t) {__p0, __p0};
5314  return __ret;
5315}
5316#else
5317__ai float32x2_t vdup_n_f32(float32_t __p0) {
5318  float32x2_t __ret;
5319  __ret = (float32x2_t) {__p0, __p0};
5320  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5321  return __ret;
5322}
5323#endif
5324
5325#ifdef __LITTLE_ENDIAN__
5326#define vdup_n_f16(__p0) __extension__ ({ \
5327  float16_t __s0 = __p0; \
5328  float16x4_t __ret; \
5329  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5330  __ret; \
5331})
5332#else
5333#define vdup_n_f16(__p0) __extension__ ({ \
5334  float16_t __s0 = __p0; \
5335  float16x4_t __ret; \
5336  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5337  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5338  __ret; \
5339})
5340#endif
5341
5342#ifdef __LITTLE_ENDIAN__
5343__ai int32x2_t vdup_n_s32(int32_t __p0) {
5344  int32x2_t __ret;
5345  __ret = (int32x2_t) {__p0, __p0};
5346  return __ret;
5347}
5348#else
5349__ai int32x2_t vdup_n_s32(int32_t __p0) {
5350  int32x2_t __ret;
5351  __ret = (int32x2_t) {__p0, __p0};
5352  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5353  return __ret;
5354}
5355#endif
5356
5357#ifdef __LITTLE_ENDIAN__
5358__ai int64x1_t vdup_n_s64(int64_t __p0) {
5359  int64x1_t __ret;
5360  __ret = (int64x1_t) {__p0};
5361  return __ret;
5362}
5363#else
5364__ai int64x1_t vdup_n_s64(int64_t __p0) {
5365  int64x1_t __ret;
5366  __ret = (int64x1_t) {__p0};
5367  return __ret;
5368}
5369#endif
5370
5371#ifdef __LITTLE_ENDIAN__
5372__ai int16x4_t vdup_n_s16(int16_t __p0) {
5373  int16x4_t __ret;
5374  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5375  return __ret;
5376}
5377#else
5378__ai int16x4_t vdup_n_s16(int16_t __p0) {
5379  int16x4_t __ret;
5380  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5381  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5382  return __ret;
5383}
5384#endif
5385
5386#ifdef __LITTLE_ENDIAN__
5387__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5388  uint8x16_t __ret;
5389  __ret = __p0 ^ __p1;
5390  return __ret;
5391}
5392#else
5393__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5394  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5395  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5396  uint8x16_t __ret;
5397  __ret = __rev0 ^ __rev1;
5398  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5399  return __ret;
5400}
5401#endif
5402
5403#ifdef __LITTLE_ENDIAN__
5404__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5405  uint32x4_t __ret;
5406  __ret = __p0 ^ __p1;
5407  return __ret;
5408}
5409#else
5410__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5411  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5412  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5413  uint32x4_t __ret;
5414  __ret = __rev0 ^ __rev1;
5415  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5416  return __ret;
5417}
5418#endif
5419
5420#ifdef __LITTLE_ENDIAN__
5421__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5422  uint64x2_t __ret;
5423  __ret = __p0 ^ __p1;
5424  return __ret;
5425}
5426#else
5427__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5428  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5429  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5430  uint64x2_t __ret;
5431  __ret = __rev0 ^ __rev1;
5432  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5433  return __ret;
5434}
5435#endif
5436
5437#ifdef __LITTLE_ENDIAN__
5438__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5439  uint16x8_t __ret;
5440  __ret = __p0 ^ __p1;
5441  return __ret;
5442}
5443#else
5444__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5445  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5446  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5447  uint16x8_t __ret;
5448  __ret = __rev0 ^ __rev1;
5449  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5450  return __ret;
5451}
5452#endif
5453
5454#ifdef __LITTLE_ENDIAN__
5455__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5456  int8x16_t __ret;
5457  __ret = __p0 ^ __p1;
5458  return __ret;
5459}
5460#else
5461__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5462  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5463  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5464  int8x16_t __ret;
5465  __ret = __rev0 ^ __rev1;
5466  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5467  return __ret;
5468}
5469#endif
5470
5471#ifdef __LITTLE_ENDIAN__
5472__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5473  int32x4_t __ret;
5474  __ret = __p0 ^ __p1;
5475  return __ret;
5476}
5477#else
5478__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5479  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5480  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5481  int32x4_t __ret;
5482  __ret = __rev0 ^ __rev1;
5483  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5484  return __ret;
5485}
5486#endif
5487
5488#ifdef __LITTLE_ENDIAN__
5489__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5490  int64x2_t __ret;
5491  __ret = __p0 ^ __p1;
5492  return __ret;
5493}
5494#else
5495__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5496  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5497  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5498  int64x2_t __ret;
5499  __ret = __rev0 ^ __rev1;
5500  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5501  return __ret;
5502}
5503#endif
5504
5505#ifdef __LITTLE_ENDIAN__
5506__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5507  int16x8_t __ret;
5508  __ret = __p0 ^ __p1;
5509  return __ret;
5510}
5511#else
5512__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5513  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5514  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5515  int16x8_t __ret;
5516  __ret = __rev0 ^ __rev1;
5517  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5518  return __ret;
5519}
5520#endif
5521
5522#ifdef __LITTLE_ENDIAN__
5523__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5524  uint8x8_t __ret;
5525  __ret = __p0 ^ __p1;
5526  return __ret;
5527}
5528#else
5529__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5530  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5531  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5532  uint8x8_t __ret;
5533  __ret = __rev0 ^ __rev1;
5534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5535  return __ret;
5536}
5537#endif
5538
5539#ifdef __LITTLE_ENDIAN__
5540__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5541  uint32x2_t __ret;
5542  __ret = __p0 ^ __p1;
5543  return __ret;
5544}
5545#else
5546__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5547  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5548  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5549  uint32x2_t __ret;
5550  __ret = __rev0 ^ __rev1;
5551  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5552  return __ret;
5553}
5554#endif
5555
5556#ifdef __LITTLE_ENDIAN__
5557__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5558  uint64x1_t __ret;
5559  __ret = __p0 ^ __p1;
5560  return __ret;
5561}
5562#else
5563__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5564  uint64x1_t __ret;
5565  __ret = __p0 ^ __p1;
5566  return __ret;
5567}
5568#endif
5569
5570#ifdef __LITTLE_ENDIAN__
5571__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5572  uint16x4_t __ret;
5573  __ret = __p0 ^ __p1;
5574  return __ret;
5575}
5576#else
5577__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5578  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5579  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5580  uint16x4_t __ret;
5581  __ret = __rev0 ^ __rev1;
5582  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5583  return __ret;
5584}
5585#endif
5586
5587#ifdef __LITTLE_ENDIAN__
5588__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5589  int8x8_t __ret;
5590  __ret = __p0 ^ __p1;
5591  return __ret;
5592}
5593#else
5594__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5595  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5596  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5597  int8x8_t __ret;
5598  __ret = __rev0 ^ __rev1;
5599  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5600  return __ret;
5601}
5602#endif
5603
5604#ifdef __LITTLE_ENDIAN__
5605__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5606  int32x2_t __ret;
5607  __ret = __p0 ^ __p1;
5608  return __ret;
5609}
5610#else
5611__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5612  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5613  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5614  int32x2_t __ret;
5615  __ret = __rev0 ^ __rev1;
5616  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5617  return __ret;
5618}
5619#endif
5620
5621#ifdef __LITTLE_ENDIAN__
5622__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5623  int64x1_t __ret;
5624  __ret = __p0 ^ __p1;
5625  return __ret;
5626}
5627#else
5628__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5629  int64x1_t __ret;
5630  __ret = __p0 ^ __p1;
5631  return __ret;
5632}
5633#endif
5634
5635#ifdef __LITTLE_ENDIAN__
5636__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5637  int16x4_t __ret;
5638  __ret = __p0 ^ __p1;
5639  return __ret;
5640}
5641#else
5642__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5643  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5644  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5645  int16x4_t __ret;
5646  __ret = __rev0 ^ __rev1;
5647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5648  return __ret;
5649}
5650#endif
5651
5652#ifdef __LITTLE_ENDIAN__
5653#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5654  poly8x8_t __s0 = __p0; \
5655  poly8x8_t __s1 = __p1; \
5656  poly8x8_t __ret; \
5657  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
5658  __ret; \
5659})
5660#else
5661#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5662  poly8x8_t __s0 = __p0; \
5663  poly8x8_t __s1 = __p1; \
5664  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5665  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5666  poly8x8_t __ret; \
5667  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
5668  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5669  __ret; \
5670})
5671#endif
5672
5673#ifdef __LITTLE_ENDIAN__
5674#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5675  poly16x4_t __s0 = __p0; \
5676  poly16x4_t __s1 = __p1; \
5677  poly16x4_t __ret; \
5678  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
5679  __ret; \
5680})
5681#else
5682#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5683  poly16x4_t __s0 = __p0; \
5684  poly16x4_t __s1 = __p1; \
5685  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5686  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5687  poly16x4_t __ret; \
5688  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
5689  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5690  __ret; \
5691})
5692#endif
5693
5694#ifdef __LITTLE_ENDIAN__
5695#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5696  poly8x16_t __s0 = __p0; \
5697  poly8x16_t __s1 = __p1; \
5698  poly8x16_t __ret; \
5699  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
5700  __ret; \
5701})
5702#else
5703#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5704  poly8x16_t __s0 = __p0; \
5705  poly8x16_t __s1 = __p1; \
5706  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5707  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5708  poly8x16_t __ret; \
5709  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
5710  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5711  __ret; \
5712})
5713#endif
5714
5715#ifdef __LITTLE_ENDIAN__
5716#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5717  poly16x8_t __s0 = __p0; \
5718  poly16x8_t __s1 = __p1; \
5719  poly16x8_t __ret; \
5720  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
5721  __ret; \
5722})
5723#else
5724#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5725  poly16x8_t __s0 = __p0; \
5726  poly16x8_t __s1 = __p1; \
5727  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5728  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5729  poly16x8_t __ret; \
5730  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
5731  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5732  __ret; \
5733})
5734#endif
5735
5736#ifdef __LITTLE_ENDIAN__
5737#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5738  uint8x16_t __s0 = __p0; \
5739  uint8x16_t __s1 = __p1; \
5740  uint8x16_t __ret; \
5741  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
5742  __ret; \
5743})
5744#else
5745#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5746  uint8x16_t __s0 = __p0; \
5747  uint8x16_t __s1 = __p1; \
5748  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5749  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5750  uint8x16_t __ret; \
5751  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
5752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5753  __ret; \
5754})
5755#endif
5756
5757#ifdef __LITTLE_ENDIAN__
5758#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5759  uint32x4_t __s0 = __p0; \
5760  uint32x4_t __s1 = __p1; \
5761  uint32x4_t __ret; \
5762  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
5763  __ret; \
5764})
5765#else
5766#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5767  uint32x4_t __s0 = __p0; \
5768  uint32x4_t __s1 = __p1; \
5769  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5770  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5771  uint32x4_t __ret; \
5772  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
5773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5774  __ret; \
5775})
5776#endif
5777
5778#ifdef __LITTLE_ENDIAN__
5779#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5780  uint64x2_t __s0 = __p0; \
5781  uint64x2_t __s1 = __p1; \
5782  uint64x2_t __ret; \
5783  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
5784  __ret; \
5785})
5786#else
5787#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5788  uint64x2_t __s0 = __p0; \
5789  uint64x2_t __s1 = __p1; \
5790  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5791  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5792  uint64x2_t __ret; \
5793  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
5794  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5795  __ret; \
5796})
5797#endif
5798
5799#ifdef __LITTLE_ENDIAN__
5800#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5801  uint16x8_t __s0 = __p0; \
5802  uint16x8_t __s1 = __p1; \
5803  uint16x8_t __ret; \
5804  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
5805  __ret; \
5806})
5807#else
5808#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5809  uint16x8_t __s0 = __p0; \
5810  uint16x8_t __s1 = __p1; \
5811  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5812  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5813  uint16x8_t __ret; \
5814  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
5815  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5816  __ret; \
5817})
5818#endif
5819
5820#ifdef __LITTLE_ENDIAN__
5821#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5822  int8x16_t __s0 = __p0; \
5823  int8x16_t __s1 = __p1; \
5824  int8x16_t __ret; \
5825  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
5826  __ret; \
5827})
5828#else
5829#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5830  int8x16_t __s0 = __p0; \
5831  int8x16_t __s1 = __p1; \
5832  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5833  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5834  int8x16_t __ret; \
5835  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
5836  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5837  __ret; \
5838})
5839#endif
5840
5841#ifdef __LITTLE_ENDIAN__
5842#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5843  float32x4_t __s0 = __p0; \
5844  float32x4_t __s1 = __p1; \
5845  float32x4_t __ret; \
5846  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
5847  __ret; \
5848})
5849#else
5850#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5851  float32x4_t __s0 = __p0; \
5852  float32x4_t __s1 = __p1; \
5853  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5854  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5855  float32x4_t __ret; \
5856  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
5857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5858  __ret; \
5859})
5860#endif
5861
5862#ifdef __LITTLE_ENDIAN__
5863#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5864  int32x4_t __s0 = __p0; \
5865  int32x4_t __s1 = __p1; \
5866  int32x4_t __ret; \
5867  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
5868  __ret; \
5869})
5870#else
5871#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5872  int32x4_t __s0 = __p0; \
5873  int32x4_t __s1 = __p1; \
5874  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5875  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5876  int32x4_t __ret; \
5877  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
5878  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5879  __ret; \
5880})
5881#endif
5882
5883#ifdef __LITTLE_ENDIAN__
5884#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5885  int64x2_t __s0 = __p0; \
5886  int64x2_t __s1 = __p1; \
5887  int64x2_t __ret; \
5888  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
5889  __ret; \
5890})
5891#else
5892#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5893  int64x2_t __s0 = __p0; \
5894  int64x2_t __s1 = __p1; \
5895  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5896  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5897  int64x2_t __ret; \
5898  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
5899  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5900  __ret; \
5901})
5902#endif
5903
5904#ifdef __LITTLE_ENDIAN__
5905#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5906  int16x8_t __s0 = __p0; \
5907  int16x8_t __s1 = __p1; \
5908  int16x8_t __ret; \
5909  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
5910  __ret; \
5911})
5912#else
5913#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5914  int16x8_t __s0 = __p0; \
5915  int16x8_t __s1 = __p1; \
5916  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5917  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5918  int16x8_t __ret; \
5919  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
5920  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5921  __ret; \
5922})
5923#endif
5924
5925#ifdef __LITTLE_ENDIAN__
5926#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5927  uint8x8_t __s0 = __p0; \
5928  uint8x8_t __s1 = __p1; \
5929  uint8x8_t __ret; \
5930  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
5931  __ret; \
5932})
5933#else
5934#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5935  uint8x8_t __s0 = __p0; \
5936  uint8x8_t __s1 = __p1; \
5937  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5938  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5939  uint8x8_t __ret; \
5940  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
5941  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5942  __ret; \
5943})
5944#endif
5945
5946#ifdef __LITTLE_ENDIAN__
5947#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5948  uint32x2_t __s0 = __p0; \
5949  uint32x2_t __s1 = __p1; \
5950  uint32x2_t __ret; \
5951  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
5952  __ret; \
5953})
5954#else
5955#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5956  uint32x2_t __s0 = __p0; \
5957  uint32x2_t __s1 = __p1; \
5958  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5959  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5960  uint32x2_t __ret; \
5961  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
5962  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5963  __ret; \
5964})
5965#endif
5966
5967#ifdef __LITTLE_ENDIAN__
5968#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
5969  uint64x1_t __s0 = __p0; \
5970  uint64x1_t __s1 = __p1; \
5971  uint64x1_t __ret; \
5972  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
5973  __ret; \
5974})
5975#else
5976#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
5977  uint64x1_t __s0 = __p0; \
5978  uint64x1_t __s1 = __p1; \
5979  uint64x1_t __ret; \
5980  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
5981  __ret; \
5982})
5983#endif
5984
5985#ifdef __LITTLE_ENDIAN__
5986#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
5987  uint16x4_t __s0 = __p0; \
5988  uint16x4_t __s1 = __p1; \
5989  uint16x4_t __ret; \
5990  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
5991  __ret; \
5992})
5993#else
5994#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
5995  uint16x4_t __s0 = __p0; \
5996  uint16x4_t __s1 = __p1; \
5997  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5998  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5999  uint16x4_t __ret; \
6000  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
6001  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6002  __ret; \
6003})
6004#endif
6005
6006#ifdef __LITTLE_ENDIAN__
6007#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6008  int8x8_t __s0 = __p0; \
6009  int8x8_t __s1 = __p1; \
6010  int8x8_t __ret; \
6011  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
6012  __ret; \
6013})
6014#else
6015#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6016  int8x8_t __s0 = __p0; \
6017  int8x8_t __s1 = __p1; \
6018  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6019  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6020  int8x8_t __ret; \
6021  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
6022  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6023  __ret; \
6024})
6025#endif
6026
6027#ifdef __LITTLE_ENDIAN__
6028#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6029  float32x2_t __s0 = __p0; \
6030  float32x2_t __s1 = __p1; \
6031  float32x2_t __ret; \
6032  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
6033  __ret; \
6034})
6035#else
6036#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6037  float32x2_t __s0 = __p0; \
6038  float32x2_t __s1 = __p1; \
6039  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6040  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6041  float32x2_t __ret; \
6042  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
6043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6044  __ret; \
6045})
6046#endif
6047
6048#ifdef __LITTLE_ENDIAN__
6049#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6050  int32x2_t __s0 = __p0; \
6051  int32x2_t __s1 = __p1; \
6052  int32x2_t __ret; \
6053  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
6054  __ret; \
6055})
6056#else
6057#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6058  int32x2_t __s0 = __p0; \
6059  int32x2_t __s1 = __p1; \
6060  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6061  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6062  int32x2_t __ret; \
6063  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
6064  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6065  __ret; \
6066})
6067#endif
6068
6069#ifdef __LITTLE_ENDIAN__
6070#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6071  int64x1_t __s0 = __p0; \
6072  int64x1_t __s1 = __p1; \
6073  int64x1_t __ret; \
6074  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6075  __ret; \
6076})
6077#else
6078#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6079  int64x1_t __s0 = __p0; \
6080  int64x1_t __s1 = __p1; \
6081  int64x1_t __ret; \
6082  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6083  __ret; \
6084})
6085#endif
6086
6087#ifdef __LITTLE_ENDIAN__
6088#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6089  int16x4_t __s0 = __p0; \
6090  int16x4_t __s1 = __p1; \
6091  int16x4_t __ret; \
6092  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
6093  __ret; \
6094})
6095#else
6096#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6097  int16x4_t __s0 = __p0; \
6098  int16x4_t __s1 = __p1; \
6099  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6100  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6101  int16x4_t __ret; \
6102  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
6103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6104  __ret; \
6105})
6106#endif
6107
6108#ifdef __LITTLE_ENDIAN__
6109__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6110  poly8x8_t __ret;
6111  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6112  return __ret;
6113}
6114#else
6115__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6116  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6117  poly8x8_t __ret;
6118  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6119  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6120  return __ret;
6121}
6122__ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
6123  poly8x8_t __ret;
6124  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6125  return __ret;
6126}
6127#endif
6128
6129#ifdef __LITTLE_ENDIAN__
6130__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6131  poly16x4_t __ret;
6132  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6133  return __ret;
6134}
6135#else
6136__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6137  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6138  poly16x4_t __ret;
6139  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6140  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6141  return __ret;
6142}
6143#endif
6144
6145#ifdef __LITTLE_ENDIAN__
6146__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6147  uint8x8_t __ret;
6148  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6149  return __ret;
6150}
6151#else
6152__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6153  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6154  uint8x8_t __ret;
6155  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6156  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6157  return __ret;
6158}
6159__ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
6160  uint8x8_t __ret;
6161  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6162  return __ret;
6163}
6164#endif
6165
6166#ifdef __LITTLE_ENDIAN__
6167__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6168  uint32x2_t __ret;
6169  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6170  return __ret;
6171}
6172#else
6173__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6174  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6175  uint32x2_t __ret;
6176  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6177  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6178  return __ret;
6179}
6180__ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
6181  uint32x2_t __ret;
6182  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6183  return __ret;
6184}
6185#endif
6186
6187#ifdef __LITTLE_ENDIAN__
6188__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6189  uint64x1_t __ret;
6190  __ret = __builtin_shufflevector(__p0, __p0, 1);
6191  return __ret;
6192}
6193#else
6194__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6195  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6196  uint64x1_t __ret;
6197  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6198  return __ret;
6199}
6200#endif
6201
6202#ifdef __LITTLE_ENDIAN__
6203__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6204  uint16x4_t __ret;
6205  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6206  return __ret;
6207}
6208#else
6209__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6210  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6211  uint16x4_t __ret;
6212  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6213  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6214  return __ret;
6215}
6216__ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
6217  uint16x4_t __ret;
6218  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6219  return __ret;
6220}
6221#endif
6222
6223#ifdef __LITTLE_ENDIAN__
6224__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6225  int8x8_t __ret;
6226  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6227  return __ret;
6228}
6229#else
6230__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6231  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6232  int8x8_t __ret;
6233  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6234  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6235  return __ret;
6236}
6237__ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
6238  int8x8_t __ret;
6239  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6240  return __ret;
6241}
6242#endif
6243
6244#ifdef __LITTLE_ENDIAN__
6245__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6246  float32x2_t __ret;
6247  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6248  return __ret;
6249}
6250#else
6251__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6252  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6253  float32x2_t __ret;
6254  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6255  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6256  return __ret;
6257}
6258__ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
6259  float32x2_t __ret;
6260  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6261  return __ret;
6262}
6263#endif
6264
6265#ifdef __LITTLE_ENDIAN__
6266__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6267  float16x4_t __ret;
6268  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6269  return __ret;
6270}
6271#else
6272__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6273  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6274  float16x4_t __ret;
6275  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6276  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6277  return __ret;
6278}
6279__ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
6280  float16x4_t __ret;
6281  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6282  return __ret;
6283}
6284#endif
6285
6286#ifdef __LITTLE_ENDIAN__
6287__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6288  int32x2_t __ret;
6289  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6290  return __ret;
6291}
6292#else
6293__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6294  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6295  int32x2_t __ret;
6296  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6297  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6298  return __ret;
6299}
6300__ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
6301  int32x2_t __ret;
6302  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6303  return __ret;
6304}
6305#endif
6306
6307#ifdef __LITTLE_ENDIAN__
6308__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6309  int64x1_t __ret;
6310  __ret = __builtin_shufflevector(__p0, __p0, 1);
6311  return __ret;
6312}
6313#else
6314__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6315  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6316  int64x1_t __ret;
6317  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6318  return __ret;
6319}
6320#endif
6321
6322#ifdef __LITTLE_ENDIAN__
6323__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6324  int16x4_t __ret;
6325  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6326  return __ret;
6327}
6328#else
6329__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6330  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6331  int16x4_t __ret;
6332  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6333  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6334  return __ret;
6335}
6336__ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
6337  int16x4_t __ret;
6338  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6339  return __ret;
6340}
6341#endif
6342
6343#ifdef __LITTLE_ENDIAN__
6344#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6345  poly8x8_t __s0 = __p0; \
6346  poly8_t __ret; \
6347  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6348  __ret; \
6349})
6350#else
6351#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6352  poly8x8_t __s0 = __p0; \
6353  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6354  poly8_t __ret; \
6355  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6356  __ret; \
6357})
6358#define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
6359  poly8x8_t __s0 = __p0; \
6360  poly8_t __ret; \
6361  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6362  __ret; \
6363})
6364#endif
6365
6366#ifdef __LITTLE_ENDIAN__
6367#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6368  poly16x4_t __s0 = __p0; \
6369  poly16_t __ret; \
6370  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6371  __ret; \
6372})
6373#else
6374#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6375  poly16x4_t __s0 = __p0; \
6376  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6377  poly16_t __ret; \
6378  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6379  __ret; \
6380})
6381#define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
6382  poly16x4_t __s0 = __p0; \
6383  poly16_t __ret; \
6384  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6385  __ret; \
6386})
6387#endif
6388
6389#ifdef __LITTLE_ENDIAN__
6390#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6391  poly8x16_t __s0 = __p0; \
6392  poly8_t __ret; \
6393  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6394  __ret; \
6395})
6396#else
6397#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6398  poly8x16_t __s0 = __p0; \
6399  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6400  poly8_t __ret; \
6401  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6402  __ret; \
6403})
6404#define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6405  poly8x16_t __s0 = __p0; \
6406  poly8_t __ret; \
6407  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6408  __ret; \
6409})
6410#endif
6411
6412#ifdef __LITTLE_ENDIAN__
6413#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6414  poly16x8_t __s0 = __p0; \
6415  poly16_t __ret; \
6416  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6417  __ret; \
6418})
6419#else
6420#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6421  poly16x8_t __s0 = __p0; \
6422  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6423  poly16_t __ret; \
6424  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6425  __ret; \
6426})
6427#define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6428  poly16x8_t __s0 = __p0; \
6429  poly16_t __ret; \
6430  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6431  __ret; \
6432})
6433#endif
6434
6435#ifdef __LITTLE_ENDIAN__
6436#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6437  uint8x16_t __s0 = __p0; \
6438  uint8_t __ret; \
6439  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6440  __ret; \
6441})
6442#else
6443#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6444  uint8x16_t __s0 = __p0; \
6445  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6446  uint8_t __ret; \
6447  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6448  __ret; \
6449})
6450#define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6451  uint8x16_t __s0 = __p0; \
6452  uint8_t __ret; \
6453  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6454  __ret; \
6455})
6456#endif
6457
6458#ifdef __LITTLE_ENDIAN__
6459#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6460  uint32x4_t __s0 = __p0; \
6461  uint32_t __ret; \
6462  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6463  __ret; \
6464})
6465#else
6466#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6467  uint32x4_t __s0 = __p0; \
6468  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6469  uint32_t __ret; \
6470  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6471  __ret; \
6472})
6473#define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6474  uint32x4_t __s0 = __p0; \
6475  uint32_t __ret; \
6476  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6477  __ret; \
6478})
6479#endif
6480
6481#ifdef __LITTLE_ENDIAN__
6482#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6483  uint64x2_t __s0 = __p0; \
6484  uint64_t __ret; \
6485  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6486  __ret; \
6487})
6488#else
6489#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6490  uint64x2_t __s0 = __p0; \
6491  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6492  uint64_t __ret; \
6493  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6494  __ret; \
6495})
6496#define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6497  uint64x2_t __s0 = __p0; \
6498  uint64_t __ret; \
6499  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6500  __ret; \
6501})
6502#endif
6503
6504#ifdef __LITTLE_ENDIAN__
6505#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6506  uint16x8_t __s0 = __p0; \
6507  uint16_t __ret; \
6508  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6509  __ret; \
6510})
6511#else
6512#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6513  uint16x8_t __s0 = __p0; \
6514  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6515  uint16_t __ret; \
6516  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6517  __ret; \
6518})
6519#define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6520  uint16x8_t __s0 = __p0; \
6521  uint16_t __ret; \
6522  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6523  __ret; \
6524})
6525#endif
6526
6527#ifdef __LITTLE_ENDIAN__
6528#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6529  int8x16_t __s0 = __p0; \
6530  int8_t __ret; \
6531  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6532  __ret; \
6533})
6534#else
6535#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6536  int8x16_t __s0 = __p0; \
6537  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6538  int8_t __ret; \
6539  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6540  __ret; \
6541})
6542#define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6543  int8x16_t __s0 = __p0; \
6544  int8_t __ret; \
6545  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6546  __ret; \
6547})
6548#endif
6549
6550#ifdef __LITTLE_ENDIAN__
6551#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6552  float32x4_t __s0 = __p0; \
6553  float32_t __ret; \
6554  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6555  __ret; \
6556})
6557#else
6558#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6559  float32x4_t __s0 = __p0; \
6560  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6561  float32_t __ret; \
6562  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__rev0, __p1); \
6563  __ret; \
6564})
6565#define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6566  float32x4_t __s0 = __p0; \
6567  float32_t __ret; \
6568  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6569  __ret; \
6570})
6571#endif
6572
6573#ifdef __LITTLE_ENDIAN__
6574#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6575  int32x4_t __s0 = __p0; \
6576  int32_t __ret; \
6577  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6578  __ret; \
6579})
6580#else
6581#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6582  int32x4_t __s0 = __p0; \
6583  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6584  int32_t __ret; \
6585  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6586  __ret; \
6587})
6588#define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6589  int32x4_t __s0 = __p0; \
6590  int32_t __ret; \
6591  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6592  __ret; \
6593})
6594#endif
6595
6596#ifdef __LITTLE_ENDIAN__
6597#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6598  int64x2_t __s0 = __p0; \
6599  int64_t __ret; \
6600  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6601  __ret; \
6602})
6603#else
6604#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6605  int64x2_t __s0 = __p0; \
6606  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6607  int64_t __ret; \
6608  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6609  __ret; \
6610})
6611#define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6612  int64x2_t __s0 = __p0; \
6613  int64_t __ret; \
6614  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6615  __ret; \
6616})
6617#endif
6618
6619#ifdef __LITTLE_ENDIAN__
6620#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6621  int16x8_t __s0 = __p0; \
6622  int16_t __ret; \
6623  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6624  __ret; \
6625})
6626#else
6627#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6628  int16x8_t __s0 = __p0; \
6629  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6630  int16_t __ret; \
6631  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6632  __ret; \
6633})
6634#define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6635  int16x8_t __s0 = __p0; \
6636  int16_t __ret; \
6637  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6638  __ret; \
6639})
6640#endif
6641
6642#ifdef __LITTLE_ENDIAN__
6643#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6644  uint8x8_t __s0 = __p0; \
6645  uint8_t __ret; \
6646  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6647  __ret; \
6648})
6649#else
6650#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6651  uint8x8_t __s0 = __p0; \
6652  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6653  uint8_t __ret; \
6654  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6655  __ret; \
6656})
6657#define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
6658  uint8x8_t __s0 = __p0; \
6659  uint8_t __ret; \
6660  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6661  __ret; \
6662})
6663#endif
6664
6665#ifdef __LITTLE_ENDIAN__
6666#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6667  uint32x2_t __s0 = __p0; \
6668  uint32_t __ret; \
6669  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6670  __ret; \
6671})
6672#else
6673#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6674  uint32x2_t __s0 = __p0; \
6675  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6676  uint32_t __ret; \
6677  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6678  __ret; \
6679})
6680#define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
6681  uint32x2_t __s0 = __p0; \
6682  uint32_t __ret; \
6683  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6684  __ret; \
6685})
6686#endif
6687
6688#ifdef __LITTLE_ENDIAN__
6689#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6690  uint64x1_t __s0 = __p0; \
6691  uint64_t __ret; \
6692  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6693  __ret; \
6694})
6695#else
6696#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6697  uint64x1_t __s0 = __p0; \
6698  uint64_t __ret; \
6699  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6700  __ret; \
6701})
6702#define __noswap_vget_lane_u64(__p0, __p1) __extension__ ({ \
6703  uint64x1_t __s0 = __p0; \
6704  uint64_t __ret; \
6705  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6706  __ret; \
6707})
6708#endif
6709
6710#ifdef __LITTLE_ENDIAN__
6711#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6712  uint16x4_t __s0 = __p0; \
6713  uint16_t __ret; \
6714  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6715  __ret; \
6716})
6717#else
6718#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6719  uint16x4_t __s0 = __p0; \
6720  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6721  uint16_t __ret; \
6722  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6723  __ret; \
6724})
6725#define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
6726  uint16x4_t __s0 = __p0; \
6727  uint16_t __ret; \
6728  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6729  __ret; \
6730})
6731#endif
6732
6733#ifdef __LITTLE_ENDIAN__
6734#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6735  int8x8_t __s0 = __p0; \
6736  int8_t __ret; \
6737  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6738  __ret; \
6739})
6740#else
6741#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6742  int8x8_t __s0 = __p0; \
6743  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6744  int8_t __ret; \
6745  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6746  __ret; \
6747})
6748#define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
6749  int8x8_t __s0 = __p0; \
6750  int8_t __ret; \
6751  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6752  __ret; \
6753})
6754#endif
6755
6756#ifdef __LITTLE_ENDIAN__
6757#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6758  float32x2_t __s0 = __p0; \
6759  float32_t __ret; \
6760  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6761  __ret; \
6762})
6763#else
6764#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6765  float32x2_t __s0 = __p0; \
6766  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6767  float32_t __ret; \
6768  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__rev0, __p1); \
6769  __ret; \
6770})
6771#define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
6772  float32x2_t __s0 = __p0; \
6773  float32_t __ret; \
6774  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6775  __ret; \
6776})
6777#endif
6778
6779#ifdef __LITTLE_ENDIAN__
6780#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6781  int32x2_t __s0 = __p0; \
6782  int32_t __ret; \
6783  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6784  __ret; \
6785})
6786#else
6787#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6788  int32x2_t __s0 = __p0; \
6789  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6790  int32_t __ret; \
6791  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6792  __ret; \
6793})
6794#define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
6795  int32x2_t __s0 = __p0; \
6796  int32_t __ret; \
6797  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6798  __ret; \
6799})
6800#endif
6801
6802#ifdef __LITTLE_ENDIAN__
6803#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6804  int64x1_t __s0 = __p0; \
6805  int64_t __ret; \
6806  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6807  __ret; \
6808})
6809#else
6810#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6811  int64x1_t __s0 = __p0; \
6812  int64_t __ret; \
6813  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6814  __ret; \
6815})
6816#define __noswap_vget_lane_s64(__p0, __p1) __extension__ ({ \
6817  int64x1_t __s0 = __p0; \
6818  int64_t __ret; \
6819  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6820  __ret; \
6821})
6822#endif
6823
6824#ifdef __LITTLE_ENDIAN__
6825#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6826  int16x4_t __s0 = __p0; \
6827  int16_t __ret; \
6828  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6829  __ret; \
6830})
6831#else
6832#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6833  int16x4_t __s0 = __p0; \
6834  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6835  int16_t __ret; \
6836  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6837  __ret; \
6838})
6839#define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
6840  int16x4_t __s0 = __p0; \
6841  int16_t __ret; \
6842  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6843  __ret; \
6844})
6845#endif
6846
6847#ifdef __LITTLE_ENDIAN__
6848__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6849  poly8x8_t __ret;
6850  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6851  return __ret;
6852}
6853#else
6854__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6855  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6856  poly8x8_t __ret;
6857  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6858  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6859  return __ret;
6860}
6861#endif
6862
6863#ifdef __LITTLE_ENDIAN__
6864__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6865  poly16x4_t __ret;
6866  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6867  return __ret;
6868}
6869#else
6870__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6871  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6872  poly16x4_t __ret;
6873  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6874  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6875  return __ret;
6876}
6877#endif
6878
6879#ifdef __LITTLE_ENDIAN__
6880__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6881  uint8x8_t __ret;
6882  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6883  return __ret;
6884}
6885#else
6886__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6887  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6888  uint8x8_t __ret;
6889  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6890  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6891  return __ret;
6892}
6893#endif
6894
6895#ifdef __LITTLE_ENDIAN__
6896__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6897  uint32x2_t __ret;
6898  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6899  return __ret;
6900}
6901#else
6902__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6903  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6904  uint32x2_t __ret;
6905  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
6906  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6907  return __ret;
6908}
6909#endif
6910
6911#ifdef __LITTLE_ENDIAN__
6912__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
6913  uint64x1_t __ret;
6914  __ret = __builtin_shufflevector(__p0, __p0, 0);
6915  return __ret;
6916}
6917#else
6918__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
6919  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6920  uint64x1_t __ret;
6921  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
6922  return __ret;
6923}
6924#endif
6925
6926#ifdef __LITTLE_ENDIAN__
6927__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
6928  uint16x4_t __ret;
6929  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6930  return __ret;
6931}
6932#else
6933__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
6934  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6935  uint16x4_t __ret;
6936  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6937  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6938  return __ret;
6939}
6940#endif
6941
6942#ifdef __LITTLE_ENDIAN__
6943__ai int8x8_t vget_low_s8(int8x16_t __p0) {
6944  int8x8_t __ret;
6945  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6946  return __ret;
6947}
6948#else
6949__ai int8x8_t vget_low_s8(int8x16_t __p0) {
6950  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6951  int8x8_t __ret;
6952  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6953  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6954  return __ret;
6955}
6956#endif
6957
6958#ifdef __LITTLE_ENDIAN__
6959__ai float32x2_t vget_low_f32(float32x4_t __p0) {
6960  float32x2_t __ret;
6961  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6962  return __ret;
6963}
6964#else
6965__ai float32x2_t vget_low_f32(float32x4_t __p0) {
6966  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6967  float32x2_t __ret;
6968  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
6969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6970  return __ret;
6971}
6972#endif
6973
6974#ifdef __LITTLE_ENDIAN__
6975__ai float16x4_t vget_low_f16(float16x8_t __p0) {
6976  float16x4_t __ret;
6977  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6978  return __ret;
6979}
6980#else
6981__ai float16x4_t vget_low_f16(float16x8_t __p0) {
6982  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6983  float16x4_t __ret;
6984  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6985  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6986  return __ret;
6987}
6988#endif
6989
6990#ifdef __LITTLE_ENDIAN__
6991__ai int32x2_t vget_low_s32(int32x4_t __p0) {
6992  int32x2_t __ret;
6993  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6994  return __ret;
6995}
6996#else
6997__ai int32x2_t vget_low_s32(int32x4_t __p0) {
6998  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6999  int32x2_t __ret;
7000  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7002  return __ret;
7003}
7004#endif
7005
7006#ifdef __LITTLE_ENDIAN__
7007__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7008  int64x1_t __ret;
7009  __ret = __builtin_shufflevector(__p0, __p0, 0);
7010  return __ret;
7011}
7012#else
7013__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7014  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7015  int64x1_t __ret;
7016  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7017  return __ret;
7018}
7019#endif
7020
7021#ifdef __LITTLE_ENDIAN__
7022__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7023  int16x4_t __ret;
7024  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7025  return __ret;
7026}
7027#else
7028__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7029  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7030  int16x4_t __ret;
7031  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7032  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7033  return __ret;
7034}
7035#endif
7036
7037#ifdef __LITTLE_ENDIAN__
7038__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7039  uint8x16_t __ret;
7040  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7041  return __ret;
7042}
7043#else
7044__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7045  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7046  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7047  uint8x16_t __ret;
7048  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7049  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7050  return __ret;
7051}
7052#endif
7053
7054#ifdef __LITTLE_ENDIAN__
7055__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7056  uint32x4_t __ret;
7057  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7058  return __ret;
7059}
7060#else
7061__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7062  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7063  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7064  uint32x4_t __ret;
7065  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7066  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7067  return __ret;
7068}
7069#endif
7070
7071#ifdef __LITTLE_ENDIAN__
7072__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7073  uint16x8_t __ret;
7074  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7075  return __ret;
7076}
7077#else
7078__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7079  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7080  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7081  uint16x8_t __ret;
7082  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7083  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7084  return __ret;
7085}
7086#endif
7087
7088#ifdef __LITTLE_ENDIAN__
7089__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7090  int8x16_t __ret;
7091  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7092  return __ret;
7093}
7094#else
7095__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7096  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7097  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7098  int8x16_t __ret;
7099  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7100  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7101  return __ret;
7102}
7103#endif
7104
7105#ifdef __LITTLE_ENDIAN__
7106__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7107  int32x4_t __ret;
7108  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7109  return __ret;
7110}
7111#else
7112__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7113  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7114  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7115  int32x4_t __ret;
7116  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7117  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7118  return __ret;
7119}
7120#endif
7121
7122#ifdef __LITTLE_ENDIAN__
7123__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7124  int16x8_t __ret;
7125  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7126  return __ret;
7127}
7128#else
7129__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7130  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7131  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7132  int16x8_t __ret;
7133  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7134  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7135  return __ret;
7136}
7137#endif
7138
7139#ifdef __LITTLE_ENDIAN__
7140__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7141  uint8x8_t __ret;
7142  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7143  return __ret;
7144}
7145#else
7146__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7147  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7148  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7149  uint8x8_t __ret;
7150  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7151  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7152  return __ret;
7153}
7154#endif
7155
7156#ifdef __LITTLE_ENDIAN__
7157__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7158  uint32x2_t __ret;
7159  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7160  return __ret;
7161}
7162#else
7163__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7164  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7165  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7166  uint32x2_t __ret;
7167  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7168  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7169  return __ret;
7170}
7171#endif
7172
7173#ifdef __LITTLE_ENDIAN__
7174__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7175  uint16x4_t __ret;
7176  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7177  return __ret;
7178}
7179#else
7180__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7181  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7182  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7183  uint16x4_t __ret;
7184  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7185  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7186  return __ret;
7187}
7188#endif
7189
7190#ifdef __LITTLE_ENDIAN__
7191__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7192  int8x8_t __ret;
7193  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7194  return __ret;
7195}
7196#else
7197__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7198  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7199  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7200  int8x8_t __ret;
7201  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7202  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7203  return __ret;
7204}
7205#endif
7206
7207#ifdef __LITTLE_ENDIAN__
7208__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7209  int32x2_t __ret;
7210  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7211  return __ret;
7212}
7213#else
7214__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7215  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7216  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7217  int32x2_t __ret;
7218  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7219  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7220  return __ret;
7221}
7222#endif
7223
7224#ifdef __LITTLE_ENDIAN__
7225__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7226  int16x4_t __ret;
7227  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7228  return __ret;
7229}
7230#else
7231__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7232  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7233  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7234  int16x4_t __ret;
7235  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7237  return __ret;
7238}
7239#endif
7240
7241#ifdef __LITTLE_ENDIAN__
7242__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7243  uint8x16_t __ret;
7244  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7245  return __ret;
7246}
7247#else
7248__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7249  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7250  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7251  uint8x16_t __ret;
7252  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7253  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7254  return __ret;
7255}
7256#endif
7257
7258#ifdef __LITTLE_ENDIAN__
7259__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7260  uint32x4_t __ret;
7261  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7262  return __ret;
7263}
7264#else
7265__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7266  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7267  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7268  uint32x4_t __ret;
7269  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7271  return __ret;
7272}
7273#endif
7274
7275#ifdef __LITTLE_ENDIAN__
7276__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7277  uint16x8_t __ret;
7278  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7279  return __ret;
7280}
7281#else
7282__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7283  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7284  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7285  uint16x8_t __ret;
7286  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7287  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7288  return __ret;
7289}
7290#endif
7291
7292#ifdef __LITTLE_ENDIAN__
7293__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7294  int8x16_t __ret;
7295  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7296  return __ret;
7297}
7298#else
7299__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7300  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7301  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7302  int8x16_t __ret;
7303  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7304  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7305  return __ret;
7306}
7307#endif
7308
7309#ifdef __LITTLE_ENDIAN__
7310__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7311  int32x4_t __ret;
7312  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7313  return __ret;
7314}
7315#else
7316__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7317  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7318  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7319  int32x4_t __ret;
7320  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7321  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7322  return __ret;
7323}
7324#endif
7325
7326#ifdef __LITTLE_ENDIAN__
7327__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7328  int16x8_t __ret;
7329  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7330  return __ret;
7331}
7332#else
7333__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7334  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7335  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7336  int16x8_t __ret;
7337  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7338  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7339  return __ret;
7340}
7341#endif
7342
7343#ifdef __LITTLE_ENDIAN__
7344__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7345  uint8x8_t __ret;
7346  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7347  return __ret;
7348}
7349#else
7350__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7351  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7352  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7353  uint8x8_t __ret;
7354  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7355  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7356  return __ret;
7357}
7358#endif
7359
7360#ifdef __LITTLE_ENDIAN__
7361__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7362  uint32x2_t __ret;
7363  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7364  return __ret;
7365}
7366#else
7367__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7368  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7369  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7370  uint32x2_t __ret;
7371  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7372  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7373  return __ret;
7374}
7375#endif
7376
7377#ifdef __LITTLE_ENDIAN__
7378__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7379  uint16x4_t __ret;
7380  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7381  return __ret;
7382}
7383#else
7384__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7385  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7386  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7387  uint16x4_t __ret;
7388  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7389  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7390  return __ret;
7391}
7392#endif
7393
7394#ifdef __LITTLE_ENDIAN__
7395__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7396  int8x8_t __ret;
7397  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7398  return __ret;
7399}
7400#else
7401__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7402  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7403  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7404  int8x8_t __ret;
7405  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7406  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7407  return __ret;
7408}
7409#endif
7410
7411#ifdef __LITTLE_ENDIAN__
7412__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7413  int32x2_t __ret;
7414  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7415  return __ret;
7416}
7417#else
7418__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7419  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7420  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7421  int32x2_t __ret;
7422  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7423  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7424  return __ret;
7425}
7426#endif
7427
7428#ifdef __LITTLE_ENDIAN__
7429__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7430  int16x4_t __ret;
7431  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7432  return __ret;
7433}
7434#else
7435__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7436  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7437  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7438  int16x4_t __ret;
7439  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7441  return __ret;
7442}
7443#endif
7444
7445#ifdef __LITTLE_ENDIAN__
7446#define vld1_p8(__p0) __extension__ ({ \
7447  poly8x8_t __ret; \
7448  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7449  __ret; \
7450})
7451#else
7452#define vld1_p8(__p0) __extension__ ({ \
7453  poly8x8_t __ret; \
7454  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7456  __ret; \
7457})
7458#endif
7459
7460#ifdef __LITTLE_ENDIAN__
7461#define vld1_p16(__p0) __extension__ ({ \
7462  poly16x4_t __ret; \
7463  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7464  __ret; \
7465})
7466#else
7467#define vld1_p16(__p0) __extension__ ({ \
7468  poly16x4_t __ret; \
7469  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7470  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7471  __ret; \
7472})
7473#endif
7474
7475#ifdef __LITTLE_ENDIAN__
7476#define vld1q_p8(__p0) __extension__ ({ \
7477  poly8x16_t __ret; \
7478  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7479  __ret; \
7480})
7481#else
7482#define vld1q_p8(__p0) __extension__ ({ \
7483  poly8x16_t __ret; \
7484  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7485  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7486  __ret; \
7487})
7488#endif
7489
7490#ifdef __LITTLE_ENDIAN__
7491#define vld1q_p16(__p0) __extension__ ({ \
7492  poly16x8_t __ret; \
7493  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7494  __ret; \
7495})
7496#else
7497#define vld1q_p16(__p0) __extension__ ({ \
7498  poly16x8_t __ret; \
7499  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7500  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7501  __ret; \
7502})
7503#endif
7504
7505#ifdef __LITTLE_ENDIAN__
7506#define vld1q_u8(__p0) __extension__ ({ \
7507  uint8x16_t __ret; \
7508  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7509  __ret; \
7510})
7511#else
7512#define vld1q_u8(__p0) __extension__ ({ \
7513  uint8x16_t __ret; \
7514  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7515  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7516  __ret; \
7517})
7518#endif
7519
7520#ifdef __LITTLE_ENDIAN__
7521#define vld1q_u32(__p0) __extension__ ({ \
7522  uint32x4_t __ret; \
7523  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7524  __ret; \
7525})
7526#else
7527#define vld1q_u32(__p0) __extension__ ({ \
7528  uint32x4_t __ret; \
7529  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7530  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7531  __ret; \
7532})
7533#endif
7534
7535#ifdef __LITTLE_ENDIAN__
7536#define vld1q_u64(__p0) __extension__ ({ \
7537  uint64x2_t __ret; \
7538  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7539  __ret; \
7540})
7541#else
7542#define vld1q_u64(__p0) __extension__ ({ \
7543  uint64x2_t __ret; \
7544  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7546  __ret; \
7547})
7548#endif
7549
7550#ifdef __LITTLE_ENDIAN__
7551#define vld1q_u16(__p0) __extension__ ({ \
7552  uint16x8_t __ret; \
7553  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7554  __ret; \
7555})
7556#else
7557#define vld1q_u16(__p0) __extension__ ({ \
7558  uint16x8_t __ret; \
7559  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7560  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7561  __ret; \
7562})
7563#endif
7564
7565#ifdef __LITTLE_ENDIAN__
7566#define vld1q_s8(__p0) __extension__ ({ \
7567  int8x16_t __ret; \
7568  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7569  __ret; \
7570})
7571#else
7572#define vld1q_s8(__p0) __extension__ ({ \
7573  int8x16_t __ret; \
7574  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7575  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7576  __ret; \
7577})
7578#endif
7579
7580#ifdef __LITTLE_ENDIAN__
7581#define vld1q_f32(__p0) __extension__ ({ \
7582  float32x4_t __ret; \
7583  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7584  __ret; \
7585})
7586#else
7587#define vld1q_f32(__p0) __extension__ ({ \
7588  float32x4_t __ret; \
7589  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7590  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7591  __ret; \
7592})
7593#endif
7594
7595#ifdef __LITTLE_ENDIAN__
7596#define vld1q_f16(__p0) __extension__ ({ \
7597  float16x8_t __ret; \
7598  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7599  __ret; \
7600})
7601#else
7602#define vld1q_f16(__p0) __extension__ ({ \
7603  float16x8_t __ret; \
7604  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7605  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7606  __ret; \
7607})
7608#endif
7609
7610#ifdef __LITTLE_ENDIAN__
7611#define vld1q_s32(__p0) __extension__ ({ \
7612  int32x4_t __ret; \
7613  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7614  __ret; \
7615})
7616#else
7617#define vld1q_s32(__p0) __extension__ ({ \
7618  int32x4_t __ret; \
7619  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7620  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7621  __ret; \
7622})
7623#endif
7624
7625#ifdef __LITTLE_ENDIAN__
7626#define vld1q_s64(__p0) __extension__ ({ \
7627  int64x2_t __ret; \
7628  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7629  __ret; \
7630})
7631#else
7632#define vld1q_s64(__p0) __extension__ ({ \
7633  int64x2_t __ret; \
7634  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7635  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7636  __ret; \
7637})
7638#endif
7639
7640#ifdef __LITTLE_ENDIAN__
7641#define vld1q_s16(__p0) __extension__ ({ \
7642  int16x8_t __ret; \
7643  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7644  __ret; \
7645})
7646#else
7647#define vld1q_s16(__p0) __extension__ ({ \
7648  int16x8_t __ret; \
7649  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7650  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7651  __ret; \
7652})
7653#endif
7654
7655#ifdef __LITTLE_ENDIAN__
7656#define vld1_u8(__p0) __extension__ ({ \
7657  uint8x8_t __ret; \
7658  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7659  __ret; \
7660})
7661#else
7662#define vld1_u8(__p0) __extension__ ({ \
7663  uint8x8_t __ret; \
7664  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7665  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7666  __ret; \
7667})
7668#endif
7669
7670#ifdef __LITTLE_ENDIAN__
7671#define vld1_u32(__p0) __extension__ ({ \
7672  uint32x2_t __ret; \
7673  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7674  __ret; \
7675})
7676#else
7677#define vld1_u32(__p0) __extension__ ({ \
7678  uint32x2_t __ret; \
7679  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7681  __ret; \
7682})
7683#endif
7684
7685#ifdef __LITTLE_ENDIAN__
7686#define vld1_u64(__p0) __extension__ ({ \
7687  uint64x1_t __ret; \
7688  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7689  __ret; \
7690})
7691#else
7692#define vld1_u64(__p0) __extension__ ({ \
7693  uint64x1_t __ret; \
7694  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7695  __ret; \
7696})
7697#endif
7698
7699#ifdef __LITTLE_ENDIAN__
7700#define vld1_u16(__p0) __extension__ ({ \
7701  uint16x4_t __ret; \
7702  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7703  __ret; \
7704})
7705#else
7706#define vld1_u16(__p0) __extension__ ({ \
7707  uint16x4_t __ret; \
7708  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7709  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7710  __ret; \
7711})
7712#endif
7713
7714#ifdef __LITTLE_ENDIAN__
7715#define vld1_s8(__p0) __extension__ ({ \
7716  int8x8_t __ret; \
7717  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7718  __ret; \
7719})
7720#else
7721#define vld1_s8(__p0) __extension__ ({ \
7722  int8x8_t __ret; \
7723  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7724  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7725  __ret; \
7726})
7727#endif
7728
7729#ifdef __LITTLE_ENDIAN__
7730#define vld1_f32(__p0) __extension__ ({ \
7731  float32x2_t __ret; \
7732  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7733  __ret; \
7734})
7735#else
7736#define vld1_f32(__p0) __extension__ ({ \
7737  float32x2_t __ret; \
7738  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7739  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7740  __ret; \
7741})
7742#endif
7743
7744#ifdef __LITTLE_ENDIAN__
7745#define vld1_f16(__p0) __extension__ ({ \
7746  float16x4_t __ret; \
7747  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7748  __ret; \
7749})
7750#else
7751#define vld1_f16(__p0) __extension__ ({ \
7752  float16x4_t __ret; \
7753  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7754  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7755  __ret; \
7756})
7757#endif
7758
7759#ifdef __LITTLE_ENDIAN__
7760#define vld1_s32(__p0) __extension__ ({ \
7761  int32x2_t __ret; \
7762  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7763  __ret; \
7764})
7765#else
7766#define vld1_s32(__p0) __extension__ ({ \
7767  int32x2_t __ret; \
7768  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7770  __ret; \
7771})
7772#endif
7773
7774#ifdef __LITTLE_ENDIAN__
7775#define vld1_s64(__p0) __extension__ ({ \
7776  int64x1_t __ret; \
7777  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7778  __ret; \
7779})
7780#else
7781#define vld1_s64(__p0) __extension__ ({ \
7782  int64x1_t __ret; \
7783  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7784  __ret; \
7785})
7786#endif
7787
7788#ifdef __LITTLE_ENDIAN__
7789#define vld1_s16(__p0) __extension__ ({ \
7790  int16x4_t __ret; \
7791  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7792  __ret; \
7793})
7794#else
7795#define vld1_s16(__p0) __extension__ ({ \
7796  int16x4_t __ret; \
7797  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7798  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7799  __ret; \
7800})
7801#endif
7802
7803#ifdef __LITTLE_ENDIAN__
7804#define vld1_dup_p8(__p0) __extension__ ({ \
7805  poly8x8_t __ret; \
7806  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7807  __ret; \
7808})
7809#else
7810#define vld1_dup_p8(__p0) __extension__ ({ \
7811  poly8x8_t __ret; \
7812  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7813  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7814  __ret; \
7815})
7816#endif
7817
7818#ifdef __LITTLE_ENDIAN__
7819#define vld1_dup_p16(__p0) __extension__ ({ \
7820  poly16x4_t __ret; \
7821  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7822  __ret; \
7823})
7824#else
7825#define vld1_dup_p16(__p0) __extension__ ({ \
7826  poly16x4_t __ret; \
7827  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7828  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7829  __ret; \
7830})
7831#endif
7832
7833#ifdef __LITTLE_ENDIAN__
7834#define vld1q_dup_p8(__p0) __extension__ ({ \
7835  poly8x16_t __ret; \
7836  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7837  __ret; \
7838})
7839#else
7840#define vld1q_dup_p8(__p0) __extension__ ({ \
7841  poly8x16_t __ret; \
7842  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7843  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7844  __ret; \
7845})
7846#endif
7847
7848#ifdef __LITTLE_ENDIAN__
7849#define vld1q_dup_p16(__p0) __extension__ ({ \
7850  poly16x8_t __ret; \
7851  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7852  __ret; \
7853})
7854#else
7855#define vld1q_dup_p16(__p0) __extension__ ({ \
7856  poly16x8_t __ret; \
7857  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7858  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7859  __ret; \
7860})
7861#endif
7862
7863#ifdef __LITTLE_ENDIAN__
7864#define vld1q_dup_u8(__p0) __extension__ ({ \
7865  uint8x16_t __ret; \
7866  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7867  __ret; \
7868})
7869#else
7870#define vld1q_dup_u8(__p0) __extension__ ({ \
7871  uint8x16_t __ret; \
7872  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7873  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7874  __ret; \
7875})
7876#endif
7877
7878#ifdef __LITTLE_ENDIAN__
7879#define vld1q_dup_u32(__p0) __extension__ ({ \
7880  uint32x4_t __ret; \
7881  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7882  __ret; \
7883})
7884#else
7885#define vld1q_dup_u32(__p0) __extension__ ({ \
7886  uint32x4_t __ret; \
7887  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7888  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7889  __ret; \
7890})
7891#endif
7892
7893#ifdef __LITTLE_ENDIAN__
7894#define vld1q_dup_u64(__p0) __extension__ ({ \
7895  uint64x2_t __ret; \
7896  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7897  __ret; \
7898})
7899#else
7900#define vld1q_dup_u64(__p0) __extension__ ({ \
7901  uint64x2_t __ret; \
7902  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7903  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7904  __ret; \
7905})
7906#endif
7907
7908#ifdef __LITTLE_ENDIAN__
7909#define vld1q_dup_u16(__p0) __extension__ ({ \
7910  uint16x8_t __ret; \
7911  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
7912  __ret; \
7913})
7914#else
7915#define vld1q_dup_u16(__p0) __extension__ ({ \
7916  uint16x8_t __ret; \
7917  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
7918  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7919  __ret; \
7920})
7921#endif
7922
7923#ifdef __LITTLE_ENDIAN__
7924#define vld1q_dup_s8(__p0) __extension__ ({ \
7925  int8x16_t __ret; \
7926  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
7927  __ret; \
7928})
7929#else
7930#define vld1q_dup_s8(__p0) __extension__ ({ \
7931  int8x16_t __ret; \
7932  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
7933  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7934  __ret; \
7935})
7936#endif
7937
7938#ifdef __LITTLE_ENDIAN__
7939#define vld1q_dup_f32(__p0) __extension__ ({ \
7940  float32x4_t __ret; \
7941  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
7942  __ret; \
7943})
7944#else
7945#define vld1q_dup_f32(__p0) __extension__ ({ \
7946  float32x4_t __ret; \
7947  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
7948  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7949  __ret; \
7950})
7951#endif
7952
7953#ifdef __LITTLE_ENDIAN__
7954#define vld1q_dup_f16(__p0) __extension__ ({ \
7955  float16x8_t __ret; \
7956  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
7957  __ret; \
7958})
7959#else
7960#define vld1q_dup_f16(__p0) __extension__ ({ \
7961  float16x8_t __ret; \
7962  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
7963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7964  __ret; \
7965})
7966#endif
7967
7968#ifdef __LITTLE_ENDIAN__
7969#define vld1q_dup_s32(__p0) __extension__ ({ \
7970  int32x4_t __ret; \
7971  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
7972  __ret; \
7973})
7974#else
7975#define vld1q_dup_s32(__p0) __extension__ ({ \
7976  int32x4_t __ret; \
7977  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
7978  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7979  __ret; \
7980})
7981#endif
7982
7983#ifdef __LITTLE_ENDIAN__
7984#define vld1q_dup_s64(__p0) __extension__ ({ \
7985  int64x2_t __ret; \
7986  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
7987  __ret; \
7988})
7989#else
7990#define vld1q_dup_s64(__p0) __extension__ ({ \
7991  int64x2_t __ret; \
7992  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
7993  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7994  __ret; \
7995})
7996#endif
7997
7998#ifdef __LITTLE_ENDIAN__
7999#define vld1q_dup_s16(__p0) __extension__ ({ \
8000  int16x8_t __ret; \
8001  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8002  __ret; \
8003})
8004#else
8005#define vld1q_dup_s16(__p0) __extension__ ({ \
8006  int16x8_t __ret; \
8007  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8008  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8009  __ret; \
8010})
8011#endif
8012
8013#ifdef __LITTLE_ENDIAN__
8014#define vld1_dup_u8(__p0) __extension__ ({ \
8015  uint8x8_t __ret; \
8016  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8017  __ret; \
8018})
8019#else
8020#define vld1_dup_u8(__p0) __extension__ ({ \
8021  uint8x8_t __ret; \
8022  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8024  __ret; \
8025})
8026#endif
8027
8028#ifdef __LITTLE_ENDIAN__
8029#define vld1_dup_u32(__p0) __extension__ ({ \
8030  uint32x2_t __ret; \
8031  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8032  __ret; \
8033})
8034#else
8035#define vld1_dup_u32(__p0) __extension__ ({ \
8036  uint32x2_t __ret; \
8037  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8039  __ret; \
8040})
8041#endif
8042
8043#ifdef __LITTLE_ENDIAN__
8044#define vld1_dup_u64(__p0) __extension__ ({ \
8045  uint64x1_t __ret; \
8046  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8047  __ret; \
8048})
8049#else
8050#define vld1_dup_u64(__p0) __extension__ ({ \
8051  uint64x1_t __ret; \
8052  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8053  __ret; \
8054})
8055#endif
8056
8057#ifdef __LITTLE_ENDIAN__
8058#define vld1_dup_u16(__p0) __extension__ ({ \
8059  uint16x4_t __ret; \
8060  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8061  __ret; \
8062})
8063#else
8064#define vld1_dup_u16(__p0) __extension__ ({ \
8065  uint16x4_t __ret; \
8066  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8067  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8068  __ret; \
8069})
8070#endif
8071
8072#ifdef __LITTLE_ENDIAN__
8073#define vld1_dup_s8(__p0) __extension__ ({ \
8074  int8x8_t __ret; \
8075  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8076  __ret; \
8077})
8078#else
8079#define vld1_dup_s8(__p0) __extension__ ({ \
8080  int8x8_t __ret; \
8081  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8082  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8083  __ret; \
8084})
8085#endif
8086
8087#ifdef __LITTLE_ENDIAN__
8088#define vld1_dup_f32(__p0) __extension__ ({ \
8089  float32x2_t __ret; \
8090  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8091  __ret; \
8092})
8093#else
8094#define vld1_dup_f32(__p0) __extension__ ({ \
8095  float32x2_t __ret; \
8096  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8097  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8098  __ret; \
8099})
8100#endif
8101
8102#ifdef __LITTLE_ENDIAN__
8103#define vld1_dup_f16(__p0) __extension__ ({ \
8104  float16x4_t __ret; \
8105  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8106  __ret; \
8107})
8108#else
8109#define vld1_dup_f16(__p0) __extension__ ({ \
8110  float16x4_t __ret; \
8111  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8112  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8113  __ret; \
8114})
8115#endif
8116
8117#ifdef __LITTLE_ENDIAN__
8118#define vld1_dup_s32(__p0) __extension__ ({ \
8119  int32x2_t __ret; \
8120  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8121  __ret; \
8122})
8123#else
8124#define vld1_dup_s32(__p0) __extension__ ({ \
8125  int32x2_t __ret; \
8126  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8127  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8128  __ret; \
8129})
8130#endif
8131
8132#ifdef __LITTLE_ENDIAN__
8133#define vld1_dup_s64(__p0) __extension__ ({ \
8134  int64x1_t __ret; \
8135  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8136  __ret; \
8137})
8138#else
8139#define vld1_dup_s64(__p0) __extension__ ({ \
8140  int64x1_t __ret; \
8141  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8142  __ret; \
8143})
8144#endif
8145
8146#ifdef __LITTLE_ENDIAN__
8147#define vld1_dup_s16(__p0) __extension__ ({ \
8148  int16x4_t __ret; \
8149  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8150  __ret; \
8151})
8152#else
8153#define vld1_dup_s16(__p0) __extension__ ({ \
8154  int16x4_t __ret; \
8155  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8156  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8157  __ret; \
8158})
8159#endif
8160
8161#ifdef __LITTLE_ENDIAN__
8162#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8163  poly8x8_t __s1 = __p1; \
8164  poly8x8_t __ret; \
8165  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
8166  __ret; \
8167})
8168#else
8169#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8170  poly8x8_t __s1 = __p1; \
8171  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8172  poly8x8_t __ret; \
8173  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
8174  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8175  __ret; \
8176})
8177#endif
8178
8179#ifdef __LITTLE_ENDIAN__
8180#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8181  poly16x4_t __s1 = __p1; \
8182  poly16x4_t __ret; \
8183  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
8184  __ret; \
8185})
8186#else
8187#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8188  poly16x4_t __s1 = __p1; \
8189  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8190  poly16x4_t __ret; \
8191  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
8192  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8193  __ret; \
8194})
8195#endif
8196
8197#ifdef __LITTLE_ENDIAN__
8198#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8199  poly8x16_t __s1 = __p1; \
8200  poly8x16_t __ret; \
8201  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
8202  __ret; \
8203})
8204#else
8205#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8206  poly8x16_t __s1 = __p1; \
8207  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8208  poly8x16_t __ret; \
8209  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
8210  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8211  __ret; \
8212})
8213#endif
8214
8215#ifdef __LITTLE_ENDIAN__
8216#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8217  poly16x8_t __s1 = __p1; \
8218  poly16x8_t __ret; \
8219  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
8220  __ret; \
8221})
8222#else
8223#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8224  poly16x8_t __s1 = __p1; \
8225  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8226  poly16x8_t __ret; \
8227  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
8228  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8229  __ret; \
8230})
8231#endif
8232
8233#ifdef __LITTLE_ENDIAN__
8234#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8235  uint8x16_t __s1 = __p1; \
8236  uint8x16_t __ret; \
8237  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
8238  __ret; \
8239})
8240#else
8241#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8242  uint8x16_t __s1 = __p1; \
8243  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8244  uint8x16_t __ret; \
8245  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
8246  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8247  __ret; \
8248})
8249#endif
8250
8251#ifdef __LITTLE_ENDIAN__
8252#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8253  uint32x4_t __s1 = __p1; \
8254  uint32x4_t __ret; \
8255  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
8256  __ret; \
8257})
8258#else
8259#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8260  uint32x4_t __s1 = __p1; \
8261  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8262  uint32x4_t __ret; \
8263  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
8264  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8265  __ret; \
8266})
8267#endif
8268
8269#ifdef __LITTLE_ENDIAN__
8270#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8271  uint64x2_t __s1 = __p1; \
8272  uint64x2_t __ret; \
8273  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
8274  __ret; \
8275})
8276#else
8277#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8278  uint64x2_t __s1 = __p1; \
8279  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8280  uint64x2_t __ret; \
8281  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
8282  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8283  __ret; \
8284})
8285#endif
8286
8287#ifdef __LITTLE_ENDIAN__
8288#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8289  uint16x8_t __s1 = __p1; \
8290  uint16x8_t __ret; \
8291  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
8292  __ret; \
8293})
8294#else
8295#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8296  uint16x8_t __s1 = __p1; \
8297  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8298  uint16x8_t __ret; \
8299  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
8300  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8301  __ret; \
8302})
8303#endif
8304
8305#ifdef __LITTLE_ENDIAN__
8306#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8307  int8x16_t __s1 = __p1; \
8308  int8x16_t __ret; \
8309  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
8310  __ret; \
8311})
8312#else
8313#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8314  int8x16_t __s1 = __p1; \
8315  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8316  int8x16_t __ret; \
8317  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
8318  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8319  __ret; \
8320})
8321#endif
8322
8323#ifdef __LITTLE_ENDIAN__
8324#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8325  float32x4_t __s1 = __p1; \
8326  float32x4_t __ret; \
8327  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
8328  __ret; \
8329})
8330#else
8331#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8332  float32x4_t __s1 = __p1; \
8333  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8334  float32x4_t __ret; \
8335  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
8336  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8337  __ret; \
8338})
8339#endif
8340
8341#ifdef __LITTLE_ENDIAN__
8342#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8343  float16x8_t __s1 = __p1; \
8344  float16x8_t __ret; \
8345  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
8346  __ret; \
8347})
8348#else
8349#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8350  float16x8_t __s1 = __p1; \
8351  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8352  float16x8_t __ret; \
8353  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
8354  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8355  __ret; \
8356})
8357#endif
8358
8359#ifdef __LITTLE_ENDIAN__
8360#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8361  int32x4_t __s1 = __p1; \
8362  int32x4_t __ret; \
8363  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
8364  __ret; \
8365})
8366#else
8367#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8368  int32x4_t __s1 = __p1; \
8369  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8370  int32x4_t __ret; \
8371  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
8372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8373  __ret; \
8374})
8375#endif
8376
8377#ifdef __LITTLE_ENDIAN__
8378#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8379  int64x2_t __s1 = __p1; \
8380  int64x2_t __ret; \
8381  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
8382  __ret; \
8383})
8384#else
8385#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8386  int64x2_t __s1 = __p1; \
8387  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8388  int64x2_t __ret; \
8389  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
8390  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8391  __ret; \
8392})
8393#endif
8394
8395#ifdef __LITTLE_ENDIAN__
8396#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8397  int16x8_t __s1 = __p1; \
8398  int16x8_t __ret; \
8399  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
8400  __ret; \
8401})
8402#else
8403#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8404  int16x8_t __s1 = __p1; \
8405  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8406  int16x8_t __ret; \
8407  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
8408  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8409  __ret; \
8410})
8411#endif
8412
8413#ifdef __LITTLE_ENDIAN__
8414#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8415  uint8x8_t __s1 = __p1; \
8416  uint8x8_t __ret; \
8417  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
8418  __ret; \
8419})
8420#else
8421#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8422  uint8x8_t __s1 = __p1; \
8423  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8424  uint8x8_t __ret; \
8425  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
8426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8427  __ret; \
8428})
8429#endif
8430
8431#ifdef __LITTLE_ENDIAN__
8432#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8433  uint32x2_t __s1 = __p1; \
8434  uint32x2_t __ret; \
8435  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
8436  __ret; \
8437})
8438#else
8439#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8440  uint32x2_t __s1 = __p1; \
8441  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8442  uint32x2_t __ret; \
8443  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
8444  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8445  __ret; \
8446})
8447#endif
8448
8449#ifdef __LITTLE_ENDIAN__
8450#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8451  uint64x1_t __s1 = __p1; \
8452  uint64x1_t __ret; \
8453  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8454  __ret; \
8455})
8456#else
8457#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8458  uint64x1_t __s1 = __p1; \
8459  uint64x1_t __ret; \
8460  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8461  __ret; \
8462})
8463#endif
8464
8465#ifdef __LITTLE_ENDIAN__
8466#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8467  uint16x4_t __s1 = __p1; \
8468  uint16x4_t __ret; \
8469  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
8470  __ret; \
8471})
8472#else
8473#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8474  uint16x4_t __s1 = __p1; \
8475  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8476  uint16x4_t __ret; \
8477  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
8478  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8479  __ret; \
8480})
8481#endif
8482
8483#ifdef __LITTLE_ENDIAN__
8484#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8485  int8x8_t __s1 = __p1; \
8486  int8x8_t __ret; \
8487  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
8488  __ret; \
8489})
8490#else
8491#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8492  int8x8_t __s1 = __p1; \
8493  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8494  int8x8_t __ret; \
8495  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
8496  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8497  __ret; \
8498})
8499#endif
8500
8501#ifdef __LITTLE_ENDIAN__
8502#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8503  float32x2_t __s1 = __p1; \
8504  float32x2_t __ret; \
8505  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
8506  __ret; \
8507})
8508#else
8509#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8510  float32x2_t __s1 = __p1; \
8511  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8512  float32x2_t __ret; \
8513  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
8514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8515  __ret; \
8516})
8517#endif
8518
8519#ifdef __LITTLE_ENDIAN__
8520#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8521  float16x4_t __s1 = __p1; \
8522  float16x4_t __ret; \
8523  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
8524  __ret; \
8525})
8526#else
8527#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8528  float16x4_t __s1 = __p1; \
8529  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8530  float16x4_t __ret; \
8531  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
8532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8533  __ret; \
8534})
8535#endif
8536
8537#ifdef __LITTLE_ENDIAN__
8538#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8539  int32x2_t __s1 = __p1; \
8540  int32x2_t __ret; \
8541  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
8542  __ret; \
8543})
8544#else
8545#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8546  int32x2_t __s1 = __p1; \
8547  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8548  int32x2_t __ret; \
8549  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
8550  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8551  __ret; \
8552})
8553#endif
8554
8555#ifdef __LITTLE_ENDIAN__
8556#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8557  int64x1_t __s1 = __p1; \
8558  int64x1_t __ret; \
8559  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8560  __ret; \
8561})
8562#else
8563#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8564  int64x1_t __s1 = __p1; \
8565  int64x1_t __ret; \
8566  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8567  __ret; \
8568})
8569#endif
8570
8571#ifdef __LITTLE_ENDIAN__
8572#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8573  int16x4_t __s1 = __p1; \
8574  int16x4_t __ret; \
8575  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
8576  __ret; \
8577})
8578#else
8579#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8580  int16x4_t __s1 = __p1; \
8581  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8582  int16x4_t __ret; \
8583  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
8584  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8585  __ret; \
8586})
8587#endif
8588
8589#ifdef __LITTLE_ENDIAN__
8590#define vld2_p8(__p0) __extension__ ({ \
8591  poly8x8x2_t __ret; \
8592  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8593  __ret; \
8594})
8595#else
8596#define vld2_p8(__p0) __extension__ ({ \
8597  poly8x8x2_t __ret; \
8598  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8599 \
8600  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8601  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8602  __ret; \
8603})
8604#endif
8605
8606#ifdef __LITTLE_ENDIAN__
8607#define vld2_p16(__p0) __extension__ ({ \
8608  poly16x4x2_t __ret; \
8609  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8610  __ret; \
8611})
8612#else
8613#define vld2_p16(__p0) __extension__ ({ \
8614  poly16x4x2_t __ret; \
8615  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8616 \
8617  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8618  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8619  __ret; \
8620})
8621#endif
8622
8623#ifdef __LITTLE_ENDIAN__
8624#define vld2q_p8(__p0) __extension__ ({ \
8625  poly8x16x2_t __ret; \
8626  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8627  __ret; \
8628})
8629#else
8630#define vld2q_p8(__p0) __extension__ ({ \
8631  poly8x16x2_t __ret; \
8632  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8633 \
8634  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8635  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8636  __ret; \
8637})
8638#endif
8639
8640#ifdef __LITTLE_ENDIAN__
8641#define vld2q_p16(__p0) __extension__ ({ \
8642  poly16x8x2_t __ret; \
8643  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8644  __ret; \
8645})
8646#else
8647#define vld2q_p16(__p0) __extension__ ({ \
8648  poly16x8x2_t __ret; \
8649  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8650 \
8651  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8652  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8653  __ret; \
8654})
8655#endif
8656
8657#ifdef __LITTLE_ENDIAN__
8658#define vld2q_u8(__p0) __extension__ ({ \
8659  uint8x16x2_t __ret; \
8660  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8661  __ret; \
8662})
8663#else
8664#define vld2q_u8(__p0) __extension__ ({ \
8665  uint8x16x2_t __ret; \
8666  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8667 \
8668  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8669  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8670  __ret; \
8671})
8672#endif
8673
8674#ifdef __LITTLE_ENDIAN__
8675#define vld2q_u32(__p0) __extension__ ({ \
8676  uint32x4x2_t __ret; \
8677  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8678  __ret; \
8679})
8680#else
8681#define vld2q_u32(__p0) __extension__ ({ \
8682  uint32x4x2_t __ret; \
8683  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8684 \
8685  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8686  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8687  __ret; \
8688})
8689#endif
8690
8691#ifdef __LITTLE_ENDIAN__
8692#define vld2q_u16(__p0) __extension__ ({ \
8693  uint16x8x2_t __ret; \
8694  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8695  __ret; \
8696})
8697#else
8698#define vld2q_u16(__p0) __extension__ ({ \
8699  uint16x8x2_t __ret; \
8700  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8701 \
8702  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8703  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8704  __ret; \
8705})
8706#endif
8707
8708#ifdef __LITTLE_ENDIAN__
8709#define vld2q_s8(__p0) __extension__ ({ \
8710  int8x16x2_t __ret; \
8711  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8712  __ret; \
8713})
8714#else
8715#define vld2q_s8(__p0) __extension__ ({ \
8716  int8x16x2_t __ret; \
8717  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8718 \
8719  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8720  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8721  __ret; \
8722})
8723#endif
8724
8725#ifdef __LITTLE_ENDIAN__
8726#define vld2q_f32(__p0) __extension__ ({ \
8727  float32x4x2_t __ret; \
8728  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8729  __ret; \
8730})
8731#else
8732#define vld2q_f32(__p0) __extension__ ({ \
8733  float32x4x2_t __ret; \
8734  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8735 \
8736  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8737  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8738  __ret; \
8739})
8740#endif
8741
8742#ifdef __LITTLE_ENDIAN__
8743#define vld2q_f16(__p0) __extension__ ({ \
8744  float16x8x2_t __ret; \
8745  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8746  __ret; \
8747})
8748#else
8749#define vld2q_f16(__p0) __extension__ ({ \
8750  float16x8x2_t __ret; \
8751  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8752 \
8753  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8754  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8755  __ret; \
8756})
8757#endif
8758
8759#ifdef __LITTLE_ENDIAN__
8760#define vld2q_s32(__p0) __extension__ ({ \
8761  int32x4x2_t __ret; \
8762  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8763  __ret; \
8764})
8765#else
8766#define vld2q_s32(__p0) __extension__ ({ \
8767  int32x4x2_t __ret; \
8768  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8769 \
8770  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8771  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8772  __ret; \
8773})
8774#endif
8775
8776#ifdef __LITTLE_ENDIAN__
8777#define vld2q_s16(__p0) __extension__ ({ \
8778  int16x8x2_t __ret; \
8779  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8780  __ret; \
8781})
8782#else
8783#define vld2q_s16(__p0) __extension__ ({ \
8784  int16x8x2_t __ret; \
8785  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8786 \
8787  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8788  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8789  __ret; \
8790})
8791#endif
8792
8793#ifdef __LITTLE_ENDIAN__
8794#define vld2_u8(__p0) __extension__ ({ \
8795  uint8x8x2_t __ret; \
8796  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8797  __ret; \
8798})
8799#else
8800#define vld2_u8(__p0) __extension__ ({ \
8801  uint8x8x2_t __ret; \
8802  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8803 \
8804  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8805  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8806  __ret; \
8807})
8808#endif
8809
8810#ifdef __LITTLE_ENDIAN__
8811#define vld2_u32(__p0) __extension__ ({ \
8812  uint32x2x2_t __ret; \
8813  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8814  __ret; \
8815})
8816#else
8817#define vld2_u32(__p0) __extension__ ({ \
8818  uint32x2x2_t __ret; \
8819  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8820 \
8821  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8822  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8823  __ret; \
8824})
8825#endif
8826
8827#ifdef __LITTLE_ENDIAN__
8828#define vld2_u64(__p0) __extension__ ({ \
8829  uint64x1x2_t __ret; \
8830  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8831  __ret; \
8832})
8833#else
8834#define vld2_u64(__p0) __extension__ ({ \
8835  uint64x1x2_t __ret; \
8836  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8837  __ret; \
8838})
8839#endif
8840
8841#ifdef __LITTLE_ENDIAN__
8842#define vld2_u16(__p0) __extension__ ({ \
8843  uint16x4x2_t __ret; \
8844  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8845  __ret; \
8846})
8847#else
8848#define vld2_u16(__p0) __extension__ ({ \
8849  uint16x4x2_t __ret; \
8850  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8851 \
8852  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8853  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8854  __ret; \
8855})
8856#endif
8857
8858#ifdef __LITTLE_ENDIAN__
8859#define vld2_s8(__p0) __extension__ ({ \
8860  int8x8x2_t __ret; \
8861  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8862  __ret; \
8863})
8864#else
8865#define vld2_s8(__p0) __extension__ ({ \
8866  int8x8x2_t __ret; \
8867  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8868 \
8869  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8870  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8871  __ret; \
8872})
8873#endif
8874
8875#ifdef __LITTLE_ENDIAN__
8876#define vld2_f32(__p0) __extension__ ({ \
8877  float32x2x2_t __ret; \
8878  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8879  __ret; \
8880})
8881#else
8882#define vld2_f32(__p0) __extension__ ({ \
8883  float32x2x2_t __ret; \
8884  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8885 \
8886  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8887  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8888  __ret; \
8889})
8890#endif
8891
8892#ifdef __LITTLE_ENDIAN__
8893#define vld2_f16(__p0) __extension__ ({ \
8894  float16x4x2_t __ret; \
8895  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8896  __ret; \
8897})
8898#else
8899#define vld2_f16(__p0) __extension__ ({ \
8900  float16x4x2_t __ret; \
8901  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8902 \
8903  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8904  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8905  __ret; \
8906})
8907#endif
8908
8909#ifdef __LITTLE_ENDIAN__
8910#define vld2_s32(__p0) __extension__ ({ \
8911  int32x2x2_t __ret; \
8912  __builtin_neon_vld2_v(&__ret, __p0, 2); \
8913  __ret; \
8914})
8915#else
8916#define vld2_s32(__p0) __extension__ ({ \
8917  int32x2x2_t __ret; \
8918  __builtin_neon_vld2_v(&__ret, __p0, 2); \
8919 \
8920  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8921  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8922  __ret; \
8923})
8924#endif
8925
8926#ifdef __LITTLE_ENDIAN__
8927#define vld2_s64(__p0) __extension__ ({ \
8928  int64x1x2_t __ret; \
8929  __builtin_neon_vld2_v(&__ret, __p0, 3); \
8930  __ret; \
8931})
8932#else
8933#define vld2_s64(__p0) __extension__ ({ \
8934  int64x1x2_t __ret; \
8935  __builtin_neon_vld2_v(&__ret, __p0, 3); \
8936  __ret; \
8937})
8938#endif
8939
8940#ifdef __LITTLE_ENDIAN__
8941#define vld2_s16(__p0) __extension__ ({ \
8942  int16x4x2_t __ret; \
8943  __builtin_neon_vld2_v(&__ret, __p0, 1); \
8944  __ret; \
8945})
8946#else
8947#define vld2_s16(__p0) __extension__ ({ \
8948  int16x4x2_t __ret; \
8949  __builtin_neon_vld2_v(&__ret, __p0, 1); \
8950 \
8951  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8952  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8953  __ret; \
8954})
8955#endif
8956
8957#ifdef __LITTLE_ENDIAN__
8958#define vld2_dup_p8(__p0) __extension__ ({ \
8959  poly8x8x2_t __ret; \
8960  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
8961  __ret; \
8962})
8963#else
8964#define vld2_dup_p8(__p0) __extension__ ({ \
8965  poly8x8x2_t __ret; \
8966  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
8967 \
8968  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8969  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8970  __ret; \
8971})
8972#endif
8973
8974#ifdef __LITTLE_ENDIAN__
8975#define vld2_dup_p16(__p0) __extension__ ({ \
8976  poly16x4x2_t __ret; \
8977  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
8978  __ret; \
8979})
8980#else
8981#define vld2_dup_p16(__p0) __extension__ ({ \
8982  poly16x4x2_t __ret; \
8983  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
8984 \
8985  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8986  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8987  __ret; \
8988})
8989#endif
8990
8991#ifdef __LITTLE_ENDIAN__
8992#define vld2_dup_u8(__p0) __extension__ ({ \
8993  uint8x8x2_t __ret; \
8994  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
8995  __ret; \
8996})
8997#else
8998#define vld2_dup_u8(__p0) __extension__ ({ \
8999  uint8x8x2_t __ret; \
9000  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9001 \
9002  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9003  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9004  __ret; \
9005})
9006#endif
9007
9008#ifdef __LITTLE_ENDIAN__
9009#define vld2_dup_u32(__p0) __extension__ ({ \
9010  uint32x2x2_t __ret; \
9011  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9012  __ret; \
9013})
9014#else
9015#define vld2_dup_u32(__p0) __extension__ ({ \
9016  uint32x2x2_t __ret; \
9017  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9018 \
9019  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9020  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9021  __ret; \
9022})
9023#endif
9024
9025#ifdef __LITTLE_ENDIAN__
9026#define vld2_dup_u64(__p0) __extension__ ({ \
9027  uint64x1x2_t __ret; \
9028  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9029  __ret; \
9030})
9031#else
9032#define vld2_dup_u64(__p0) __extension__ ({ \
9033  uint64x1x2_t __ret; \
9034  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9035  __ret; \
9036})
9037#endif
9038
9039#ifdef __LITTLE_ENDIAN__
9040#define vld2_dup_u16(__p0) __extension__ ({ \
9041  uint16x4x2_t __ret; \
9042  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9043  __ret; \
9044})
9045#else
9046#define vld2_dup_u16(__p0) __extension__ ({ \
9047  uint16x4x2_t __ret; \
9048  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9049 \
9050  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9051  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9052  __ret; \
9053})
9054#endif
9055
9056#ifdef __LITTLE_ENDIAN__
9057#define vld2_dup_s8(__p0) __extension__ ({ \
9058  int8x8x2_t __ret; \
9059  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9060  __ret; \
9061})
9062#else
9063#define vld2_dup_s8(__p0) __extension__ ({ \
9064  int8x8x2_t __ret; \
9065  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9066 \
9067  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9068  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9069  __ret; \
9070})
9071#endif
9072
9073#ifdef __LITTLE_ENDIAN__
9074#define vld2_dup_f32(__p0) __extension__ ({ \
9075  float32x2x2_t __ret; \
9076  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9077  __ret; \
9078})
9079#else
9080#define vld2_dup_f32(__p0) __extension__ ({ \
9081  float32x2x2_t __ret; \
9082  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9083 \
9084  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9085  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9086  __ret; \
9087})
9088#endif
9089
9090#ifdef __LITTLE_ENDIAN__
9091#define vld2_dup_f16(__p0) __extension__ ({ \
9092  float16x4x2_t __ret; \
9093  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9094  __ret; \
9095})
9096#else
9097#define vld2_dup_f16(__p0) __extension__ ({ \
9098  float16x4x2_t __ret; \
9099  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9100 \
9101  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9102  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9103  __ret; \
9104})
9105#endif
9106
9107#ifdef __LITTLE_ENDIAN__
9108#define vld2_dup_s32(__p0) __extension__ ({ \
9109  int32x2x2_t __ret; \
9110  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9111  __ret; \
9112})
9113#else
9114#define vld2_dup_s32(__p0) __extension__ ({ \
9115  int32x2x2_t __ret; \
9116  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9117 \
9118  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9119  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9120  __ret; \
9121})
9122#endif
9123
9124#ifdef __LITTLE_ENDIAN__
9125#define vld2_dup_s64(__p0) __extension__ ({ \
9126  int64x1x2_t __ret; \
9127  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9128  __ret; \
9129})
9130#else
9131#define vld2_dup_s64(__p0) __extension__ ({ \
9132  int64x1x2_t __ret; \
9133  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9134  __ret; \
9135})
9136#endif
9137
9138#ifdef __LITTLE_ENDIAN__
9139#define vld2_dup_s16(__p0) __extension__ ({ \
9140  int16x4x2_t __ret; \
9141  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9142  __ret; \
9143})
9144#else
9145#define vld2_dup_s16(__p0) __extension__ ({ \
9146  int16x4x2_t __ret; \
9147  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9148 \
9149  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9150  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9151  __ret; \
9152})
9153#endif
9154
9155#ifdef __LITTLE_ENDIAN__
9156#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9157  poly8x8x2_t __s1 = __p1; \
9158  poly8x8x2_t __ret; \
9159  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
9160  __ret; \
9161})
9162#else
9163#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9164  poly8x8x2_t __s1 = __p1; \
9165  poly8x8x2_t __rev1; \
9166  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9167  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9168  poly8x8x2_t __ret; \
9169  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
9170 \
9171  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9172  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9173  __ret; \
9174})
9175#endif
9176
9177#ifdef __LITTLE_ENDIAN__
9178#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9179  poly16x4x2_t __s1 = __p1; \
9180  poly16x4x2_t __ret; \
9181  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
9182  __ret; \
9183})
9184#else
9185#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9186  poly16x4x2_t __s1 = __p1; \
9187  poly16x4x2_t __rev1; \
9188  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9189  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9190  poly16x4x2_t __ret; \
9191  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
9192 \
9193  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9194  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9195  __ret; \
9196})
9197#endif
9198
9199#ifdef __LITTLE_ENDIAN__
9200#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9201  poly16x8x2_t __s1 = __p1; \
9202  poly16x8x2_t __ret; \
9203  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
9204  __ret; \
9205})
9206#else
9207#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9208  poly16x8x2_t __s1 = __p1; \
9209  poly16x8x2_t __rev1; \
9210  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9211  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9212  poly16x8x2_t __ret; \
9213  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
9214 \
9215  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9216  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9217  __ret; \
9218})
9219#endif
9220
9221#ifdef __LITTLE_ENDIAN__
9222#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9223  uint32x4x2_t __s1 = __p1; \
9224  uint32x4x2_t __ret; \
9225  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
9226  __ret; \
9227})
9228#else
9229#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9230  uint32x4x2_t __s1 = __p1; \
9231  uint32x4x2_t __rev1; \
9232  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9233  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9234  uint32x4x2_t __ret; \
9235  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
9236 \
9237  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9238  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9239  __ret; \
9240})
9241#endif
9242
9243#ifdef __LITTLE_ENDIAN__
9244#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9245  uint16x8x2_t __s1 = __p1; \
9246  uint16x8x2_t __ret; \
9247  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
9248  __ret; \
9249})
9250#else
9251#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9252  uint16x8x2_t __s1 = __p1; \
9253  uint16x8x2_t __rev1; \
9254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9256  uint16x8x2_t __ret; \
9257  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
9258 \
9259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9261  __ret; \
9262})
9263#endif
9264
9265#ifdef __LITTLE_ENDIAN__
9266#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9267  float32x4x2_t __s1 = __p1; \
9268  float32x4x2_t __ret; \
9269  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 41); \
9270  __ret; \
9271})
9272#else
9273#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9274  float32x4x2_t __s1 = __p1; \
9275  float32x4x2_t __rev1; \
9276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9278  float32x4x2_t __ret; \
9279  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
9280 \
9281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9283  __ret; \
9284})
9285#endif
9286
9287#ifdef __LITTLE_ENDIAN__
9288#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9289  float16x8x2_t __s1 = __p1; \
9290  float16x8x2_t __ret; \
9291  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 40); \
9292  __ret; \
9293})
9294#else
9295#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9296  float16x8x2_t __s1 = __p1; \
9297  float16x8x2_t __rev1; \
9298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9300  float16x8x2_t __ret; \
9301  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
9302 \
9303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9305  __ret; \
9306})
9307#endif
9308
9309#ifdef __LITTLE_ENDIAN__
9310#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9311  int32x4x2_t __s1 = __p1; \
9312  int32x4x2_t __ret; \
9313  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 34); \
9314  __ret; \
9315})
9316#else
9317#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9318  int32x4x2_t __s1 = __p1; \
9319  int32x4x2_t __rev1; \
9320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9322  int32x4x2_t __ret; \
9323  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
9324 \
9325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9327  __ret; \
9328})
9329#endif
9330
9331#ifdef __LITTLE_ENDIAN__
9332#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9333  int16x8x2_t __s1 = __p1; \
9334  int16x8x2_t __ret; \
9335  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 33); \
9336  __ret; \
9337})
9338#else
9339#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9340  int16x8x2_t __s1 = __p1; \
9341  int16x8x2_t __rev1; \
9342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9344  int16x8x2_t __ret; \
9345  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
9346 \
9347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9349  __ret; \
9350})
9351#endif
9352
9353#ifdef __LITTLE_ENDIAN__
9354#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9355  uint8x8x2_t __s1 = __p1; \
9356  uint8x8x2_t __ret; \
9357  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
9358  __ret; \
9359})
9360#else
9361#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9362  uint8x8x2_t __s1 = __p1; \
9363  uint8x8x2_t __rev1; \
9364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9366  uint8x8x2_t __ret; \
9367  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
9368 \
9369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9371  __ret; \
9372})
9373#endif
9374
9375#ifdef __LITTLE_ENDIAN__
9376#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9377  uint32x2x2_t __s1 = __p1; \
9378  uint32x2x2_t __ret; \
9379  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
9380  __ret; \
9381})
9382#else
9383#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9384  uint32x2x2_t __s1 = __p1; \
9385  uint32x2x2_t __rev1; \
9386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9388  uint32x2x2_t __ret; \
9389  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
9390 \
9391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9393  __ret; \
9394})
9395#endif
9396
9397#ifdef __LITTLE_ENDIAN__
9398#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9399  uint16x4x2_t __s1 = __p1; \
9400  uint16x4x2_t __ret; \
9401  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
9402  __ret; \
9403})
9404#else
9405#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9406  uint16x4x2_t __s1 = __p1; \
9407  uint16x4x2_t __rev1; \
9408  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9409  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9410  uint16x4x2_t __ret; \
9411  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
9412 \
9413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9415  __ret; \
9416})
9417#endif
9418
9419#ifdef __LITTLE_ENDIAN__
9420#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9421  int8x8x2_t __s1 = __p1; \
9422  int8x8x2_t __ret; \
9423  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
9424  __ret; \
9425})
9426#else
9427#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9428  int8x8x2_t __s1 = __p1; \
9429  int8x8x2_t __rev1; \
9430  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9431  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9432  int8x8x2_t __ret; \
9433  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
9434 \
9435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9437  __ret; \
9438})
9439#endif
9440
9441#ifdef __LITTLE_ENDIAN__
9442#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9443  float32x2x2_t __s1 = __p1; \
9444  float32x2x2_t __ret; \
9445  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 9); \
9446  __ret; \
9447})
9448#else
9449#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9450  float32x2x2_t __s1 = __p1; \
9451  float32x2x2_t __rev1; \
9452  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9453  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9454  float32x2x2_t __ret; \
9455  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
9456 \
9457  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9458  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9459  __ret; \
9460})
9461#endif
9462
9463#ifdef __LITTLE_ENDIAN__
9464#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9465  float16x4x2_t __s1 = __p1; \
9466  float16x4x2_t __ret; \
9467  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 8); \
9468  __ret; \
9469})
9470#else
9471#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9472  float16x4x2_t __s1 = __p1; \
9473  float16x4x2_t __rev1; \
9474  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9475  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9476  float16x4x2_t __ret; \
9477  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
9478 \
9479  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9480  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9481  __ret; \
9482})
9483#endif
9484
9485#ifdef __LITTLE_ENDIAN__
9486#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9487  int32x2x2_t __s1 = __p1; \
9488  int32x2x2_t __ret; \
9489  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 2); \
9490  __ret; \
9491})
9492#else
9493#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9494  int32x2x2_t __s1 = __p1; \
9495  int32x2x2_t __rev1; \
9496  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9497  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9498  int32x2x2_t __ret; \
9499  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
9500 \
9501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9503  __ret; \
9504})
9505#endif
9506
9507#ifdef __LITTLE_ENDIAN__
9508#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9509  int16x4x2_t __s1 = __p1; \
9510  int16x4x2_t __ret; \
9511  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 1); \
9512  __ret; \
9513})
9514#else
9515#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9516  int16x4x2_t __s1 = __p1; \
9517  int16x4x2_t __rev1; \
9518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9520  int16x4x2_t __ret; \
9521  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
9522 \
9523  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9524  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9525  __ret; \
9526})
9527#endif
9528
9529#ifdef __LITTLE_ENDIAN__
9530#define vld3_p8(__p0) __extension__ ({ \
9531  poly8x8x3_t __ret; \
9532  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9533  __ret; \
9534})
9535#else
9536#define vld3_p8(__p0) __extension__ ({ \
9537  poly8x8x3_t __ret; \
9538  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9539 \
9540  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9541  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9542  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9543  __ret; \
9544})
9545#endif
9546
9547#ifdef __LITTLE_ENDIAN__
9548#define vld3_p16(__p0) __extension__ ({ \
9549  poly16x4x3_t __ret; \
9550  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9551  __ret; \
9552})
9553#else
9554#define vld3_p16(__p0) __extension__ ({ \
9555  poly16x4x3_t __ret; \
9556  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9557 \
9558  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9559  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9560  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9561  __ret; \
9562})
9563#endif
9564
9565#ifdef __LITTLE_ENDIAN__
9566#define vld3q_p8(__p0) __extension__ ({ \
9567  poly8x16x3_t __ret; \
9568  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9569  __ret; \
9570})
9571#else
9572#define vld3q_p8(__p0) __extension__ ({ \
9573  poly8x16x3_t __ret; \
9574  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9575 \
9576  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9577  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9578  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9579  __ret; \
9580})
9581#endif
9582
9583#ifdef __LITTLE_ENDIAN__
9584#define vld3q_p16(__p0) __extension__ ({ \
9585  poly16x8x3_t __ret; \
9586  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9587  __ret; \
9588})
9589#else
9590#define vld3q_p16(__p0) __extension__ ({ \
9591  poly16x8x3_t __ret; \
9592  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9593 \
9594  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9595  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9596  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9597  __ret; \
9598})
9599#endif
9600
9601#ifdef __LITTLE_ENDIAN__
9602#define vld3q_u8(__p0) __extension__ ({ \
9603  uint8x16x3_t __ret; \
9604  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9605  __ret; \
9606})
9607#else
9608#define vld3q_u8(__p0) __extension__ ({ \
9609  uint8x16x3_t __ret; \
9610  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9611 \
9612  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9613  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9614  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9615  __ret; \
9616})
9617#endif
9618
9619#ifdef __LITTLE_ENDIAN__
9620#define vld3q_u32(__p0) __extension__ ({ \
9621  uint32x4x3_t __ret; \
9622  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9623  __ret; \
9624})
9625#else
9626#define vld3q_u32(__p0) __extension__ ({ \
9627  uint32x4x3_t __ret; \
9628  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9629 \
9630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9632  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9633  __ret; \
9634})
9635#endif
9636
9637#ifdef __LITTLE_ENDIAN__
9638#define vld3q_u16(__p0) __extension__ ({ \
9639  uint16x8x3_t __ret; \
9640  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9641  __ret; \
9642})
9643#else
9644#define vld3q_u16(__p0) __extension__ ({ \
9645  uint16x8x3_t __ret; \
9646  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9647 \
9648  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9649  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9650  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9651  __ret; \
9652})
9653#endif
9654
9655#ifdef __LITTLE_ENDIAN__
9656#define vld3q_s8(__p0) __extension__ ({ \
9657  int8x16x3_t __ret; \
9658  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9659  __ret; \
9660})
9661#else
9662#define vld3q_s8(__p0) __extension__ ({ \
9663  int8x16x3_t __ret; \
9664  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9665 \
9666  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9667  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9668  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9669  __ret; \
9670})
9671#endif
9672
9673#ifdef __LITTLE_ENDIAN__
9674#define vld3q_f32(__p0) __extension__ ({ \
9675  float32x4x3_t __ret; \
9676  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9677  __ret; \
9678})
9679#else
9680#define vld3q_f32(__p0) __extension__ ({ \
9681  float32x4x3_t __ret; \
9682  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9683 \
9684  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9685  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9686  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9687  __ret; \
9688})
9689#endif
9690
9691#ifdef __LITTLE_ENDIAN__
9692#define vld3q_f16(__p0) __extension__ ({ \
9693  float16x8x3_t __ret; \
9694  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9695  __ret; \
9696})
9697#else
9698#define vld3q_f16(__p0) __extension__ ({ \
9699  float16x8x3_t __ret; \
9700  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9701 \
9702  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9703  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9704  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9705  __ret; \
9706})
9707#endif
9708
9709#ifdef __LITTLE_ENDIAN__
9710#define vld3q_s32(__p0) __extension__ ({ \
9711  int32x4x3_t __ret; \
9712  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9713  __ret; \
9714})
9715#else
9716#define vld3q_s32(__p0) __extension__ ({ \
9717  int32x4x3_t __ret; \
9718  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9719 \
9720  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9721  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9722  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9723  __ret; \
9724})
9725#endif
9726
9727#ifdef __LITTLE_ENDIAN__
9728#define vld3q_s16(__p0) __extension__ ({ \
9729  int16x8x3_t __ret; \
9730  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9731  __ret; \
9732})
9733#else
9734#define vld3q_s16(__p0) __extension__ ({ \
9735  int16x8x3_t __ret; \
9736  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9737 \
9738  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9739  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9740  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9741  __ret; \
9742})
9743#endif
9744
9745#ifdef __LITTLE_ENDIAN__
9746#define vld3_u8(__p0) __extension__ ({ \
9747  uint8x8x3_t __ret; \
9748  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9749  __ret; \
9750})
9751#else
9752#define vld3_u8(__p0) __extension__ ({ \
9753  uint8x8x3_t __ret; \
9754  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9755 \
9756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9758  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9759  __ret; \
9760})
9761#endif
9762
9763#ifdef __LITTLE_ENDIAN__
9764#define vld3_u32(__p0) __extension__ ({ \
9765  uint32x2x3_t __ret; \
9766  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9767  __ret; \
9768})
9769#else
9770#define vld3_u32(__p0) __extension__ ({ \
9771  uint32x2x3_t __ret; \
9772  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9773 \
9774  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9775  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9776  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9777  __ret; \
9778})
9779#endif
9780
9781#ifdef __LITTLE_ENDIAN__
9782#define vld3_u64(__p0) __extension__ ({ \
9783  uint64x1x3_t __ret; \
9784  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9785  __ret; \
9786})
9787#else
9788#define vld3_u64(__p0) __extension__ ({ \
9789  uint64x1x3_t __ret; \
9790  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9791  __ret; \
9792})
9793#endif
9794
9795#ifdef __LITTLE_ENDIAN__
9796#define vld3_u16(__p0) __extension__ ({ \
9797  uint16x4x3_t __ret; \
9798  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9799  __ret; \
9800})
9801#else
9802#define vld3_u16(__p0) __extension__ ({ \
9803  uint16x4x3_t __ret; \
9804  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9805 \
9806  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9807  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9808  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9809  __ret; \
9810})
9811#endif
9812
9813#ifdef __LITTLE_ENDIAN__
9814#define vld3_s8(__p0) __extension__ ({ \
9815  int8x8x3_t __ret; \
9816  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9817  __ret; \
9818})
9819#else
9820#define vld3_s8(__p0) __extension__ ({ \
9821  int8x8x3_t __ret; \
9822  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9823 \
9824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9826  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9827  __ret; \
9828})
9829#endif
9830
9831#ifdef __LITTLE_ENDIAN__
9832#define vld3_f32(__p0) __extension__ ({ \
9833  float32x2x3_t __ret; \
9834  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9835  __ret; \
9836})
9837#else
9838#define vld3_f32(__p0) __extension__ ({ \
9839  float32x2x3_t __ret; \
9840  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9841 \
9842  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9843  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9844  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9845  __ret; \
9846})
9847#endif
9848
9849#ifdef __LITTLE_ENDIAN__
9850#define vld3_f16(__p0) __extension__ ({ \
9851  float16x4x3_t __ret; \
9852  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9853  __ret; \
9854})
9855#else
9856#define vld3_f16(__p0) __extension__ ({ \
9857  float16x4x3_t __ret; \
9858  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9859 \
9860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9863  __ret; \
9864})
9865#endif
9866
9867#ifdef __LITTLE_ENDIAN__
9868#define vld3_s32(__p0) __extension__ ({ \
9869  int32x2x3_t __ret; \
9870  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9871  __ret; \
9872})
9873#else
9874#define vld3_s32(__p0) __extension__ ({ \
9875  int32x2x3_t __ret; \
9876  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9877 \
9878  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9879  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9880  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9881  __ret; \
9882})
9883#endif
9884
9885#ifdef __LITTLE_ENDIAN__
9886#define vld3_s64(__p0) __extension__ ({ \
9887  int64x1x3_t __ret; \
9888  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9889  __ret; \
9890})
9891#else
9892#define vld3_s64(__p0) __extension__ ({ \
9893  int64x1x3_t __ret; \
9894  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9895  __ret; \
9896})
9897#endif
9898
9899#ifdef __LITTLE_ENDIAN__
9900#define vld3_s16(__p0) __extension__ ({ \
9901  int16x4x3_t __ret; \
9902  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9903  __ret; \
9904})
9905#else
9906#define vld3_s16(__p0) __extension__ ({ \
9907  int16x4x3_t __ret; \
9908  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9909 \
9910  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9911  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9912  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9913  __ret; \
9914})
9915#endif
9916
9917#ifdef __LITTLE_ENDIAN__
9918#define vld3_dup_p8(__p0) __extension__ ({ \
9919  poly8x8x3_t __ret; \
9920  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
9921  __ret; \
9922})
9923#else
9924#define vld3_dup_p8(__p0) __extension__ ({ \
9925  poly8x8x3_t __ret; \
9926  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
9927 \
9928  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9929  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9930  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9931  __ret; \
9932})
9933#endif
9934
9935#ifdef __LITTLE_ENDIAN__
9936#define vld3_dup_p16(__p0) __extension__ ({ \
9937  poly16x4x3_t __ret; \
9938  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
9939  __ret; \
9940})
9941#else
9942#define vld3_dup_p16(__p0) __extension__ ({ \
9943  poly16x4x3_t __ret; \
9944  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
9945 \
9946  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9947  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9948  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9949  __ret; \
9950})
9951#endif
9952
9953#ifdef __LITTLE_ENDIAN__
9954#define vld3_dup_u8(__p0) __extension__ ({ \
9955  uint8x8x3_t __ret; \
9956  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
9957  __ret; \
9958})
9959#else
9960#define vld3_dup_u8(__p0) __extension__ ({ \
9961  uint8x8x3_t __ret; \
9962  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
9963 \
9964  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9965  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9966  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9967  __ret; \
9968})
9969#endif
9970
9971#ifdef __LITTLE_ENDIAN__
9972#define vld3_dup_u32(__p0) __extension__ ({ \
9973  uint32x2x3_t __ret; \
9974  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
9975  __ret; \
9976})
9977#else
9978#define vld3_dup_u32(__p0) __extension__ ({ \
9979  uint32x2x3_t __ret; \
9980  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
9981 \
9982  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9983  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9984  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9985  __ret; \
9986})
9987#endif
9988
9989#ifdef __LITTLE_ENDIAN__
9990#define vld3_dup_u64(__p0) __extension__ ({ \
9991  uint64x1x3_t __ret; \
9992  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
9993  __ret; \
9994})
9995#else
9996#define vld3_dup_u64(__p0) __extension__ ({ \
9997  uint64x1x3_t __ret; \
9998  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
9999  __ret; \
10000})
10001#endif
10002
10003#ifdef __LITTLE_ENDIAN__
10004#define vld3_dup_u16(__p0) __extension__ ({ \
10005  uint16x4x3_t __ret; \
10006  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10007  __ret; \
10008})
10009#else
10010#define vld3_dup_u16(__p0) __extension__ ({ \
10011  uint16x4x3_t __ret; \
10012  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10013 \
10014  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10015  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10016  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10017  __ret; \
10018})
10019#endif
10020
10021#ifdef __LITTLE_ENDIAN__
10022#define vld3_dup_s8(__p0) __extension__ ({ \
10023  int8x8x3_t __ret; \
10024  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10025  __ret; \
10026})
10027#else
10028#define vld3_dup_s8(__p0) __extension__ ({ \
10029  int8x8x3_t __ret; \
10030  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10031 \
10032  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10033  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10034  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10035  __ret; \
10036})
10037#endif
10038
10039#ifdef __LITTLE_ENDIAN__
10040#define vld3_dup_f32(__p0) __extension__ ({ \
10041  float32x2x3_t __ret; \
10042  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10043  __ret; \
10044})
10045#else
10046#define vld3_dup_f32(__p0) __extension__ ({ \
10047  float32x2x3_t __ret; \
10048  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10049 \
10050  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10051  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10052  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10053  __ret; \
10054})
10055#endif
10056
10057#ifdef __LITTLE_ENDIAN__
10058#define vld3_dup_f16(__p0) __extension__ ({ \
10059  float16x4x3_t __ret; \
10060  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10061  __ret; \
10062})
10063#else
10064#define vld3_dup_f16(__p0) __extension__ ({ \
10065  float16x4x3_t __ret; \
10066  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10067 \
10068  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10069  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10070  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10071  __ret; \
10072})
10073#endif
10074
10075#ifdef __LITTLE_ENDIAN__
10076#define vld3_dup_s32(__p0) __extension__ ({ \
10077  int32x2x3_t __ret; \
10078  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10079  __ret; \
10080})
10081#else
10082#define vld3_dup_s32(__p0) __extension__ ({ \
10083  int32x2x3_t __ret; \
10084  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10085 \
10086  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10087  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10088  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10089  __ret; \
10090})
10091#endif
10092
10093#ifdef __LITTLE_ENDIAN__
10094#define vld3_dup_s64(__p0) __extension__ ({ \
10095  int64x1x3_t __ret; \
10096  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10097  __ret; \
10098})
10099#else
10100#define vld3_dup_s64(__p0) __extension__ ({ \
10101  int64x1x3_t __ret; \
10102  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10103  __ret; \
10104})
10105#endif
10106
10107#ifdef __LITTLE_ENDIAN__
10108#define vld3_dup_s16(__p0) __extension__ ({ \
10109  int16x4x3_t __ret; \
10110  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10111  __ret; \
10112})
10113#else
10114#define vld3_dup_s16(__p0) __extension__ ({ \
10115  int16x4x3_t __ret; \
10116  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10117 \
10118  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10119  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10120  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10121  __ret; \
10122})
10123#endif
10124
10125#ifdef __LITTLE_ENDIAN__
10126#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10127  poly8x8x3_t __s1 = __p1; \
10128  poly8x8x3_t __ret; \
10129  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
10130  __ret; \
10131})
10132#else
10133#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10134  poly8x8x3_t __s1 = __p1; \
10135  poly8x8x3_t __rev1; \
10136  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10137  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10138  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10139  poly8x8x3_t __ret; \
10140  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
10141 \
10142  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10143  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10144  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10145  __ret; \
10146})
10147#endif
10148
10149#ifdef __LITTLE_ENDIAN__
10150#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10151  poly16x4x3_t __s1 = __p1; \
10152  poly16x4x3_t __ret; \
10153  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
10154  __ret; \
10155})
10156#else
10157#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10158  poly16x4x3_t __s1 = __p1; \
10159  poly16x4x3_t __rev1; \
10160  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10161  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10162  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10163  poly16x4x3_t __ret; \
10164  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
10165 \
10166  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10167  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10168  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10169  __ret; \
10170})
10171#endif
10172
10173#ifdef __LITTLE_ENDIAN__
10174#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10175  poly16x8x3_t __s1 = __p1; \
10176  poly16x8x3_t __ret; \
10177  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
10178  __ret; \
10179})
10180#else
10181#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10182  poly16x8x3_t __s1 = __p1; \
10183  poly16x8x3_t __rev1; \
10184  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10185  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10186  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10187  poly16x8x3_t __ret; \
10188  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
10189 \
10190  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10191  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10192  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10193  __ret; \
10194})
10195#endif
10196
10197#ifdef __LITTLE_ENDIAN__
10198#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10199  uint32x4x3_t __s1 = __p1; \
10200  uint32x4x3_t __ret; \
10201  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
10202  __ret; \
10203})
10204#else
10205#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10206  uint32x4x3_t __s1 = __p1; \
10207  uint32x4x3_t __rev1; \
10208  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10209  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10210  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10211  uint32x4x3_t __ret; \
10212  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
10213 \
10214  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10215  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10216  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10217  __ret; \
10218})
10219#endif
10220
10221#ifdef __LITTLE_ENDIAN__
10222#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10223  uint16x8x3_t __s1 = __p1; \
10224  uint16x8x3_t __ret; \
10225  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
10226  __ret; \
10227})
10228#else
10229#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10230  uint16x8x3_t __s1 = __p1; \
10231  uint16x8x3_t __rev1; \
10232  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10233  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10234  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10235  uint16x8x3_t __ret; \
10236  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
10237 \
10238  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10239  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10240  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10241  __ret; \
10242})
10243#endif
10244
10245#ifdef __LITTLE_ENDIAN__
10246#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10247  float32x4x3_t __s1 = __p1; \
10248  float32x4x3_t __ret; \
10249  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
10250  __ret; \
10251})
10252#else
10253#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10254  float32x4x3_t __s1 = __p1; \
10255  float32x4x3_t __rev1; \
10256  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10257  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10258  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10259  float32x4x3_t __ret; \
10260  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
10261 \
10262  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10263  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10264  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10265  __ret; \
10266})
10267#endif
10268
10269#ifdef __LITTLE_ENDIAN__
10270#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10271  float16x8x3_t __s1 = __p1; \
10272  float16x8x3_t __ret; \
10273  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
10274  __ret; \
10275})
10276#else
10277#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10278  float16x8x3_t __s1 = __p1; \
10279  float16x8x3_t __rev1; \
10280  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10281  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10282  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10283  float16x8x3_t __ret; \
10284  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
10285 \
10286  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10287  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10288  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10289  __ret; \
10290})
10291#endif
10292
10293#ifdef __LITTLE_ENDIAN__
10294#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10295  int32x4x3_t __s1 = __p1; \
10296  int32x4x3_t __ret; \
10297  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
10298  __ret; \
10299})
10300#else
10301#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10302  int32x4x3_t __s1 = __p1; \
10303  int32x4x3_t __rev1; \
10304  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10305  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10306  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10307  int32x4x3_t __ret; \
10308  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
10309 \
10310  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10311  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10312  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10313  __ret; \
10314})
10315#endif
10316
10317#ifdef __LITTLE_ENDIAN__
10318#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10319  int16x8x3_t __s1 = __p1; \
10320  int16x8x3_t __ret; \
10321  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
10322  __ret; \
10323})
10324#else
10325#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10326  int16x8x3_t __s1 = __p1; \
10327  int16x8x3_t __rev1; \
10328  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10329  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10330  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10331  int16x8x3_t __ret; \
10332  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
10333 \
10334  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10335  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10336  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10337  __ret; \
10338})
10339#endif
10340
10341#ifdef __LITTLE_ENDIAN__
10342#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10343  uint8x8x3_t __s1 = __p1; \
10344  uint8x8x3_t __ret; \
10345  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
10346  __ret; \
10347})
10348#else
10349#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10350  uint8x8x3_t __s1 = __p1; \
10351  uint8x8x3_t __rev1; \
10352  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10353  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10354  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10355  uint8x8x3_t __ret; \
10356  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
10357 \
10358  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10359  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10360  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10361  __ret; \
10362})
10363#endif
10364
10365#ifdef __LITTLE_ENDIAN__
10366#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10367  uint32x2x3_t __s1 = __p1; \
10368  uint32x2x3_t __ret; \
10369  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
10370  __ret; \
10371})
10372#else
10373#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10374  uint32x2x3_t __s1 = __p1; \
10375  uint32x2x3_t __rev1; \
10376  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10377  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10378  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10379  uint32x2x3_t __ret; \
10380  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
10381 \
10382  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10383  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10384  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10385  __ret; \
10386})
10387#endif
10388
10389#ifdef __LITTLE_ENDIAN__
10390#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10391  uint16x4x3_t __s1 = __p1; \
10392  uint16x4x3_t __ret; \
10393  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
10394  __ret; \
10395})
10396#else
10397#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10398  uint16x4x3_t __s1 = __p1; \
10399  uint16x4x3_t __rev1; \
10400  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10401  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10402  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10403  uint16x4x3_t __ret; \
10404  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
10405 \
10406  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10407  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10408  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10409  __ret; \
10410})
10411#endif
10412
10413#ifdef __LITTLE_ENDIAN__
10414#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10415  int8x8x3_t __s1 = __p1; \
10416  int8x8x3_t __ret; \
10417  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
10418  __ret; \
10419})
10420#else
10421#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10422  int8x8x3_t __s1 = __p1; \
10423  int8x8x3_t __rev1; \
10424  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10425  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10426  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10427  int8x8x3_t __ret; \
10428  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
10429 \
10430  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10431  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10432  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10433  __ret; \
10434})
10435#endif
10436
10437#ifdef __LITTLE_ENDIAN__
10438#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10439  float32x2x3_t __s1 = __p1; \
10440  float32x2x3_t __ret; \
10441  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
10442  __ret; \
10443})
10444#else
10445#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10446  float32x2x3_t __s1 = __p1; \
10447  float32x2x3_t __rev1; \
10448  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10449  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10450  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10451  float32x2x3_t __ret; \
10452  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
10453 \
10454  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10455  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10456  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10457  __ret; \
10458})
10459#endif
10460
10461#ifdef __LITTLE_ENDIAN__
10462#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10463  float16x4x3_t __s1 = __p1; \
10464  float16x4x3_t __ret; \
10465  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
10466  __ret; \
10467})
10468#else
10469#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10470  float16x4x3_t __s1 = __p1; \
10471  float16x4x3_t __rev1; \
10472  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10473  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10474  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10475  float16x4x3_t __ret; \
10476  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
10477 \
10478  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10479  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10480  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10481  __ret; \
10482})
10483#endif
10484
10485#ifdef __LITTLE_ENDIAN__
10486#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10487  int32x2x3_t __s1 = __p1; \
10488  int32x2x3_t __ret; \
10489  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
10490  __ret; \
10491})
10492#else
10493#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10494  int32x2x3_t __s1 = __p1; \
10495  int32x2x3_t __rev1; \
10496  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10497  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10498  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10499  int32x2x3_t __ret; \
10500  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
10501 \
10502  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10503  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10504  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10505  __ret; \
10506})
10507#endif
10508
10509#ifdef __LITTLE_ENDIAN__
10510#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10511  int16x4x3_t __s1 = __p1; \
10512  int16x4x3_t __ret; \
10513  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
10514  __ret; \
10515})
10516#else
10517#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10518  int16x4x3_t __s1 = __p1; \
10519  int16x4x3_t __rev1; \
10520  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10521  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10522  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10523  int16x4x3_t __ret; \
10524  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
10525 \
10526  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10527  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10528  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10529  __ret; \
10530})
10531#endif
10532
10533#ifdef __LITTLE_ENDIAN__
10534#define vld4_p8(__p0) __extension__ ({ \
10535  poly8x8x4_t __ret; \
10536  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10537  __ret; \
10538})
10539#else
10540#define vld4_p8(__p0) __extension__ ({ \
10541  poly8x8x4_t __ret; \
10542  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10543 \
10544  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10545  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10546  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10547  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10548  __ret; \
10549})
10550#endif
10551
10552#ifdef __LITTLE_ENDIAN__
10553#define vld4_p16(__p0) __extension__ ({ \
10554  poly16x4x4_t __ret; \
10555  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10556  __ret; \
10557})
10558#else
10559#define vld4_p16(__p0) __extension__ ({ \
10560  poly16x4x4_t __ret; \
10561  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10562 \
10563  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10564  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10565  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10566  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10567  __ret; \
10568})
10569#endif
10570
10571#ifdef __LITTLE_ENDIAN__
10572#define vld4q_p8(__p0) __extension__ ({ \
10573  poly8x16x4_t __ret; \
10574  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10575  __ret; \
10576})
10577#else
10578#define vld4q_p8(__p0) __extension__ ({ \
10579  poly8x16x4_t __ret; \
10580  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10581 \
10582  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10583  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10584  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10585  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10586  __ret; \
10587})
10588#endif
10589
10590#ifdef __LITTLE_ENDIAN__
10591#define vld4q_p16(__p0) __extension__ ({ \
10592  poly16x8x4_t __ret; \
10593  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10594  __ret; \
10595})
10596#else
10597#define vld4q_p16(__p0) __extension__ ({ \
10598  poly16x8x4_t __ret; \
10599  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10600 \
10601  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10602  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10603  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10604  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10605  __ret; \
10606})
10607#endif
10608
10609#ifdef __LITTLE_ENDIAN__
10610#define vld4q_u8(__p0) __extension__ ({ \
10611  uint8x16x4_t __ret; \
10612  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10613  __ret; \
10614})
10615#else
10616#define vld4q_u8(__p0) __extension__ ({ \
10617  uint8x16x4_t __ret; \
10618  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10619 \
10620  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10621  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10622  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10623  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10624  __ret; \
10625})
10626#endif
10627
10628#ifdef __LITTLE_ENDIAN__
10629#define vld4q_u32(__p0) __extension__ ({ \
10630  uint32x4x4_t __ret; \
10631  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10632  __ret; \
10633})
10634#else
10635#define vld4q_u32(__p0) __extension__ ({ \
10636  uint32x4x4_t __ret; \
10637  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10638 \
10639  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10640  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10641  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10642  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10643  __ret; \
10644})
10645#endif
10646
10647#ifdef __LITTLE_ENDIAN__
10648#define vld4q_u16(__p0) __extension__ ({ \
10649  uint16x8x4_t __ret; \
10650  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10651  __ret; \
10652})
10653#else
10654#define vld4q_u16(__p0) __extension__ ({ \
10655  uint16x8x4_t __ret; \
10656  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10657 \
10658  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10659  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10660  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10661  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10662  __ret; \
10663})
10664#endif
10665
10666#ifdef __LITTLE_ENDIAN__
10667#define vld4q_s8(__p0) __extension__ ({ \
10668  int8x16x4_t __ret; \
10669  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10670  __ret; \
10671})
10672#else
10673#define vld4q_s8(__p0) __extension__ ({ \
10674  int8x16x4_t __ret; \
10675  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10676 \
10677  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10678  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10679  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10680  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10681  __ret; \
10682})
10683#endif
10684
10685#ifdef __LITTLE_ENDIAN__
10686#define vld4q_f32(__p0) __extension__ ({ \
10687  float32x4x4_t __ret; \
10688  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10689  __ret; \
10690})
10691#else
10692#define vld4q_f32(__p0) __extension__ ({ \
10693  float32x4x4_t __ret; \
10694  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10695 \
10696  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10697  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10698  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10699  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10700  __ret; \
10701})
10702#endif
10703
10704#ifdef __LITTLE_ENDIAN__
10705#define vld4q_f16(__p0) __extension__ ({ \
10706  float16x8x4_t __ret; \
10707  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10708  __ret; \
10709})
10710#else
10711#define vld4q_f16(__p0) __extension__ ({ \
10712  float16x8x4_t __ret; \
10713  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10714 \
10715  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10716  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10717  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10718  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10719  __ret; \
10720})
10721#endif
10722
10723#ifdef __LITTLE_ENDIAN__
10724#define vld4q_s32(__p0) __extension__ ({ \
10725  int32x4x4_t __ret; \
10726  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10727  __ret; \
10728})
10729#else
10730#define vld4q_s32(__p0) __extension__ ({ \
10731  int32x4x4_t __ret; \
10732  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10733 \
10734  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10735  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10736  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10737  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10738  __ret; \
10739})
10740#endif
10741
10742#ifdef __LITTLE_ENDIAN__
10743#define vld4q_s16(__p0) __extension__ ({ \
10744  int16x8x4_t __ret; \
10745  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10746  __ret; \
10747})
10748#else
10749#define vld4q_s16(__p0) __extension__ ({ \
10750  int16x8x4_t __ret; \
10751  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10752 \
10753  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10754  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10755  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10756  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10757  __ret; \
10758})
10759#endif
10760
10761#ifdef __LITTLE_ENDIAN__
10762#define vld4_u8(__p0) __extension__ ({ \
10763  uint8x8x4_t __ret; \
10764  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10765  __ret; \
10766})
10767#else
10768#define vld4_u8(__p0) __extension__ ({ \
10769  uint8x8x4_t __ret; \
10770  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10771 \
10772  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10773  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10774  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10775  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10776  __ret; \
10777})
10778#endif
10779
10780#ifdef __LITTLE_ENDIAN__
10781#define vld4_u32(__p0) __extension__ ({ \
10782  uint32x2x4_t __ret; \
10783  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10784  __ret; \
10785})
10786#else
10787#define vld4_u32(__p0) __extension__ ({ \
10788  uint32x2x4_t __ret; \
10789  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10790 \
10791  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10792  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10793  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10794  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10795  __ret; \
10796})
10797#endif
10798
10799#ifdef __LITTLE_ENDIAN__
10800#define vld4_u64(__p0) __extension__ ({ \
10801  uint64x1x4_t __ret; \
10802  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10803  __ret; \
10804})
10805#else
10806#define vld4_u64(__p0) __extension__ ({ \
10807  uint64x1x4_t __ret; \
10808  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10809  __ret; \
10810})
10811#endif
10812
10813#ifdef __LITTLE_ENDIAN__
10814#define vld4_u16(__p0) __extension__ ({ \
10815  uint16x4x4_t __ret; \
10816  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10817  __ret; \
10818})
10819#else
10820#define vld4_u16(__p0) __extension__ ({ \
10821  uint16x4x4_t __ret; \
10822  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10823 \
10824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10826  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10827  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10828  __ret; \
10829})
10830#endif
10831
10832#ifdef __LITTLE_ENDIAN__
10833#define vld4_s8(__p0) __extension__ ({ \
10834  int8x8x4_t __ret; \
10835  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10836  __ret; \
10837})
10838#else
10839#define vld4_s8(__p0) __extension__ ({ \
10840  int8x8x4_t __ret; \
10841  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10842 \
10843  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10844  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10845  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10846  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10847  __ret; \
10848})
10849#endif
10850
10851#ifdef __LITTLE_ENDIAN__
10852#define vld4_f32(__p0) __extension__ ({ \
10853  float32x2x4_t __ret; \
10854  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10855  __ret; \
10856})
10857#else
10858#define vld4_f32(__p0) __extension__ ({ \
10859  float32x2x4_t __ret; \
10860  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10861 \
10862  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10863  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10864  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10865  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10866  __ret; \
10867})
10868#endif
10869
10870#ifdef __LITTLE_ENDIAN__
10871#define vld4_f16(__p0) __extension__ ({ \
10872  float16x4x4_t __ret; \
10873  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10874  __ret; \
10875})
10876#else
10877#define vld4_f16(__p0) __extension__ ({ \
10878  float16x4x4_t __ret; \
10879  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10880 \
10881  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10882  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10883  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10884  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10885  __ret; \
10886})
10887#endif
10888
10889#ifdef __LITTLE_ENDIAN__
10890#define vld4_s32(__p0) __extension__ ({ \
10891  int32x2x4_t __ret; \
10892  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10893  __ret; \
10894})
10895#else
10896#define vld4_s32(__p0) __extension__ ({ \
10897  int32x2x4_t __ret; \
10898  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10899 \
10900  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10901  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10902  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10903  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10904  __ret; \
10905})
10906#endif
10907
10908#ifdef __LITTLE_ENDIAN__
10909#define vld4_s64(__p0) __extension__ ({ \
10910  int64x1x4_t __ret; \
10911  __builtin_neon_vld4_v(&__ret, __p0, 3); \
10912  __ret; \
10913})
10914#else
10915#define vld4_s64(__p0) __extension__ ({ \
10916  int64x1x4_t __ret; \
10917  __builtin_neon_vld4_v(&__ret, __p0, 3); \
10918  __ret; \
10919})
10920#endif
10921
10922#ifdef __LITTLE_ENDIAN__
10923#define vld4_s16(__p0) __extension__ ({ \
10924  int16x4x4_t __ret; \
10925  __builtin_neon_vld4_v(&__ret, __p0, 1); \
10926  __ret; \
10927})
10928#else
10929#define vld4_s16(__p0) __extension__ ({ \
10930  int16x4x4_t __ret; \
10931  __builtin_neon_vld4_v(&__ret, __p0, 1); \
10932 \
10933  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10934  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10935  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10936  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10937  __ret; \
10938})
10939#endif
10940
10941#ifdef __LITTLE_ENDIAN__
10942#define vld4_dup_p8(__p0) __extension__ ({ \
10943  poly8x8x4_t __ret; \
10944  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
10945  __ret; \
10946})
10947#else
10948#define vld4_dup_p8(__p0) __extension__ ({ \
10949  poly8x8x4_t __ret; \
10950  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
10951 \
10952  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10953  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10954  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10955  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10956  __ret; \
10957})
10958#endif
10959
10960#ifdef __LITTLE_ENDIAN__
10961#define vld4_dup_p16(__p0) __extension__ ({ \
10962  poly16x4x4_t __ret; \
10963  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
10964  __ret; \
10965})
10966#else
10967#define vld4_dup_p16(__p0) __extension__ ({ \
10968  poly16x4x4_t __ret; \
10969  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
10970 \
10971  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10972  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10973  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10974  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10975  __ret; \
10976})
10977#endif
10978
10979#ifdef __LITTLE_ENDIAN__
10980#define vld4_dup_u8(__p0) __extension__ ({ \
10981  uint8x8x4_t __ret; \
10982  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
10983  __ret; \
10984})
10985#else
10986#define vld4_dup_u8(__p0) __extension__ ({ \
10987  uint8x8x4_t __ret; \
10988  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
10989 \
10990  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10991  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10992  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10993  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10994  __ret; \
10995})
10996#endif
10997
10998#ifdef __LITTLE_ENDIAN__
10999#define vld4_dup_u32(__p0) __extension__ ({ \
11000  uint32x2x4_t __ret; \
11001  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11002  __ret; \
11003})
11004#else
11005#define vld4_dup_u32(__p0) __extension__ ({ \
11006  uint32x2x4_t __ret; \
11007  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11008 \
11009  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11010  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11011  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11012  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11013  __ret; \
11014})
11015#endif
11016
11017#ifdef __LITTLE_ENDIAN__
11018#define vld4_dup_u64(__p0) __extension__ ({ \
11019  uint64x1x4_t __ret; \
11020  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11021  __ret; \
11022})
11023#else
11024#define vld4_dup_u64(__p0) __extension__ ({ \
11025  uint64x1x4_t __ret; \
11026  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11027  __ret; \
11028})
11029#endif
11030
11031#ifdef __LITTLE_ENDIAN__
11032#define vld4_dup_u16(__p0) __extension__ ({ \
11033  uint16x4x4_t __ret; \
11034  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11035  __ret; \
11036})
11037#else
11038#define vld4_dup_u16(__p0) __extension__ ({ \
11039  uint16x4x4_t __ret; \
11040  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11041 \
11042  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11043  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11044  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11045  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11046  __ret; \
11047})
11048#endif
11049
11050#ifdef __LITTLE_ENDIAN__
11051#define vld4_dup_s8(__p0) __extension__ ({ \
11052  int8x8x4_t __ret; \
11053  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11054  __ret; \
11055})
11056#else
11057#define vld4_dup_s8(__p0) __extension__ ({ \
11058  int8x8x4_t __ret; \
11059  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11060 \
11061  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11062  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11063  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11064  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11065  __ret; \
11066})
11067#endif
11068
11069#ifdef __LITTLE_ENDIAN__
11070#define vld4_dup_f32(__p0) __extension__ ({ \
11071  float32x2x4_t __ret; \
11072  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11073  __ret; \
11074})
11075#else
11076#define vld4_dup_f32(__p0) __extension__ ({ \
11077  float32x2x4_t __ret; \
11078  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11079 \
11080  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11081  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11082  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11083  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11084  __ret; \
11085})
11086#endif
11087
11088#ifdef __LITTLE_ENDIAN__
11089#define vld4_dup_f16(__p0) __extension__ ({ \
11090  float16x4x4_t __ret; \
11091  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11092  __ret; \
11093})
11094#else
11095#define vld4_dup_f16(__p0) __extension__ ({ \
11096  float16x4x4_t __ret; \
11097  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11098 \
11099  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11100  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11101  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11102  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11103  __ret; \
11104})
11105#endif
11106
11107#ifdef __LITTLE_ENDIAN__
11108#define vld4_dup_s32(__p0) __extension__ ({ \
11109  int32x2x4_t __ret; \
11110  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11111  __ret; \
11112})
11113#else
11114#define vld4_dup_s32(__p0) __extension__ ({ \
11115  int32x2x4_t __ret; \
11116  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11117 \
11118  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11119  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11120  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11121  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11122  __ret; \
11123})
11124#endif
11125
11126#ifdef __LITTLE_ENDIAN__
11127#define vld4_dup_s64(__p0) __extension__ ({ \
11128  int64x1x4_t __ret; \
11129  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11130  __ret; \
11131})
11132#else
11133#define vld4_dup_s64(__p0) __extension__ ({ \
11134  int64x1x4_t __ret; \
11135  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11136  __ret; \
11137})
11138#endif
11139
11140#ifdef __LITTLE_ENDIAN__
11141#define vld4_dup_s16(__p0) __extension__ ({ \
11142  int16x4x4_t __ret; \
11143  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11144  __ret; \
11145})
11146#else
11147#define vld4_dup_s16(__p0) __extension__ ({ \
11148  int16x4x4_t __ret; \
11149  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11150 \
11151  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11152  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11153  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11154  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11155  __ret; \
11156})
11157#endif
11158
11159#ifdef __LITTLE_ENDIAN__
11160#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11161  poly8x8x4_t __s1 = __p1; \
11162  poly8x8x4_t __ret; \
11163  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
11164  __ret; \
11165})
11166#else
11167#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11168  poly8x8x4_t __s1 = __p1; \
11169  poly8x8x4_t __rev1; \
11170  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11171  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11172  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11173  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11174  poly8x8x4_t __ret; \
11175  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
11176 \
11177  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11178  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11179  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11180  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11181  __ret; \
11182})
11183#endif
11184
11185#ifdef __LITTLE_ENDIAN__
11186#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11187  poly16x4x4_t __s1 = __p1; \
11188  poly16x4x4_t __ret; \
11189  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
11190  __ret; \
11191})
11192#else
11193#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11194  poly16x4x4_t __s1 = __p1; \
11195  poly16x4x4_t __rev1; \
11196  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11197  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11198  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11199  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11200  poly16x4x4_t __ret; \
11201  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
11202 \
11203  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11204  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11205  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11206  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11207  __ret; \
11208})
11209#endif
11210
11211#ifdef __LITTLE_ENDIAN__
11212#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11213  poly16x8x4_t __s1 = __p1; \
11214  poly16x8x4_t __ret; \
11215  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
11216  __ret; \
11217})
11218#else
11219#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11220  poly16x8x4_t __s1 = __p1; \
11221  poly16x8x4_t __rev1; \
11222  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11223  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11224  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11225  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11226  poly16x8x4_t __ret; \
11227  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
11228 \
11229  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11230  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11231  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11232  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11233  __ret; \
11234})
11235#endif
11236
11237#ifdef __LITTLE_ENDIAN__
11238#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11239  uint32x4x4_t __s1 = __p1; \
11240  uint32x4x4_t __ret; \
11241  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
11242  __ret; \
11243})
11244#else
11245#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11246  uint32x4x4_t __s1 = __p1; \
11247  uint32x4x4_t __rev1; \
11248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11251  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11252  uint32x4x4_t __ret; \
11253  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
11254 \
11255  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11256  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11257  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11258  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11259  __ret; \
11260})
11261#endif
11262
11263#ifdef __LITTLE_ENDIAN__
11264#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11265  uint16x8x4_t __s1 = __p1; \
11266  uint16x8x4_t __ret; \
11267  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
11268  __ret; \
11269})
11270#else
11271#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11272  uint16x8x4_t __s1 = __p1; \
11273  uint16x8x4_t __rev1; \
11274  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11275  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11276  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11277  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11278  uint16x8x4_t __ret; \
11279  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
11280 \
11281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11283  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11284  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11285  __ret; \
11286})
11287#endif
11288
11289#ifdef __LITTLE_ENDIAN__
11290#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11291  float32x4x4_t __s1 = __p1; \
11292  float32x4x4_t __ret; \
11293  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
11294  __ret; \
11295})
11296#else
11297#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11298  float32x4x4_t __s1 = __p1; \
11299  float32x4x4_t __rev1; \
11300  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11301  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11302  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11303  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11304  float32x4x4_t __ret; \
11305  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
11306 \
11307  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11308  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11309  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11310  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11311  __ret; \
11312})
11313#endif
11314
11315#ifdef __LITTLE_ENDIAN__
11316#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11317  float16x8x4_t __s1 = __p1; \
11318  float16x8x4_t __ret; \
11319  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
11320  __ret; \
11321})
11322#else
11323#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11324  float16x8x4_t __s1 = __p1; \
11325  float16x8x4_t __rev1; \
11326  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11327  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11328  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11329  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11330  float16x8x4_t __ret; \
11331  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
11332 \
11333  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11334  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11335  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11336  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11337  __ret; \
11338})
11339#endif
11340
11341#ifdef __LITTLE_ENDIAN__
11342#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11343  int32x4x4_t __s1 = __p1; \
11344  int32x4x4_t __ret; \
11345  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
11346  __ret; \
11347})
11348#else
11349#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11350  int32x4x4_t __s1 = __p1; \
11351  int32x4x4_t __rev1; \
11352  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11353  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11354  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11355  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11356  int32x4x4_t __ret; \
11357  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
11358 \
11359  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11360  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11361  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11362  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11363  __ret; \
11364})
11365#endif
11366
11367#ifdef __LITTLE_ENDIAN__
11368#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11369  int16x8x4_t __s1 = __p1; \
11370  int16x8x4_t __ret; \
11371  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
11372  __ret; \
11373})
11374#else
11375#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11376  int16x8x4_t __s1 = __p1; \
11377  int16x8x4_t __rev1; \
11378  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11379  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11380  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11381  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11382  int16x8x4_t __ret; \
11383  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
11384 \
11385  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11386  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11387  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11388  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11389  __ret; \
11390})
11391#endif
11392
11393#ifdef __LITTLE_ENDIAN__
11394#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11395  uint8x8x4_t __s1 = __p1; \
11396  uint8x8x4_t __ret; \
11397  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
11398  __ret; \
11399})
11400#else
11401#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11402  uint8x8x4_t __s1 = __p1; \
11403  uint8x8x4_t __rev1; \
11404  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11405  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11406  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11407  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11408  uint8x8x4_t __ret; \
11409  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
11410 \
11411  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11412  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11413  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11414  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11415  __ret; \
11416})
11417#endif
11418
11419#ifdef __LITTLE_ENDIAN__
11420#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11421  uint32x2x4_t __s1 = __p1; \
11422  uint32x2x4_t __ret; \
11423  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
11424  __ret; \
11425})
11426#else
11427#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11428  uint32x2x4_t __s1 = __p1; \
11429  uint32x2x4_t __rev1; \
11430  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11431  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11432  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11433  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11434  uint32x2x4_t __ret; \
11435  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
11436 \
11437  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11438  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11439  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11440  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11441  __ret; \
11442})
11443#endif
11444
11445#ifdef __LITTLE_ENDIAN__
11446#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11447  uint16x4x4_t __s1 = __p1; \
11448  uint16x4x4_t __ret; \
11449  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
11450  __ret; \
11451})
11452#else
11453#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11454  uint16x4x4_t __s1 = __p1; \
11455  uint16x4x4_t __rev1; \
11456  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11457  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11458  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11459  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11460  uint16x4x4_t __ret; \
11461  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
11462 \
11463  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11464  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11465  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11466  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11467  __ret; \
11468})
11469#endif
11470
11471#ifdef __LITTLE_ENDIAN__
11472#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11473  int8x8x4_t __s1 = __p1; \
11474  int8x8x4_t __ret; \
11475  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
11476  __ret; \
11477})
11478#else
11479#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11480  int8x8x4_t __s1 = __p1; \
11481  int8x8x4_t __rev1; \
11482  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11483  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11484  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11485  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11486  int8x8x4_t __ret; \
11487  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
11488 \
11489  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11490  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11491  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11492  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11493  __ret; \
11494})
11495#endif
11496
11497#ifdef __LITTLE_ENDIAN__
11498#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11499  float32x2x4_t __s1 = __p1; \
11500  float32x2x4_t __ret; \
11501  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
11502  __ret; \
11503})
11504#else
11505#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11506  float32x2x4_t __s1 = __p1; \
11507  float32x2x4_t __rev1; \
11508  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11509  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11510  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11511  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11512  float32x2x4_t __ret; \
11513  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
11514 \
11515  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11516  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11517  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11518  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11519  __ret; \
11520})
11521#endif
11522
11523#ifdef __LITTLE_ENDIAN__
11524#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11525  float16x4x4_t __s1 = __p1; \
11526  float16x4x4_t __ret; \
11527  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
11528  __ret; \
11529})
11530#else
11531#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11532  float16x4x4_t __s1 = __p1; \
11533  float16x4x4_t __rev1; \
11534  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11535  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11536  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11537  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11538  float16x4x4_t __ret; \
11539  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
11540 \
11541  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11542  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11543  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11544  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11545  __ret; \
11546})
11547#endif
11548
11549#ifdef __LITTLE_ENDIAN__
11550#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11551  int32x2x4_t __s1 = __p1; \
11552  int32x2x4_t __ret; \
11553  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
11554  __ret; \
11555})
11556#else
11557#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11558  int32x2x4_t __s1 = __p1; \
11559  int32x2x4_t __rev1; \
11560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11563  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11564  int32x2x4_t __ret; \
11565  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
11566 \
11567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11569  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11570  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11571  __ret; \
11572})
11573#endif
11574
11575#ifdef __LITTLE_ENDIAN__
11576#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11577  int16x4x4_t __s1 = __p1; \
11578  int16x4x4_t __ret; \
11579  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
11580  __ret; \
11581})
11582#else
11583#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11584  int16x4x4_t __s1 = __p1; \
11585  int16x4x4_t __rev1; \
11586  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11587  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11588  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11589  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11590  int16x4x4_t __ret; \
11591  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
11592 \
11593  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11594  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11595  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11596  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11597  __ret; \
11598})
11599#endif
11600
11601#ifdef __LITTLE_ENDIAN__
11602__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11603  uint8x16_t __ret;
11604  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11605  return __ret;
11606}
11607#else
11608__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11609  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11610  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11611  uint8x16_t __ret;
11612  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11613  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11614  return __ret;
11615}
11616#endif
11617
11618#ifdef __LITTLE_ENDIAN__
11619__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11620  uint32x4_t __ret;
11621  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11622  return __ret;
11623}
11624#else
11625__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11626  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11627  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11628  uint32x4_t __ret;
11629  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11630  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11631  return __ret;
11632}
11633#endif
11634
11635#ifdef __LITTLE_ENDIAN__
11636__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11637  uint16x8_t __ret;
11638  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11639  return __ret;
11640}
11641#else
11642__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11643  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11644  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11645  uint16x8_t __ret;
11646  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11647  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11648  return __ret;
11649}
11650#endif
11651
11652#ifdef __LITTLE_ENDIAN__
11653__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11654  int8x16_t __ret;
11655  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11656  return __ret;
11657}
11658#else
11659__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11660  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11661  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11662  int8x16_t __ret;
11663  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11664  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11665  return __ret;
11666}
11667#endif
11668
11669#ifdef __LITTLE_ENDIAN__
11670__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11671  float32x4_t __ret;
11672  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11673  return __ret;
11674}
11675#else
11676__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11677  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11678  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11679  float32x4_t __ret;
11680  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
11681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11682  return __ret;
11683}
11684#endif
11685
11686#ifdef __LITTLE_ENDIAN__
11687__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11688  int32x4_t __ret;
11689  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
11690  return __ret;
11691}
11692#else
11693__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11694  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11695  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11696  int32x4_t __ret;
11697  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
11698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11699  return __ret;
11700}
11701#endif
11702
11703#ifdef __LITTLE_ENDIAN__
11704__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11705  int16x8_t __ret;
11706  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
11707  return __ret;
11708}
11709#else
11710__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11711  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11712  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11713  int16x8_t __ret;
11714  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
11715  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11716  return __ret;
11717}
11718#endif
11719
11720#ifdef __LITTLE_ENDIAN__
11721__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11722  uint8x8_t __ret;
11723  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
11724  return __ret;
11725}
11726#else
11727__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11728  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11729  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11730  uint8x8_t __ret;
11731  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
11732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11733  return __ret;
11734}
11735#endif
11736
11737#ifdef __LITTLE_ENDIAN__
11738__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11739  uint32x2_t __ret;
11740  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
11741  return __ret;
11742}
11743#else
11744__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11745  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11746  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11747  uint32x2_t __ret;
11748  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
11749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11750  return __ret;
11751}
11752#endif
11753
11754#ifdef __LITTLE_ENDIAN__
11755__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11756  uint16x4_t __ret;
11757  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
11758  return __ret;
11759}
11760#else
11761__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11762  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11763  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11764  uint16x4_t __ret;
11765  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
11766  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11767  return __ret;
11768}
11769#endif
11770
11771#ifdef __LITTLE_ENDIAN__
11772__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11773  int8x8_t __ret;
11774  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
11775  return __ret;
11776}
11777#else
11778__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11779  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11780  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11781  int8x8_t __ret;
11782  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
11783  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11784  return __ret;
11785}
11786#endif
11787
11788#ifdef __LITTLE_ENDIAN__
11789__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11790  float32x2_t __ret;
11791  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
11792  return __ret;
11793}
11794#else
11795__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11796  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11797  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11798  float32x2_t __ret;
11799  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
11800  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11801  return __ret;
11802}
11803#endif
11804
11805#ifdef __LITTLE_ENDIAN__
11806__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11807  int32x2_t __ret;
11808  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
11809  return __ret;
11810}
11811#else
11812__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11813  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11814  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11815  int32x2_t __ret;
11816  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
11817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11818  return __ret;
11819}
11820#endif
11821
11822#ifdef __LITTLE_ENDIAN__
11823__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11824  int16x4_t __ret;
11825  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
11826  return __ret;
11827}
11828#else
11829__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11830  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11831  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11832  int16x4_t __ret;
11833  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
11834  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11835  return __ret;
11836}
11837#endif
11838
11839#ifdef __LITTLE_ENDIAN__
11840__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11841  uint8x16_t __ret;
11842  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11843  return __ret;
11844}
11845#else
11846__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11847  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11848  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11849  uint8x16_t __ret;
11850  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11851  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11852  return __ret;
11853}
11854#endif
11855
11856#ifdef __LITTLE_ENDIAN__
11857__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11858  uint32x4_t __ret;
11859  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11860  return __ret;
11861}
11862#else
11863__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11864  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11865  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11866  uint32x4_t __ret;
11867  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11869  return __ret;
11870}
11871#endif
11872
11873#ifdef __LITTLE_ENDIAN__
11874__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11875  uint16x8_t __ret;
11876  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11877  return __ret;
11878}
11879#else
11880__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11881  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11882  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11883  uint16x8_t __ret;
11884  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11885  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11886  return __ret;
11887}
11888#endif
11889
11890#ifdef __LITTLE_ENDIAN__
11891__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11892  int8x16_t __ret;
11893  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11894  return __ret;
11895}
11896#else
11897__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11898  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11899  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11900  int8x16_t __ret;
11901  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11902  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11903  return __ret;
11904}
11905#endif
11906
11907#ifdef __LITTLE_ENDIAN__
11908__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
11909  float32x4_t __ret;
11910  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11911  return __ret;
11912}
11913#else
11914__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
11915  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11916  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11917  float32x4_t __ret;
11918  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
11919  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11920  return __ret;
11921}
11922#endif
11923
11924#ifdef __LITTLE_ENDIAN__
11925__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
11926  int32x4_t __ret;
11927  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
11928  return __ret;
11929}
11930#else
11931__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
11932  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11933  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11934  int32x4_t __ret;
11935  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
11936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11937  return __ret;
11938}
11939#endif
11940
11941#ifdef __LITTLE_ENDIAN__
11942__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
11943  int16x8_t __ret;
11944  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
11945  return __ret;
11946}
11947#else
11948__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
11949  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11950  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11951  int16x8_t __ret;
11952  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
11953  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11954  return __ret;
11955}
11956#endif
11957
11958#ifdef __LITTLE_ENDIAN__
11959__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
11960  uint8x8_t __ret;
11961  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
11962  return __ret;
11963}
11964#else
11965__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
11966  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11967  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11968  uint8x8_t __ret;
11969  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
11970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11971  return __ret;
11972}
11973#endif
11974
11975#ifdef __LITTLE_ENDIAN__
11976__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
11977  uint32x2_t __ret;
11978  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
11979  return __ret;
11980}
11981#else
11982__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
11983  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11984  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11985  uint32x2_t __ret;
11986  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
11987  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11988  return __ret;
11989}
11990#endif
11991
11992#ifdef __LITTLE_ENDIAN__
11993__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
11994  uint16x4_t __ret;
11995  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
11996  return __ret;
11997}
11998#else
11999__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12000  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12001  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12002  uint16x4_t __ret;
12003  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
12004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12005  return __ret;
12006}
12007#endif
12008
12009#ifdef __LITTLE_ENDIAN__
12010__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12011  int8x8_t __ret;
12012  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
12013  return __ret;
12014}
12015#else
12016__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12017  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12018  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12019  int8x8_t __ret;
12020  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
12021  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12022  return __ret;
12023}
12024#endif
12025
12026#ifdef __LITTLE_ENDIAN__
12027__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12028  float32x2_t __ret;
12029  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
12030  return __ret;
12031}
12032#else
12033__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12034  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12035  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12036  float32x2_t __ret;
12037  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
12038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12039  return __ret;
12040}
12041#endif
12042
12043#ifdef __LITTLE_ENDIAN__
12044__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12045  int32x2_t __ret;
12046  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
12047  return __ret;
12048}
12049#else
12050__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12051  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12052  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12053  int32x2_t __ret;
12054  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
12055  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12056  return __ret;
12057}
12058#endif
12059
12060#ifdef __LITTLE_ENDIAN__
12061__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12062  int16x4_t __ret;
12063  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
12064  return __ret;
12065}
12066#else
12067__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12068  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12069  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12070  int16x4_t __ret;
12071  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
12072  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12073  return __ret;
12074}
12075#endif
12076
12077#ifdef __LITTLE_ENDIAN__
12078__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12079  uint8x16_t __ret;
12080  __ret = __p0 + __p1 * __p2;
12081  return __ret;
12082}
12083#else
12084__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12085  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12086  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12087  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12088  uint8x16_t __ret;
12089  __ret = __rev0 + __rev1 * __rev2;
12090  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12091  return __ret;
12092}
12093#endif
12094
12095#ifdef __LITTLE_ENDIAN__
12096__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12097  uint32x4_t __ret;
12098  __ret = __p0 + __p1 * __p2;
12099  return __ret;
12100}
12101#else
12102__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12103  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12104  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12105  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12106  uint32x4_t __ret;
12107  __ret = __rev0 + __rev1 * __rev2;
12108  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12109  return __ret;
12110}
12111#endif
12112
12113#ifdef __LITTLE_ENDIAN__
12114__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12115  uint16x8_t __ret;
12116  __ret = __p0 + __p1 * __p2;
12117  return __ret;
12118}
12119#else
12120__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12121  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12122  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12123  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12124  uint16x8_t __ret;
12125  __ret = __rev0 + __rev1 * __rev2;
12126  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12127  return __ret;
12128}
12129#endif
12130
12131#ifdef __LITTLE_ENDIAN__
12132__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12133  int8x16_t __ret;
12134  __ret = __p0 + __p1 * __p2;
12135  return __ret;
12136}
12137#else
12138__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12139  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12140  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12141  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12142  int8x16_t __ret;
12143  __ret = __rev0 + __rev1 * __rev2;
12144  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12145  return __ret;
12146}
12147#endif
12148
12149#ifdef __LITTLE_ENDIAN__
12150__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12151  float32x4_t __ret;
12152  __ret = __p0 + __p1 * __p2;
12153  return __ret;
12154}
12155#else
12156__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12157  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12158  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12159  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12160  float32x4_t __ret;
12161  __ret = __rev0 + __rev1 * __rev2;
12162  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12163  return __ret;
12164}
12165#endif
12166
12167#ifdef __LITTLE_ENDIAN__
12168__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12169  int32x4_t __ret;
12170  __ret = __p0 + __p1 * __p2;
12171  return __ret;
12172}
12173#else
12174__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12175  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12176  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12177  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12178  int32x4_t __ret;
12179  __ret = __rev0 + __rev1 * __rev2;
12180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12181  return __ret;
12182}
12183#endif
12184
12185#ifdef __LITTLE_ENDIAN__
12186__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12187  int16x8_t __ret;
12188  __ret = __p0 + __p1 * __p2;
12189  return __ret;
12190}
12191#else
12192__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12193  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12194  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12195  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12196  int16x8_t __ret;
12197  __ret = __rev0 + __rev1 * __rev2;
12198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12199  return __ret;
12200}
12201#endif
12202
12203#ifdef __LITTLE_ENDIAN__
12204__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12205  uint8x8_t __ret;
12206  __ret = __p0 + __p1 * __p2;
12207  return __ret;
12208}
12209#else
12210__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12211  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12212  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12213  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12214  uint8x8_t __ret;
12215  __ret = __rev0 + __rev1 * __rev2;
12216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12217  return __ret;
12218}
12219#endif
12220
12221#ifdef __LITTLE_ENDIAN__
12222__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12223  uint32x2_t __ret;
12224  __ret = __p0 + __p1 * __p2;
12225  return __ret;
12226}
12227#else
12228__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12229  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12230  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12231  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12232  uint32x2_t __ret;
12233  __ret = __rev0 + __rev1 * __rev2;
12234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12235  return __ret;
12236}
12237#endif
12238
12239#ifdef __LITTLE_ENDIAN__
12240__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12241  uint16x4_t __ret;
12242  __ret = __p0 + __p1 * __p2;
12243  return __ret;
12244}
12245#else
12246__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12247  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12248  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12249  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12250  uint16x4_t __ret;
12251  __ret = __rev0 + __rev1 * __rev2;
12252  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12253  return __ret;
12254}
12255#endif
12256
12257#ifdef __LITTLE_ENDIAN__
12258__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12259  int8x8_t __ret;
12260  __ret = __p0 + __p1 * __p2;
12261  return __ret;
12262}
12263#else
12264__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12265  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12266  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12267  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12268  int8x8_t __ret;
12269  __ret = __rev0 + __rev1 * __rev2;
12270  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12271  return __ret;
12272}
12273#endif
12274
12275#ifdef __LITTLE_ENDIAN__
12276__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12277  float32x2_t __ret;
12278  __ret = __p0 + __p1 * __p2;
12279  return __ret;
12280}
12281#else
12282__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12283  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12284  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12285  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12286  float32x2_t __ret;
12287  __ret = __rev0 + __rev1 * __rev2;
12288  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12289  return __ret;
12290}
12291#endif
12292
12293#ifdef __LITTLE_ENDIAN__
12294__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12295  int32x2_t __ret;
12296  __ret = __p0 + __p1 * __p2;
12297  return __ret;
12298}
12299#else
12300__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12301  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12302  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12303  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12304  int32x2_t __ret;
12305  __ret = __rev0 + __rev1 * __rev2;
12306  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12307  return __ret;
12308}
12309#endif
12310
12311#ifdef __LITTLE_ENDIAN__
12312__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12313  int16x4_t __ret;
12314  __ret = __p0 + __p1 * __p2;
12315  return __ret;
12316}
12317#else
12318__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12319  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12320  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12321  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12322  int16x4_t __ret;
12323  __ret = __rev0 + __rev1 * __rev2;
12324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12325  return __ret;
12326}
12327#endif
12328
12329#ifdef __LITTLE_ENDIAN__
12330#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12331  uint32x4_t __s0 = __p0; \
12332  uint32x4_t __s1 = __p1; \
12333  uint32x2_t __s2 = __p2; \
12334  uint32x4_t __ret; \
12335  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12336  __ret; \
12337})
12338#else
12339#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12340  uint32x4_t __s0 = __p0; \
12341  uint32x4_t __s1 = __p1; \
12342  uint32x2_t __s2 = __p2; \
12343  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12344  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12345  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12346  uint32x4_t __ret; \
12347  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12348  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12349  __ret; \
12350})
12351#endif
12352
12353#ifdef __LITTLE_ENDIAN__
12354#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12355  uint16x8_t __s0 = __p0; \
12356  uint16x8_t __s1 = __p1; \
12357  uint16x4_t __s2 = __p2; \
12358  uint16x8_t __ret; \
12359  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12360  __ret; \
12361})
12362#else
12363#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12364  uint16x8_t __s0 = __p0; \
12365  uint16x8_t __s1 = __p1; \
12366  uint16x4_t __s2 = __p2; \
12367  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12368  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12369  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12370  uint16x8_t __ret; \
12371  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12372  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12373  __ret; \
12374})
12375#endif
12376
12377#ifdef __LITTLE_ENDIAN__
12378#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12379  float32x4_t __s0 = __p0; \
12380  float32x4_t __s1 = __p1; \
12381  float32x2_t __s2 = __p2; \
12382  float32x4_t __ret; \
12383  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12384  __ret; \
12385})
12386#else
12387#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12388  float32x4_t __s0 = __p0; \
12389  float32x4_t __s1 = __p1; \
12390  float32x2_t __s2 = __p2; \
12391  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12392  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12393  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12394  float32x4_t __ret; \
12395  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12396  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12397  __ret; \
12398})
12399#endif
12400
12401#ifdef __LITTLE_ENDIAN__
12402#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12403  int32x4_t __s0 = __p0; \
12404  int32x4_t __s1 = __p1; \
12405  int32x2_t __s2 = __p2; \
12406  int32x4_t __ret; \
12407  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12408  __ret; \
12409})
12410#else
12411#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12412  int32x4_t __s0 = __p0; \
12413  int32x4_t __s1 = __p1; \
12414  int32x2_t __s2 = __p2; \
12415  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12416  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12417  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12418  int32x4_t __ret; \
12419  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12420  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12421  __ret; \
12422})
12423#endif
12424
12425#ifdef __LITTLE_ENDIAN__
12426#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12427  int16x8_t __s0 = __p0; \
12428  int16x8_t __s1 = __p1; \
12429  int16x4_t __s2 = __p2; \
12430  int16x8_t __ret; \
12431  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12432  __ret; \
12433})
12434#else
12435#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12436  int16x8_t __s0 = __p0; \
12437  int16x8_t __s1 = __p1; \
12438  int16x4_t __s2 = __p2; \
12439  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12440  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12441  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12442  int16x8_t __ret; \
12443  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12444  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12445  __ret; \
12446})
12447#endif
12448
12449#ifdef __LITTLE_ENDIAN__
12450#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12451  uint32x2_t __s0 = __p0; \
12452  uint32x2_t __s1 = __p1; \
12453  uint32x2_t __s2 = __p2; \
12454  uint32x2_t __ret; \
12455  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12456  __ret; \
12457})
12458#else
12459#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12460  uint32x2_t __s0 = __p0; \
12461  uint32x2_t __s1 = __p1; \
12462  uint32x2_t __s2 = __p2; \
12463  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12464  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12465  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12466  uint32x2_t __ret; \
12467  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12468  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12469  __ret; \
12470})
12471#endif
12472
12473#ifdef __LITTLE_ENDIAN__
12474#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12475  uint16x4_t __s0 = __p0; \
12476  uint16x4_t __s1 = __p1; \
12477  uint16x4_t __s2 = __p2; \
12478  uint16x4_t __ret; \
12479  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12480  __ret; \
12481})
12482#else
12483#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12484  uint16x4_t __s0 = __p0; \
12485  uint16x4_t __s1 = __p1; \
12486  uint16x4_t __s2 = __p2; \
12487  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12488  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12489  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12490  uint16x4_t __ret; \
12491  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12492  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12493  __ret; \
12494})
12495#endif
12496
12497#ifdef __LITTLE_ENDIAN__
12498#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12499  float32x2_t __s0 = __p0; \
12500  float32x2_t __s1 = __p1; \
12501  float32x2_t __s2 = __p2; \
12502  float32x2_t __ret; \
12503  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12504  __ret; \
12505})
12506#else
12507#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12508  float32x2_t __s0 = __p0; \
12509  float32x2_t __s1 = __p1; \
12510  float32x2_t __s2 = __p2; \
12511  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12512  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12513  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12514  float32x2_t __ret; \
12515  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12517  __ret; \
12518})
12519#endif
12520
12521#ifdef __LITTLE_ENDIAN__
12522#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12523  int32x2_t __s0 = __p0; \
12524  int32x2_t __s1 = __p1; \
12525  int32x2_t __s2 = __p2; \
12526  int32x2_t __ret; \
12527  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12528  __ret; \
12529})
12530#else
12531#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12532  int32x2_t __s0 = __p0; \
12533  int32x2_t __s1 = __p1; \
12534  int32x2_t __s2 = __p2; \
12535  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12536  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12537  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12538  int32x2_t __ret; \
12539  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12540  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12541  __ret; \
12542})
12543#endif
12544
12545#ifdef __LITTLE_ENDIAN__
12546#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12547  int16x4_t __s0 = __p0; \
12548  int16x4_t __s1 = __p1; \
12549  int16x4_t __s2 = __p2; \
12550  int16x4_t __ret; \
12551  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12552  __ret; \
12553})
12554#else
12555#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12556  int16x4_t __s0 = __p0; \
12557  int16x4_t __s1 = __p1; \
12558  int16x4_t __s2 = __p2; \
12559  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12560  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12561  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12562  int16x4_t __ret; \
12563  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12564  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12565  __ret; \
12566})
12567#endif
12568
12569#ifdef __LITTLE_ENDIAN__
12570__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12571  uint32x4_t __ret;
12572  __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12573  return __ret;
12574}
12575#else
12576__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12577  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12578  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12579  uint32x4_t __ret;
12580  __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12581  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12582  return __ret;
12583}
12584#endif
12585
12586#ifdef __LITTLE_ENDIAN__
12587__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12588  uint16x8_t __ret;
12589  __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12590  return __ret;
12591}
12592#else
12593__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12594  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12595  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12596  uint16x8_t __ret;
12597  __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12598  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12599  return __ret;
12600}
12601#endif
12602
12603#ifdef __LITTLE_ENDIAN__
12604__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12605  float32x4_t __ret;
12606  __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12607  return __ret;
12608}
12609#else
12610__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12611  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12612  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12613  float32x4_t __ret;
12614  __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12615  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12616  return __ret;
12617}
12618#endif
12619
12620#ifdef __LITTLE_ENDIAN__
12621__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12622  int32x4_t __ret;
12623  __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12624  return __ret;
12625}
12626#else
12627__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12628  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12629  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12630  int32x4_t __ret;
12631  __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12633  return __ret;
12634}
12635#endif
12636
12637#ifdef __LITTLE_ENDIAN__
12638__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12639  int16x8_t __ret;
12640  __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12641  return __ret;
12642}
12643#else
12644__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12645  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12646  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12647  int16x8_t __ret;
12648  __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12649  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12650  return __ret;
12651}
12652#endif
12653
12654#ifdef __LITTLE_ENDIAN__
12655__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12656  uint32x2_t __ret;
12657  __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
12658  return __ret;
12659}
12660#else
12661__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12662  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12663  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12664  uint32x2_t __ret;
12665  __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
12666  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12667  return __ret;
12668}
12669#endif
12670
12671#ifdef __LITTLE_ENDIAN__
12672__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12673  uint16x4_t __ret;
12674  __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12675  return __ret;
12676}
12677#else
12678__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12679  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12680  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12681  uint16x4_t __ret;
12682  __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12683  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12684  return __ret;
12685}
12686#endif
12687
12688#ifdef __LITTLE_ENDIAN__
12689__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12690  float32x2_t __ret;
12691  __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
12692  return __ret;
12693}
12694#else
12695__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12696  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12697  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12698  float32x2_t __ret;
12699  __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
12700  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12701  return __ret;
12702}
12703#endif
12704
12705#ifdef __LITTLE_ENDIAN__
12706__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12707  int32x2_t __ret;
12708  __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
12709  return __ret;
12710}
12711#else
12712__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12713  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12714  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12715  int32x2_t __ret;
12716  __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
12717  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12718  return __ret;
12719}
12720#endif
12721
12722#ifdef __LITTLE_ENDIAN__
12723__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12724  int16x4_t __ret;
12725  __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12726  return __ret;
12727}
12728#else
12729__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12730  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12731  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12732  int16x4_t __ret;
12733  __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12734  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12735  return __ret;
12736}
12737#endif
12738
12739#ifdef __LITTLE_ENDIAN__
12740__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12741  uint8x16_t __ret;
12742  __ret = __p0 - __p1 * __p2;
12743  return __ret;
12744}
12745#else
12746__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12747  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12748  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12749  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12750  uint8x16_t __ret;
12751  __ret = __rev0 - __rev1 * __rev2;
12752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12753  return __ret;
12754}
12755#endif
12756
12757#ifdef __LITTLE_ENDIAN__
12758__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12759  uint32x4_t __ret;
12760  __ret = __p0 - __p1 * __p2;
12761  return __ret;
12762}
12763#else
12764__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12765  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12766  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12767  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12768  uint32x4_t __ret;
12769  __ret = __rev0 - __rev1 * __rev2;
12770  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12771  return __ret;
12772}
12773#endif
12774
12775#ifdef __LITTLE_ENDIAN__
12776__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12777  uint16x8_t __ret;
12778  __ret = __p0 - __p1 * __p2;
12779  return __ret;
12780}
12781#else
12782__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12783  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12784  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12785  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12786  uint16x8_t __ret;
12787  __ret = __rev0 - __rev1 * __rev2;
12788  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12789  return __ret;
12790}
12791#endif
12792
12793#ifdef __LITTLE_ENDIAN__
12794__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12795  int8x16_t __ret;
12796  __ret = __p0 - __p1 * __p2;
12797  return __ret;
12798}
12799#else
12800__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12801  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12802  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12803  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12804  int8x16_t __ret;
12805  __ret = __rev0 - __rev1 * __rev2;
12806  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12807  return __ret;
12808}
12809#endif
12810
12811#ifdef __LITTLE_ENDIAN__
12812__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12813  float32x4_t __ret;
12814  __ret = __p0 - __p1 * __p2;
12815  return __ret;
12816}
12817#else
12818__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12819  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12820  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12821  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12822  float32x4_t __ret;
12823  __ret = __rev0 - __rev1 * __rev2;
12824  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12825  return __ret;
12826}
12827#endif
12828
12829#ifdef __LITTLE_ENDIAN__
12830__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12831  int32x4_t __ret;
12832  __ret = __p0 - __p1 * __p2;
12833  return __ret;
12834}
12835#else
12836__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12837  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12838  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12839  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12840  int32x4_t __ret;
12841  __ret = __rev0 - __rev1 * __rev2;
12842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12843  return __ret;
12844}
12845#endif
12846
12847#ifdef __LITTLE_ENDIAN__
12848__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12849  int16x8_t __ret;
12850  __ret = __p0 - __p1 * __p2;
12851  return __ret;
12852}
12853#else
12854__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12855  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12856  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12857  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12858  int16x8_t __ret;
12859  __ret = __rev0 - __rev1 * __rev2;
12860  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12861  return __ret;
12862}
12863#endif
12864
12865#ifdef __LITTLE_ENDIAN__
12866__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12867  uint8x8_t __ret;
12868  __ret = __p0 - __p1 * __p2;
12869  return __ret;
12870}
12871#else
12872__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12873  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12874  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12875  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12876  uint8x8_t __ret;
12877  __ret = __rev0 - __rev1 * __rev2;
12878  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12879  return __ret;
12880}
12881#endif
12882
12883#ifdef __LITTLE_ENDIAN__
12884__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12885  uint32x2_t __ret;
12886  __ret = __p0 - __p1 * __p2;
12887  return __ret;
12888}
12889#else
12890__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12891  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12892  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12893  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12894  uint32x2_t __ret;
12895  __ret = __rev0 - __rev1 * __rev2;
12896  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12897  return __ret;
12898}
12899#endif
12900
12901#ifdef __LITTLE_ENDIAN__
12902__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12903  uint16x4_t __ret;
12904  __ret = __p0 - __p1 * __p2;
12905  return __ret;
12906}
12907#else
12908__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12909  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12910  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12911  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12912  uint16x4_t __ret;
12913  __ret = __rev0 - __rev1 * __rev2;
12914  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12915  return __ret;
12916}
12917#endif
12918
12919#ifdef __LITTLE_ENDIAN__
12920__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12921  int8x8_t __ret;
12922  __ret = __p0 - __p1 * __p2;
12923  return __ret;
12924}
12925#else
12926__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12927  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12928  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12929  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12930  int8x8_t __ret;
12931  __ret = __rev0 - __rev1 * __rev2;
12932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12933  return __ret;
12934}
12935#endif
12936
12937#ifdef __LITTLE_ENDIAN__
12938__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12939  float32x2_t __ret;
12940  __ret = __p0 - __p1 * __p2;
12941  return __ret;
12942}
12943#else
12944__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12945  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12946  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12947  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12948  float32x2_t __ret;
12949  __ret = __rev0 - __rev1 * __rev2;
12950  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12951  return __ret;
12952}
12953#endif
12954
12955#ifdef __LITTLE_ENDIAN__
12956__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12957  int32x2_t __ret;
12958  __ret = __p0 - __p1 * __p2;
12959  return __ret;
12960}
12961#else
12962__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12963  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12964  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12965  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12966  int32x2_t __ret;
12967  __ret = __rev0 - __rev1 * __rev2;
12968  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12969  return __ret;
12970}
12971#endif
12972
12973#ifdef __LITTLE_ENDIAN__
12974__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12975  int16x4_t __ret;
12976  __ret = __p0 - __p1 * __p2;
12977  return __ret;
12978}
12979#else
12980__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12981  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12982  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12983  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12984  int16x4_t __ret;
12985  __ret = __rev0 - __rev1 * __rev2;
12986  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12987  return __ret;
12988}
12989#endif
12990
12991#ifdef __LITTLE_ENDIAN__
12992#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12993  uint32x4_t __s0 = __p0; \
12994  uint32x4_t __s1 = __p1; \
12995  uint32x2_t __s2 = __p2; \
12996  uint32x4_t __ret; \
12997  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12998  __ret; \
12999})
13000#else
13001#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13002  uint32x4_t __s0 = __p0; \
13003  uint32x4_t __s1 = __p1; \
13004  uint32x2_t __s2 = __p2; \
13005  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13006  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13007  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13008  uint32x4_t __ret; \
13009  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13010  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13011  __ret; \
13012})
13013#endif
13014
13015#ifdef __LITTLE_ENDIAN__
13016#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13017  uint16x8_t __s0 = __p0; \
13018  uint16x8_t __s1 = __p1; \
13019  uint16x4_t __s2 = __p2; \
13020  uint16x8_t __ret; \
13021  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13022  __ret; \
13023})
13024#else
13025#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13026  uint16x8_t __s0 = __p0; \
13027  uint16x8_t __s1 = __p1; \
13028  uint16x4_t __s2 = __p2; \
13029  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13030  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13031  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13032  uint16x8_t __ret; \
13033  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13034  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13035  __ret; \
13036})
13037#endif
13038
13039#ifdef __LITTLE_ENDIAN__
13040#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13041  float32x4_t __s0 = __p0; \
13042  float32x4_t __s1 = __p1; \
13043  float32x2_t __s2 = __p2; \
13044  float32x4_t __ret; \
13045  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13046  __ret; \
13047})
13048#else
13049#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13050  float32x4_t __s0 = __p0; \
13051  float32x4_t __s1 = __p1; \
13052  float32x2_t __s2 = __p2; \
13053  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13054  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13055  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13056  float32x4_t __ret; \
13057  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13058  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13059  __ret; \
13060})
13061#endif
13062
13063#ifdef __LITTLE_ENDIAN__
13064#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13065  int32x4_t __s0 = __p0; \
13066  int32x4_t __s1 = __p1; \
13067  int32x2_t __s2 = __p2; \
13068  int32x4_t __ret; \
13069  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13070  __ret; \
13071})
13072#else
13073#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13074  int32x4_t __s0 = __p0; \
13075  int32x4_t __s1 = __p1; \
13076  int32x2_t __s2 = __p2; \
13077  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13078  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13079  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13080  int32x4_t __ret; \
13081  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13082  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13083  __ret; \
13084})
13085#endif
13086
13087#ifdef __LITTLE_ENDIAN__
13088#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13089  int16x8_t __s0 = __p0; \
13090  int16x8_t __s1 = __p1; \
13091  int16x4_t __s2 = __p2; \
13092  int16x8_t __ret; \
13093  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13094  __ret; \
13095})
13096#else
13097#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13098  int16x8_t __s0 = __p0; \
13099  int16x8_t __s1 = __p1; \
13100  int16x4_t __s2 = __p2; \
13101  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13102  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13103  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13104  int16x8_t __ret; \
13105  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13107  __ret; \
13108})
13109#endif
13110
13111#ifdef __LITTLE_ENDIAN__
13112#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13113  uint32x2_t __s0 = __p0; \
13114  uint32x2_t __s1 = __p1; \
13115  uint32x2_t __s2 = __p2; \
13116  uint32x2_t __ret; \
13117  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13118  __ret; \
13119})
13120#else
13121#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13122  uint32x2_t __s0 = __p0; \
13123  uint32x2_t __s1 = __p1; \
13124  uint32x2_t __s2 = __p2; \
13125  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13126  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13127  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13128  uint32x2_t __ret; \
13129  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13130  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13131  __ret; \
13132})
13133#endif
13134
13135#ifdef __LITTLE_ENDIAN__
13136#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13137  uint16x4_t __s0 = __p0; \
13138  uint16x4_t __s1 = __p1; \
13139  uint16x4_t __s2 = __p2; \
13140  uint16x4_t __ret; \
13141  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13142  __ret; \
13143})
13144#else
13145#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13146  uint16x4_t __s0 = __p0; \
13147  uint16x4_t __s1 = __p1; \
13148  uint16x4_t __s2 = __p2; \
13149  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13150  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13151  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13152  uint16x4_t __ret; \
13153  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13154  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13155  __ret; \
13156})
13157#endif
13158
13159#ifdef __LITTLE_ENDIAN__
13160#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13161  float32x2_t __s0 = __p0; \
13162  float32x2_t __s1 = __p1; \
13163  float32x2_t __s2 = __p2; \
13164  float32x2_t __ret; \
13165  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13166  __ret; \
13167})
13168#else
13169#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13170  float32x2_t __s0 = __p0; \
13171  float32x2_t __s1 = __p1; \
13172  float32x2_t __s2 = __p2; \
13173  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13174  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13175  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13176  float32x2_t __ret; \
13177  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13178  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13179  __ret; \
13180})
13181#endif
13182
13183#ifdef __LITTLE_ENDIAN__
13184#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13185  int32x2_t __s0 = __p0; \
13186  int32x2_t __s1 = __p1; \
13187  int32x2_t __s2 = __p2; \
13188  int32x2_t __ret; \
13189  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13190  __ret; \
13191})
13192#else
13193#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13194  int32x2_t __s0 = __p0; \
13195  int32x2_t __s1 = __p1; \
13196  int32x2_t __s2 = __p2; \
13197  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13198  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13199  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13200  int32x2_t __ret; \
13201  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13203  __ret; \
13204})
13205#endif
13206
13207#ifdef __LITTLE_ENDIAN__
13208#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13209  int16x4_t __s0 = __p0; \
13210  int16x4_t __s1 = __p1; \
13211  int16x4_t __s2 = __p2; \
13212  int16x4_t __ret; \
13213  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13214  __ret; \
13215})
13216#else
13217#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13218  int16x4_t __s0 = __p0; \
13219  int16x4_t __s1 = __p1; \
13220  int16x4_t __s2 = __p2; \
13221  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13222  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13223  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13224  int16x4_t __ret; \
13225  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13226  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13227  __ret; \
13228})
13229#endif
13230
13231#ifdef __LITTLE_ENDIAN__
13232__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13233  uint32x4_t __ret;
13234  __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13235  return __ret;
13236}
13237#else
13238__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13239  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13240  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13241  uint32x4_t __ret;
13242  __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13244  return __ret;
13245}
13246#endif
13247
13248#ifdef __LITTLE_ENDIAN__
13249__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13250  uint16x8_t __ret;
13251  __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13252  return __ret;
13253}
13254#else
13255__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13256  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13257  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13258  uint16x8_t __ret;
13259  __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13260  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13261  return __ret;
13262}
13263#endif
13264
13265#ifdef __LITTLE_ENDIAN__
13266__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13267  float32x4_t __ret;
13268  __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13269  return __ret;
13270}
13271#else
13272__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13273  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13274  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13275  float32x4_t __ret;
13276  __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13277  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13278  return __ret;
13279}
13280#endif
13281
13282#ifdef __LITTLE_ENDIAN__
13283__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13284  int32x4_t __ret;
13285  __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13286  return __ret;
13287}
13288#else
13289__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13290  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13291  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13292  int32x4_t __ret;
13293  __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13294  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13295  return __ret;
13296}
13297#endif
13298
13299#ifdef __LITTLE_ENDIAN__
13300__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13301  int16x8_t __ret;
13302  __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13303  return __ret;
13304}
13305#else
13306__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13307  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13308  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13309  int16x8_t __ret;
13310  __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13311  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13312  return __ret;
13313}
13314#endif
13315
13316#ifdef __LITTLE_ENDIAN__
13317__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13318  uint32x2_t __ret;
13319  __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
13320  return __ret;
13321}
13322#else
13323__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13324  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13325  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13326  uint32x2_t __ret;
13327  __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
13328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13329  return __ret;
13330}
13331#endif
13332
13333#ifdef __LITTLE_ENDIAN__
13334__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13335  uint16x4_t __ret;
13336  __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13337  return __ret;
13338}
13339#else
13340__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13341  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13342  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13343  uint16x4_t __ret;
13344  __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13345  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13346  return __ret;
13347}
13348#endif
13349
13350#ifdef __LITTLE_ENDIAN__
13351__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13352  float32x2_t __ret;
13353  __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
13354  return __ret;
13355}
13356#else
13357__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13358  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13359  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13360  float32x2_t __ret;
13361  __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
13362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13363  return __ret;
13364}
13365#endif
13366
13367#ifdef __LITTLE_ENDIAN__
13368__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13369  int32x2_t __ret;
13370  __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
13371  return __ret;
13372}
13373#else
13374__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13375  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13376  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13377  int32x2_t __ret;
13378  __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
13379  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13380  return __ret;
13381}
13382#endif
13383
13384#ifdef __LITTLE_ENDIAN__
13385__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13386  int16x4_t __ret;
13387  __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13388  return __ret;
13389}
13390#else
13391__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13392  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13393  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13394  int16x4_t __ret;
13395  __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13396  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13397  return __ret;
13398}
13399#endif
13400
13401#ifdef __LITTLE_ENDIAN__
13402__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13403  poly8x8_t __ret;
13404  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13405  return __ret;
13406}
13407#else
13408__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13409  poly8x8_t __ret;
13410  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13411  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13412  return __ret;
13413}
13414#endif
13415
13416#ifdef __LITTLE_ENDIAN__
13417__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13418  poly16x4_t __ret;
13419  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13420  return __ret;
13421}
13422#else
13423__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13424  poly16x4_t __ret;
13425  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13426  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13427  return __ret;
13428}
13429#endif
13430
13431#ifdef __LITTLE_ENDIAN__
13432__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13433  poly8x16_t __ret;
13434  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13435  return __ret;
13436}
13437#else
13438__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13439  poly8x16_t __ret;
13440  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13441  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13442  return __ret;
13443}
13444#endif
13445
13446#ifdef __LITTLE_ENDIAN__
13447__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13448  poly16x8_t __ret;
13449  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13450  return __ret;
13451}
13452#else
13453__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13454  poly16x8_t __ret;
13455  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13456  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13457  return __ret;
13458}
13459#endif
13460
13461#ifdef __LITTLE_ENDIAN__
13462__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13463  uint8x16_t __ret;
13464  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13465  return __ret;
13466}
13467#else
13468__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13469  uint8x16_t __ret;
13470  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13471  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13472  return __ret;
13473}
13474#endif
13475
13476#ifdef __LITTLE_ENDIAN__
13477__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13478  uint32x4_t __ret;
13479  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13480  return __ret;
13481}
13482#else
13483__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13484  uint32x4_t __ret;
13485  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13487  return __ret;
13488}
13489#endif
13490
13491#ifdef __LITTLE_ENDIAN__
13492__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13493  uint64x2_t __ret;
13494  __ret = (uint64x2_t) {__p0, __p0};
13495  return __ret;
13496}
13497#else
13498__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13499  uint64x2_t __ret;
13500  __ret = (uint64x2_t) {__p0, __p0};
13501  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13502  return __ret;
13503}
13504#endif
13505
13506#ifdef __LITTLE_ENDIAN__
13507__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13508  uint16x8_t __ret;
13509  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13510  return __ret;
13511}
13512#else
13513__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13514  uint16x8_t __ret;
13515  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13516  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13517  return __ret;
13518}
13519#endif
13520
13521#ifdef __LITTLE_ENDIAN__
13522__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13523  int8x16_t __ret;
13524  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13525  return __ret;
13526}
13527#else
13528__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13529  int8x16_t __ret;
13530  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13531  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13532  return __ret;
13533}
13534#endif
13535
13536#ifdef __LITTLE_ENDIAN__
13537__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13538  float32x4_t __ret;
13539  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13540  return __ret;
13541}
13542#else
13543__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13544  float32x4_t __ret;
13545  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13546  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13547  return __ret;
13548}
13549#endif
13550
13551#ifdef __LITTLE_ENDIAN__
13552#define vmovq_n_f16(__p0) __extension__ ({ \
13553  float16_t __s0 = __p0; \
13554  float16x8_t __ret; \
13555  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13556  __ret; \
13557})
13558#else
13559#define vmovq_n_f16(__p0) __extension__ ({ \
13560  float16_t __s0 = __p0; \
13561  float16x8_t __ret; \
13562  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13563  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13564  __ret; \
13565})
13566#endif
13567
13568#ifdef __LITTLE_ENDIAN__
13569__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13570  int32x4_t __ret;
13571  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13572  return __ret;
13573}
13574#else
13575__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13576  int32x4_t __ret;
13577  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13578  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13579  return __ret;
13580}
13581#endif
13582
13583#ifdef __LITTLE_ENDIAN__
13584__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13585  int64x2_t __ret;
13586  __ret = (int64x2_t) {__p0, __p0};
13587  return __ret;
13588}
13589#else
13590__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13591  int64x2_t __ret;
13592  __ret = (int64x2_t) {__p0, __p0};
13593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13594  return __ret;
13595}
13596#endif
13597
13598#ifdef __LITTLE_ENDIAN__
13599__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13600  int16x8_t __ret;
13601  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13602  return __ret;
13603}
13604#else
13605__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13606  int16x8_t __ret;
13607  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13608  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13609  return __ret;
13610}
13611#endif
13612
13613#ifdef __LITTLE_ENDIAN__
13614__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13615  uint8x8_t __ret;
13616  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13617  return __ret;
13618}
13619#else
13620__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13621  uint8x8_t __ret;
13622  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13623  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13624  return __ret;
13625}
13626#endif
13627
13628#ifdef __LITTLE_ENDIAN__
13629__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13630  uint32x2_t __ret;
13631  __ret = (uint32x2_t) {__p0, __p0};
13632  return __ret;
13633}
13634#else
13635__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13636  uint32x2_t __ret;
13637  __ret = (uint32x2_t) {__p0, __p0};
13638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13639  return __ret;
13640}
13641#endif
13642
13643#ifdef __LITTLE_ENDIAN__
13644__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13645  uint64x1_t __ret;
13646  __ret = (uint64x1_t) {__p0};
13647  return __ret;
13648}
13649#else
13650__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13651  uint64x1_t __ret;
13652  __ret = (uint64x1_t) {__p0};
13653  return __ret;
13654}
13655#endif
13656
13657#ifdef __LITTLE_ENDIAN__
13658__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13659  uint16x4_t __ret;
13660  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13661  return __ret;
13662}
13663#else
13664__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13665  uint16x4_t __ret;
13666  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13667  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13668  return __ret;
13669}
13670#endif
13671
13672#ifdef __LITTLE_ENDIAN__
13673__ai int8x8_t vmov_n_s8(int8_t __p0) {
13674  int8x8_t __ret;
13675  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13676  return __ret;
13677}
13678#else
13679__ai int8x8_t vmov_n_s8(int8_t __p0) {
13680  int8x8_t __ret;
13681  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13682  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13683  return __ret;
13684}
13685#endif
13686
13687#ifdef __LITTLE_ENDIAN__
13688__ai float32x2_t vmov_n_f32(float32_t __p0) {
13689  float32x2_t __ret;
13690  __ret = (float32x2_t) {__p0, __p0};
13691  return __ret;
13692}
13693#else
13694__ai float32x2_t vmov_n_f32(float32_t __p0) {
13695  float32x2_t __ret;
13696  __ret = (float32x2_t) {__p0, __p0};
13697  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13698  return __ret;
13699}
13700#endif
13701
13702#ifdef __LITTLE_ENDIAN__
13703#define vmov_n_f16(__p0) __extension__ ({ \
13704  float16_t __s0 = __p0; \
13705  float16x4_t __ret; \
13706  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13707  __ret; \
13708})
13709#else
13710#define vmov_n_f16(__p0) __extension__ ({ \
13711  float16_t __s0 = __p0; \
13712  float16x4_t __ret; \
13713  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13714  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13715  __ret; \
13716})
13717#endif
13718
13719#ifdef __LITTLE_ENDIAN__
13720__ai int32x2_t vmov_n_s32(int32_t __p0) {
13721  int32x2_t __ret;
13722  __ret = (int32x2_t) {__p0, __p0};
13723  return __ret;
13724}
13725#else
13726__ai int32x2_t vmov_n_s32(int32_t __p0) {
13727  int32x2_t __ret;
13728  __ret = (int32x2_t) {__p0, __p0};
13729  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13730  return __ret;
13731}
13732#endif
13733
13734#ifdef __LITTLE_ENDIAN__
13735__ai int64x1_t vmov_n_s64(int64_t __p0) {
13736  int64x1_t __ret;
13737  __ret = (int64x1_t) {__p0};
13738  return __ret;
13739}
13740#else
13741__ai int64x1_t vmov_n_s64(int64_t __p0) {
13742  int64x1_t __ret;
13743  __ret = (int64x1_t) {__p0};
13744  return __ret;
13745}
13746#endif
13747
13748#ifdef __LITTLE_ENDIAN__
13749__ai int16x4_t vmov_n_s16(int16_t __p0) {
13750  int16x4_t __ret;
13751  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13752  return __ret;
13753}
13754#else
13755__ai int16x4_t vmov_n_s16(int16_t __p0) {
13756  int16x4_t __ret;
13757  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13758  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13759  return __ret;
13760}
13761#endif
13762
13763#ifdef __LITTLE_ENDIAN__
13764__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13765  uint16x8_t __ret;
13766  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13767  return __ret;
13768}
13769#else
13770__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13771  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13772  uint16x8_t __ret;
13773  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
13774  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13775  return __ret;
13776}
13777__ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
13778  uint16x8_t __ret;
13779  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13780  return __ret;
13781}
13782#endif
13783
13784#ifdef __LITTLE_ENDIAN__
13785__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13786  uint64x2_t __ret;
13787  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13788  return __ret;
13789}
13790#else
13791__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13792  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13793  uint64x2_t __ret;
13794  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
13795  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13796  return __ret;
13797}
13798__ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
13799  uint64x2_t __ret;
13800  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13801  return __ret;
13802}
13803#endif
13804
13805#ifdef __LITTLE_ENDIAN__
13806__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13807  uint32x4_t __ret;
13808  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13809  return __ret;
13810}
13811#else
13812__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13813  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13814  uint32x4_t __ret;
13815  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
13816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13817  return __ret;
13818}
13819__ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
13820  uint32x4_t __ret;
13821  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13822  return __ret;
13823}
13824#endif
13825
13826#ifdef __LITTLE_ENDIAN__
13827__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13828  int16x8_t __ret;
13829  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13830  return __ret;
13831}
13832#else
13833__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13834  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13835  int16x8_t __ret;
13836  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
13837  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13838  return __ret;
13839}
13840__ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
13841  int16x8_t __ret;
13842  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13843  return __ret;
13844}
13845#endif
13846
13847#ifdef __LITTLE_ENDIAN__
13848__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13849  int64x2_t __ret;
13850  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13851  return __ret;
13852}
13853#else
13854__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13855  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13856  int64x2_t __ret;
13857  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
13858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13859  return __ret;
13860}
13861__ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
13862  int64x2_t __ret;
13863  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13864  return __ret;
13865}
13866#endif
13867
13868#ifdef __LITTLE_ENDIAN__
13869__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13870  int32x4_t __ret;
13871  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13872  return __ret;
13873}
13874#else
13875__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13876  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13877  int32x4_t __ret;
13878  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
13879  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13880  return __ret;
13881}
13882__ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
13883  int32x4_t __ret;
13884  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13885  return __ret;
13886}
13887#endif
13888
13889#ifdef __LITTLE_ENDIAN__
13890__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13891  uint16x4_t __ret;
13892  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13893  return __ret;
13894}
13895#else
13896__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13897  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13898  uint16x4_t __ret;
13899  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
13900  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13901  return __ret;
13902}
13903__ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
13904  uint16x4_t __ret;
13905  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13906  return __ret;
13907}
13908#endif
13909
13910#ifdef __LITTLE_ENDIAN__
13911__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
13912  uint32x2_t __ret;
13913  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
13914  return __ret;
13915}
13916#else
13917__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
13918  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13919  uint32x2_t __ret;
13920  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
13921  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13922  return __ret;
13923}
13924__ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
13925  uint32x2_t __ret;
13926  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
13927  return __ret;
13928}
13929#endif
13930
13931#ifdef __LITTLE_ENDIAN__
13932__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
13933  uint8x8_t __ret;
13934  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
13935  return __ret;
13936}
13937#else
13938__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
13939  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13940  uint8x8_t __ret;
13941  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
13942  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13943  return __ret;
13944}
13945__ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
13946  uint8x8_t __ret;
13947  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
13948  return __ret;
13949}
13950#endif
13951
13952#ifdef __LITTLE_ENDIAN__
13953__ai int16x4_t vmovn_s32(int32x4_t __p0) {
13954  int16x4_t __ret;
13955  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
13956  return __ret;
13957}
13958#else
13959__ai int16x4_t vmovn_s32(int32x4_t __p0) {
13960  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13961  int16x4_t __ret;
13962  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
13963  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13964  return __ret;
13965}
13966__ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
13967  int16x4_t __ret;
13968  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
13969  return __ret;
13970}
13971#endif
13972
13973#ifdef __LITTLE_ENDIAN__
13974__ai int32x2_t vmovn_s64(int64x2_t __p0) {
13975  int32x2_t __ret;
13976  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
13977  return __ret;
13978}
13979#else
13980__ai int32x2_t vmovn_s64(int64x2_t __p0) {
13981  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13982  int32x2_t __ret;
13983  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
13984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13985  return __ret;
13986}
13987__ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
13988  int32x2_t __ret;
13989  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
13990  return __ret;
13991}
13992#endif
13993
13994#ifdef __LITTLE_ENDIAN__
13995__ai int8x8_t vmovn_s16(int16x8_t __p0) {
13996  int8x8_t __ret;
13997  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
13998  return __ret;
13999}
14000#else
14001__ai int8x8_t vmovn_s16(int16x8_t __p0) {
14002  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14003  int8x8_t __ret;
14004  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
14005  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14006  return __ret;
14007}
14008__ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
14009  int8x8_t __ret;
14010  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14011  return __ret;
14012}
14013#endif
14014
14015#ifdef __LITTLE_ENDIAN__
14016__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14017  uint8x16_t __ret;
14018  __ret = __p0 * __p1;
14019  return __ret;
14020}
14021#else
14022__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14023  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14024  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14025  uint8x16_t __ret;
14026  __ret = __rev0 * __rev1;
14027  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14028  return __ret;
14029}
14030#endif
14031
14032#ifdef __LITTLE_ENDIAN__
14033__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14034  uint32x4_t __ret;
14035  __ret = __p0 * __p1;
14036  return __ret;
14037}
14038#else
14039__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14040  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14041  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14042  uint32x4_t __ret;
14043  __ret = __rev0 * __rev1;
14044  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14045  return __ret;
14046}
14047#endif
14048
14049#ifdef __LITTLE_ENDIAN__
14050__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14051  uint16x8_t __ret;
14052  __ret = __p0 * __p1;
14053  return __ret;
14054}
14055#else
14056__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14057  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14058  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14059  uint16x8_t __ret;
14060  __ret = __rev0 * __rev1;
14061  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14062  return __ret;
14063}
14064#endif
14065
14066#ifdef __LITTLE_ENDIAN__
14067__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14068  int8x16_t __ret;
14069  __ret = __p0 * __p1;
14070  return __ret;
14071}
14072#else
14073__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14074  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14075  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14076  int8x16_t __ret;
14077  __ret = __rev0 * __rev1;
14078  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14079  return __ret;
14080}
14081#endif
14082
14083#ifdef __LITTLE_ENDIAN__
14084__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14085  float32x4_t __ret;
14086  __ret = __p0 * __p1;
14087  return __ret;
14088}
14089#else
14090__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14091  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14092  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14093  float32x4_t __ret;
14094  __ret = __rev0 * __rev1;
14095  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14096  return __ret;
14097}
14098#endif
14099
14100#ifdef __LITTLE_ENDIAN__
14101__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14102  int32x4_t __ret;
14103  __ret = __p0 * __p1;
14104  return __ret;
14105}
14106#else
14107__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14108  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14109  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14110  int32x4_t __ret;
14111  __ret = __rev0 * __rev1;
14112  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14113  return __ret;
14114}
14115#endif
14116
14117#ifdef __LITTLE_ENDIAN__
14118__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14119  int16x8_t __ret;
14120  __ret = __p0 * __p1;
14121  return __ret;
14122}
14123#else
14124__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14125  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14126  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14127  int16x8_t __ret;
14128  __ret = __rev0 * __rev1;
14129  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14130  return __ret;
14131}
14132#endif
14133
14134#ifdef __LITTLE_ENDIAN__
14135__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14136  uint8x8_t __ret;
14137  __ret = __p0 * __p1;
14138  return __ret;
14139}
14140#else
14141__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14142  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14143  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14144  uint8x8_t __ret;
14145  __ret = __rev0 * __rev1;
14146  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14147  return __ret;
14148}
14149#endif
14150
14151#ifdef __LITTLE_ENDIAN__
14152__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14153  uint32x2_t __ret;
14154  __ret = __p0 * __p1;
14155  return __ret;
14156}
14157#else
14158__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14159  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14160  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14161  uint32x2_t __ret;
14162  __ret = __rev0 * __rev1;
14163  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14164  return __ret;
14165}
14166#endif
14167
14168#ifdef __LITTLE_ENDIAN__
14169__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14170  uint16x4_t __ret;
14171  __ret = __p0 * __p1;
14172  return __ret;
14173}
14174#else
14175__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14176  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14177  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14178  uint16x4_t __ret;
14179  __ret = __rev0 * __rev1;
14180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14181  return __ret;
14182}
14183#endif
14184
14185#ifdef __LITTLE_ENDIAN__
14186__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14187  int8x8_t __ret;
14188  __ret = __p0 * __p1;
14189  return __ret;
14190}
14191#else
14192__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14193  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14194  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14195  int8x8_t __ret;
14196  __ret = __rev0 * __rev1;
14197  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14198  return __ret;
14199}
14200#endif
14201
14202#ifdef __LITTLE_ENDIAN__
14203__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14204  float32x2_t __ret;
14205  __ret = __p0 * __p1;
14206  return __ret;
14207}
14208#else
14209__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14210  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14211  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14212  float32x2_t __ret;
14213  __ret = __rev0 * __rev1;
14214  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14215  return __ret;
14216}
14217#endif
14218
14219#ifdef __LITTLE_ENDIAN__
14220__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14221  int32x2_t __ret;
14222  __ret = __p0 * __p1;
14223  return __ret;
14224}
14225#else
14226__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14227  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14228  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14229  int32x2_t __ret;
14230  __ret = __rev0 * __rev1;
14231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14232  return __ret;
14233}
14234#endif
14235
14236#ifdef __LITTLE_ENDIAN__
14237__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14238  int16x4_t __ret;
14239  __ret = __p0 * __p1;
14240  return __ret;
14241}
14242#else
14243__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14244  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14245  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14246  int16x4_t __ret;
14247  __ret = __rev0 * __rev1;
14248  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14249  return __ret;
14250}
14251#endif
14252
14253#ifdef __LITTLE_ENDIAN__
14254__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14255  poly8x8_t __ret;
14256  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
14257  return __ret;
14258}
14259#else
14260__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14261  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14262  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14263  poly8x8_t __ret;
14264  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
14265  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14266  return __ret;
14267}
14268#endif
14269
14270#ifdef __LITTLE_ENDIAN__
14271__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14272  poly8x16_t __ret;
14273  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
14274  return __ret;
14275}
14276#else
14277__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14278  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14279  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14280  poly8x16_t __ret;
14281  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
14282  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14283  return __ret;
14284}
14285#endif
14286
14287#ifdef __LITTLE_ENDIAN__
14288#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14289  uint32x4_t __s0 = __p0; \
14290  uint32x2_t __s1 = __p1; \
14291  uint32x4_t __ret; \
14292  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14293  __ret; \
14294})
14295#else
14296#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14297  uint32x4_t __s0 = __p0; \
14298  uint32x2_t __s1 = __p1; \
14299  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14300  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14301  uint32x4_t __ret; \
14302  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14303  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14304  __ret; \
14305})
14306#endif
14307
14308#ifdef __LITTLE_ENDIAN__
14309#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14310  uint16x8_t __s0 = __p0; \
14311  uint16x4_t __s1 = __p1; \
14312  uint16x8_t __ret; \
14313  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14314  __ret; \
14315})
14316#else
14317#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14318  uint16x8_t __s0 = __p0; \
14319  uint16x4_t __s1 = __p1; \
14320  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14321  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14322  uint16x8_t __ret; \
14323  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14324  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14325  __ret; \
14326})
14327#endif
14328
14329#ifdef __LITTLE_ENDIAN__
14330#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14331  float32x4_t __s0 = __p0; \
14332  float32x2_t __s1 = __p1; \
14333  float32x4_t __ret; \
14334  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14335  __ret; \
14336})
14337#else
14338#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14339  float32x4_t __s0 = __p0; \
14340  float32x2_t __s1 = __p1; \
14341  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14342  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14343  float32x4_t __ret; \
14344  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14345  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14346  __ret; \
14347})
14348#endif
14349
14350#ifdef __LITTLE_ENDIAN__
14351#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14352  int32x4_t __s0 = __p0; \
14353  int32x2_t __s1 = __p1; \
14354  int32x4_t __ret; \
14355  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14356  __ret; \
14357})
14358#else
14359#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14360  int32x4_t __s0 = __p0; \
14361  int32x2_t __s1 = __p1; \
14362  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14363  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14364  int32x4_t __ret; \
14365  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14366  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14367  __ret; \
14368})
14369#endif
14370
14371#ifdef __LITTLE_ENDIAN__
14372#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14373  int16x8_t __s0 = __p0; \
14374  int16x4_t __s1 = __p1; \
14375  int16x8_t __ret; \
14376  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14377  __ret; \
14378})
14379#else
14380#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14381  int16x8_t __s0 = __p0; \
14382  int16x4_t __s1 = __p1; \
14383  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14384  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14385  int16x8_t __ret; \
14386  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14387  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14388  __ret; \
14389})
14390#endif
14391
14392#ifdef __LITTLE_ENDIAN__
14393#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14394  uint32x2_t __s0 = __p0; \
14395  uint32x2_t __s1 = __p1; \
14396  uint32x2_t __ret; \
14397  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14398  __ret; \
14399})
14400#else
14401#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14402  uint32x2_t __s0 = __p0; \
14403  uint32x2_t __s1 = __p1; \
14404  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14405  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14406  uint32x2_t __ret; \
14407  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14409  __ret; \
14410})
14411#endif
14412
14413#ifdef __LITTLE_ENDIAN__
14414#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14415  uint16x4_t __s0 = __p0; \
14416  uint16x4_t __s1 = __p1; \
14417  uint16x4_t __ret; \
14418  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14419  __ret; \
14420})
14421#else
14422#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14423  uint16x4_t __s0 = __p0; \
14424  uint16x4_t __s1 = __p1; \
14425  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14426  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14427  uint16x4_t __ret; \
14428  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14429  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14430  __ret; \
14431})
14432#endif
14433
14434#ifdef __LITTLE_ENDIAN__
14435#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14436  float32x2_t __s0 = __p0; \
14437  float32x2_t __s1 = __p1; \
14438  float32x2_t __ret; \
14439  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14440  __ret; \
14441})
14442#else
14443#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14444  float32x2_t __s0 = __p0; \
14445  float32x2_t __s1 = __p1; \
14446  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14447  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14448  float32x2_t __ret; \
14449  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14451  __ret; \
14452})
14453#endif
14454
14455#ifdef __LITTLE_ENDIAN__
14456#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14457  int32x2_t __s0 = __p0; \
14458  int32x2_t __s1 = __p1; \
14459  int32x2_t __ret; \
14460  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14461  __ret; \
14462})
14463#else
14464#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14465  int32x2_t __s0 = __p0; \
14466  int32x2_t __s1 = __p1; \
14467  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14468  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14469  int32x2_t __ret; \
14470  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14471  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14472  __ret; \
14473})
14474#endif
14475
14476#ifdef __LITTLE_ENDIAN__
14477#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14478  int16x4_t __s0 = __p0; \
14479  int16x4_t __s1 = __p1; \
14480  int16x4_t __ret; \
14481  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14482  __ret; \
14483})
14484#else
14485#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14486  int16x4_t __s0 = __p0; \
14487  int16x4_t __s1 = __p1; \
14488  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14489  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14490  int16x4_t __ret; \
14491  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14492  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14493  __ret; \
14494})
14495#endif
14496
14497#ifdef __LITTLE_ENDIAN__
14498__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14499  uint32x4_t __ret;
14500  __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14501  return __ret;
14502}
14503#else
14504__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14505  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14506  uint32x4_t __ret;
14507  __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14509  return __ret;
14510}
14511#endif
14512
14513#ifdef __LITTLE_ENDIAN__
14514__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14515  uint16x8_t __ret;
14516  __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14517  return __ret;
14518}
14519#else
14520__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14521  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14522  uint16x8_t __ret;
14523  __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14524  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14525  return __ret;
14526}
14527#endif
14528
14529#ifdef __LITTLE_ENDIAN__
14530__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14531  float32x4_t __ret;
14532  __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14533  return __ret;
14534}
14535#else
14536__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14537  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14538  float32x4_t __ret;
14539  __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14540  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14541  return __ret;
14542}
14543#endif
14544
14545#ifdef __LITTLE_ENDIAN__
14546__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14547  int32x4_t __ret;
14548  __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14549  return __ret;
14550}
14551#else
14552__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14553  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14554  int32x4_t __ret;
14555  __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14556  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14557  return __ret;
14558}
14559#endif
14560
14561#ifdef __LITTLE_ENDIAN__
14562__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14563  int16x8_t __ret;
14564  __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14565  return __ret;
14566}
14567#else
14568__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14569  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14570  int16x8_t __ret;
14571  __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14572  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14573  return __ret;
14574}
14575#endif
14576
14577#ifdef __LITTLE_ENDIAN__
14578__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14579  uint32x2_t __ret;
14580  __ret = __p0 * (uint32x2_t) {__p1, __p1};
14581  return __ret;
14582}
14583#else
14584__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14585  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14586  uint32x2_t __ret;
14587  __ret = __rev0 * (uint32x2_t) {__p1, __p1};
14588  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14589  return __ret;
14590}
14591#endif
14592
14593#ifdef __LITTLE_ENDIAN__
14594__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14595  uint16x4_t __ret;
14596  __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14597  return __ret;
14598}
14599#else
14600__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14601  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14602  uint16x4_t __ret;
14603  __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14604  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14605  return __ret;
14606}
14607#endif
14608
14609#ifdef __LITTLE_ENDIAN__
14610__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14611  float32x2_t __ret;
14612  __ret = __p0 * (float32x2_t) {__p1, __p1};
14613  return __ret;
14614}
14615#else
14616__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14617  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14618  float32x2_t __ret;
14619  __ret = __rev0 * (float32x2_t) {__p1, __p1};
14620  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14621  return __ret;
14622}
14623#endif
14624
14625#ifdef __LITTLE_ENDIAN__
14626__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14627  int32x2_t __ret;
14628  __ret = __p0 * (int32x2_t) {__p1, __p1};
14629  return __ret;
14630}
14631#else
14632__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14633  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14634  int32x2_t __ret;
14635  __ret = __rev0 * (int32x2_t) {__p1, __p1};
14636  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14637  return __ret;
14638}
14639#endif
14640
14641#ifdef __LITTLE_ENDIAN__
14642__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14643  int16x4_t __ret;
14644  __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14645  return __ret;
14646}
14647#else
14648__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14649  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14650  int16x4_t __ret;
14651  __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14653  return __ret;
14654}
14655#endif
14656
14657#ifdef __LITTLE_ENDIAN__
14658__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14659  poly16x8_t __ret;
14660  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14661  return __ret;
14662}
14663#else
14664__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14665  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14666  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14667  poly16x8_t __ret;
14668  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
14669  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14670  return __ret;
14671}
14672__ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14673  poly16x8_t __ret;
14674  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14675  return __ret;
14676}
14677#endif
14678
14679#ifdef __LITTLE_ENDIAN__
14680__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14681  uint16x8_t __ret;
14682  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14683  return __ret;
14684}
14685#else
14686__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14687  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14688  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14689  uint16x8_t __ret;
14690  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
14691  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14692  return __ret;
14693}
14694__ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14695  uint16x8_t __ret;
14696  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14697  return __ret;
14698}
14699#endif
14700
14701#ifdef __LITTLE_ENDIAN__
14702__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14703  uint64x2_t __ret;
14704  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14705  return __ret;
14706}
14707#else
14708__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14709  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14710  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14711  uint64x2_t __ret;
14712  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
14713  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14714  return __ret;
14715}
14716__ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14717  uint64x2_t __ret;
14718  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14719  return __ret;
14720}
14721#endif
14722
14723#ifdef __LITTLE_ENDIAN__
14724__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14725  uint32x4_t __ret;
14726  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14727  return __ret;
14728}
14729#else
14730__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14731  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14732  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14733  uint32x4_t __ret;
14734  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
14735  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14736  return __ret;
14737}
14738__ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14739  uint32x4_t __ret;
14740  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14741  return __ret;
14742}
14743#endif
14744
14745#ifdef __LITTLE_ENDIAN__
14746__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14747  int16x8_t __ret;
14748  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14749  return __ret;
14750}
14751#else
14752__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14753  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14754  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14755  int16x8_t __ret;
14756  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
14757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14758  return __ret;
14759}
14760__ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14761  int16x8_t __ret;
14762  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14763  return __ret;
14764}
14765#endif
14766
14767#ifdef __LITTLE_ENDIAN__
14768__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14769  int64x2_t __ret;
14770  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14771  return __ret;
14772}
14773#else
14774__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14775  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14776  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14777  int64x2_t __ret;
14778  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
14779  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14780  return __ret;
14781}
14782__ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14783  int64x2_t __ret;
14784  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14785  return __ret;
14786}
14787#endif
14788
14789#ifdef __LITTLE_ENDIAN__
14790__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14791  int32x4_t __ret;
14792  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14793  return __ret;
14794}
14795#else
14796__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14797  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14798  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14799  int32x4_t __ret;
14800  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
14801  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14802  return __ret;
14803}
14804__ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14805  int32x4_t __ret;
14806  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14807  return __ret;
14808}
14809#endif
14810
14811#ifdef __LITTLE_ENDIAN__
14812#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14813  uint32x2_t __s0 = __p0; \
14814  uint32x2_t __s1 = __p1; \
14815  uint64x2_t __ret; \
14816  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14817  __ret; \
14818})
14819#else
14820#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14821  uint32x2_t __s0 = __p0; \
14822  uint32x2_t __s1 = __p1; \
14823  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14824  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14825  uint64x2_t __ret; \
14826  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14828  __ret; \
14829})
14830#endif
14831
14832#ifdef __LITTLE_ENDIAN__
14833#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14834  uint16x4_t __s0 = __p0; \
14835  uint16x4_t __s1 = __p1; \
14836  uint32x4_t __ret; \
14837  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14838  __ret; \
14839})
14840#else
14841#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14842  uint16x4_t __s0 = __p0; \
14843  uint16x4_t __s1 = __p1; \
14844  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14845  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14846  uint32x4_t __ret; \
14847  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14848  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14849  __ret; \
14850})
14851#endif
14852
14853#ifdef __LITTLE_ENDIAN__
14854#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14855  int32x2_t __s0 = __p0; \
14856  int32x2_t __s1 = __p1; \
14857  int64x2_t __ret; \
14858  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14859  __ret; \
14860})
14861#else
14862#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14863  int32x2_t __s0 = __p0; \
14864  int32x2_t __s1 = __p1; \
14865  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14866  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14867  int64x2_t __ret; \
14868  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14869  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14870  __ret; \
14871})
14872#endif
14873
14874#ifdef __LITTLE_ENDIAN__
14875#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14876  int16x4_t __s0 = __p0; \
14877  int16x4_t __s1 = __p1; \
14878  int32x4_t __ret; \
14879  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14880  __ret; \
14881})
14882#else
14883#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14884  int16x4_t __s0 = __p0; \
14885  int16x4_t __s1 = __p1; \
14886  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14887  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14888  int32x4_t __ret; \
14889  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14890  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14891  __ret; \
14892})
14893#endif
14894
14895#ifdef __LITTLE_ENDIAN__
14896__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14897  uint64x2_t __ret;
14898  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14899  return __ret;
14900}
14901#else
14902__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14903  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14904  uint64x2_t __ret;
14905  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14906  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14907  return __ret;
14908}
14909__ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14910  uint64x2_t __ret;
14911  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14912  return __ret;
14913}
14914#endif
14915
14916#ifdef __LITTLE_ENDIAN__
14917__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
14918  uint32x4_t __ret;
14919  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
14920  return __ret;
14921}
14922#else
14923__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
14924  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14925  uint32x4_t __ret;
14926  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
14927  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14928  return __ret;
14929}
14930__ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
14931  uint32x4_t __ret;
14932  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
14933  return __ret;
14934}
14935#endif
14936
14937#ifdef __LITTLE_ENDIAN__
14938__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
14939  int64x2_t __ret;
14940  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
14941  return __ret;
14942}
14943#else
14944__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
14945  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14946  int64x2_t __ret;
14947  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
14948  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14949  return __ret;
14950}
14951__ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
14952  int64x2_t __ret;
14953  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
14954  return __ret;
14955}
14956#endif
14957
14958#ifdef __LITTLE_ENDIAN__
14959__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
14960  int32x4_t __ret;
14961  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
14962  return __ret;
14963}
14964#else
14965__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
14966  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14967  int32x4_t __ret;
14968  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
14969  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14970  return __ret;
14971}
14972__ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
14973  int32x4_t __ret;
14974  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
14975  return __ret;
14976}
14977#endif
14978
14979#ifdef __LITTLE_ENDIAN__
14980__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
14981  poly8x8_t __ret;
14982  __ret = ~__p0;
14983  return __ret;
14984}
14985#else
14986__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
14987  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14988  poly8x8_t __ret;
14989  __ret = ~__rev0;
14990  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14991  return __ret;
14992}
14993#endif
14994
14995#ifdef __LITTLE_ENDIAN__
14996__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
14997  poly8x16_t __ret;
14998  __ret = ~__p0;
14999  return __ret;
15000}
15001#else
15002__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15003  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15004  poly8x16_t __ret;
15005  __ret = ~__rev0;
15006  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15007  return __ret;
15008}
15009#endif
15010
15011#ifdef __LITTLE_ENDIAN__
15012__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15013  uint8x16_t __ret;
15014  __ret = ~__p0;
15015  return __ret;
15016}
15017#else
15018__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15019  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15020  uint8x16_t __ret;
15021  __ret = ~__rev0;
15022  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15023  return __ret;
15024}
15025#endif
15026
15027#ifdef __LITTLE_ENDIAN__
15028__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15029  uint32x4_t __ret;
15030  __ret = ~__p0;
15031  return __ret;
15032}
15033#else
15034__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15035  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15036  uint32x4_t __ret;
15037  __ret = ~__rev0;
15038  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15039  return __ret;
15040}
15041#endif
15042
15043#ifdef __LITTLE_ENDIAN__
15044__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15045  uint16x8_t __ret;
15046  __ret = ~__p0;
15047  return __ret;
15048}
15049#else
15050__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15051  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15052  uint16x8_t __ret;
15053  __ret = ~__rev0;
15054  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15055  return __ret;
15056}
15057#endif
15058
15059#ifdef __LITTLE_ENDIAN__
15060__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15061  int8x16_t __ret;
15062  __ret = ~__p0;
15063  return __ret;
15064}
15065#else
15066__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15067  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15068  int8x16_t __ret;
15069  __ret = ~__rev0;
15070  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15071  return __ret;
15072}
15073#endif
15074
15075#ifdef __LITTLE_ENDIAN__
15076__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15077  int32x4_t __ret;
15078  __ret = ~__p0;
15079  return __ret;
15080}
15081#else
15082__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15083  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15084  int32x4_t __ret;
15085  __ret = ~__rev0;
15086  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15087  return __ret;
15088}
15089#endif
15090
15091#ifdef __LITTLE_ENDIAN__
15092__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15093  int16x8_t __ret;
15094  __ret = ~__p0;
15095  return __ret;
15096}
15097#else
15098__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15099  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15100  int16x8_t __ret;
15101  __ret = ~__rev0;
15102  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15103  return __ret;
15104}
15105#endif
15106
15107#ifdef __LITTLE_ENDIAN__
15108__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15109  uint8x8_t __ret;
15110  __ret = ~__p0;
15111  return __ret;
15112}
15113#else
15114__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15115  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15116  uint8x8_t __ret;
15117  __ret = ~__rev0;
15118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15119  return __ret;
15120}
15121#endif
15122
15123#ifdef __LITTLE_ENDIAN__
15124__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15125  uint32x2_t __ret;
15126  __ret = ~__p0;
15127  return __ret;
15128}
15129#else
15130__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15131  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15132  uint32x2_t __ret;
15133  __ret = ~__rev0;
15134  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15135  return __ret;
15136}
15137#endif
15138
15139#ifdef __LITTLE_ENDIAN__
15140__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15141  uint16x4_t __ret;
15142  __ret = ~__p0;
15143  return __ret;
15144}
15145#else
15146__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15147  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15148  uint16x4_t __ret;
15149  __ret = ~__rev0;
15150  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15151  return __ret;
15152}
15153#endif
15154
15155#ifdef __LITTLE_ENDIAN__
15156__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15157  int8x8_t __ret;
15158  __ret = ~__p0;
15159  return __ret;
15160}
15161#else
15162__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15163  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15164  int8x8_t __ret;
15165  __ret = ~__rev0;
15166  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15167  return __ret;
15168}
15169#endif
15170
15171#ifdef __LITTLE_ENDIAN__
15172__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15173  int32x2_t __ret;
15174  __ret = ~__p0;
15175  return __ret;
15176}
15177#else
15178__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15179  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15180  int32x2_t __ret;
15181  __ret = ~__rev0;
15182  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15183  return __ret;
15184}
15185#endif
15186
15187#ifdef __LITTLE_ENDIAN__
15188__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15189  int16x4_t __ret;
15190  __ret = ~__p0;
15191  return __ret;
15192}
15193#else
15194__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15195  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15196  int16x4_t __ret;
15197  __ret = ~__rev0;
15198  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15199  return __ret;
15200}
15201#endif
15202
15203#ifdef __LITTLE_ENDIAN__
15204__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15205  int8x16_t __ret;
15206  __ret = -__p0;
15207  return __ret;
15208}
15209#else
15210__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15211  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15212  int8x16_t __ret;
15213  __ret = -__rev0;
15214  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15215  return __ret;
15216}
15217#endif
15218
15219#ifdef __LITTLE_ENDIAN__
15220__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15221  float32x4_t __ret;
15222  __ret = -__p0;
15223  return __ret;
15224}
15225#else
15226__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15227  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15228  float32x4_t __ret;
15229  __ret = -__rev0;
15230  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15231  return __ret;
15232}
15233#endif
15234
15235#ifdef __LITTLE_ENDIAN__
15236__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15237  int32x4_t __ret;
15238  __ret = -__p0;
15239  return __ret;
15240}
15241#else
15242__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15243  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15244  int32x4_t __ret;
15245  __ret = -__rev0;
15246  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15247  return __ret;
15248}
15249#endif
15250
15251#ifdef __LITTLE_ENDIAN__
15252__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15253  int16x8_t __ret;
15254  __ret = -__p0;
15255  return __ret;
15256}
15257#else
15258__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15259  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15260  int16x8_t __ret;
15261  __ret = -__rev0;
15262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15263  return __ret;
15264}
15265#endif
15266
15267#ifdef __LITTLE_ENDIAN__
15268__ai int8x8_t vneg_s8(int8x8_t __p0) {
15269  int8x8_t __ret;
15270  __ret = -__p0;
15271  return __ret;
15272}
15273#else
15274__ai int8x8_t vneg_s8(int8x8_t __p0) {
15275  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15276  int8x8_t __ret;
15277  __ret = -__rev0;
15278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15279  return __ret;
15280}
15281#endif
15282
15283#ifdef __LITTLE_ENDIAN__
15284__ai float32x2_t vneg_f32(float32x2_t __p0) {
15285  float32x2_t __ret;
15286  __ret = -__p0;
15287  return __ret;
15288}
15289#else
15290__ai float32x2_t vneg_f32(float32x2_t __p0) {
15291  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15292  float32x2_t __ret;
15293  __ret = -__rev0;
15294  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15295  return __ret;
15296}
15297#endif
15298
15299#ifdef __LITTLE_ENDIAN__
15300__ai int32x2_t vneg_s32(int32x2_t __p0) {
15301  int32x2_t __ret;
15302  __ret = -__p0;
15303  return __ret;
15304}
15305#else
15306__ai int32x2_t vneg_s32(int32x2_t __p0) {
15307  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15308  int32x2_t __ret;
15309  __ret = -__rev0;
15310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15311  return __ret;
15312}
15313#endif
15314
15315#ifdef __LITTLE_ENDIAN__
15316__ai int16x4_t vneg_s16(int16x4_t __p0) {
15317  int16x4_t __ret;
15318  __ret = -__p0;
15319  return __ret;
15320}
15321#else
15322__ai int16x4_t vneg_s16(int16x4_t __p0) {
15323  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15324  int16x4_t __ret;
15325  __ret = -__rev0;
15326  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15327  return __ret;
15328}
15329#endif
15330
15331#ifdef __LITTLE_ENDIAN__
15332__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15333  uint8x16_t __ret;
15334  __ret = __p0 | ~__p1;
15335  return __ret;
15336}
15337#else
15338__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15339  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15340  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15341  uint8x16_t __ret;
15342  __ret = __rev0 | ~__rev1;
15343  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15344  return __ret;
15345}
15346#endif
15347
15348#ifdef __LITTLE_ENDIAN__
15349__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15350  uint32x4_t __ret;
15351  __ret = __p0 | ~__p1;
15352  return __ret;
15353}
15354#else
15355__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15356  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15357  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15358  uint32x4_t __ret;
15359  __ret = __rev0 | ~__rev1;
15360  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15361  return __ret;
15362}
15363#endif
15364
15365#ifdef __LITTLE_ENDIAN__
15366__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15367  uint64x2_t __ret;
15368  __ret = __p0 | ~__p1;
15369  return __ret;
15370}
15371#else
15372__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15373  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15374  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15375  uint64x2_t __ret;
15376  __ret = __rev0 | ~__rev1;
15377  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15378  return __ret;
15379}
15380#endif
15381
15382#ifdef __LITTLE_ENDIAN__
15383__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15384  uint16x8_t __ret;
15385  __ret = __p0 | ~__p1;
15386  return __ret;
15387}
15388#else
15389__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15390  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15391  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15392  uint16x8_t __ret;
15393  __ret = __rev0 | ~__rev1;
15394  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15395  return __ret;
15396}
15397#endif
15398
15399#ifdef __LITTLE_ENDIAN__
15400__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15401  int8x16_t __ret;
15402  __ret = __p0 | ~__p1;
15403  return __ret;
15404}
15405#else
15406__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15407  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15408  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15409  int8x16_t __ret;
15410  __ret = __rev0 | ~__rev1;
15411  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15412  return __ret;
15413}
15414#endif
15415
15416#ifdef __LITTLE_ENDIAN__
15417__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15418  int32x4_t __ret;
15419  __ret = __p0 | ~__p1;
15420  return __ret;
15421}
15422#else
15423__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15424  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15425  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15426  int32x4_t __ret;
15427  __ret = __rev0 | ~__rev1;
15428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15429  return __ret;
15430}
15431#endif
15432
15433#ifdef __LITTLE_ENDIAN__
15434__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15435  int64x2_t __ret;
15436  __ret = __p0 | ~__p1;
15437  return __ret;
15438}
15439#else
15440__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15441  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15442  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15443  int64x2_t __ret;
15444  __ret = __rev0 | ~__rev1;
15445  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15446  return __ret;
15447}
15448#endif
15449
15450#ifdef __LITTLE_ENDIAN__
15451__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15452  int16x8_t __ret;
15453  __ret = __p0 | ~__p1;
15454  return __ret;
15455}
15456#else
15457__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15458  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15459  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15460  int16x8_t __ret;
15461  __ret = __rev0 | ~__rev1;
15462  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15463  return __ret;
15464}
15465#endif
15466
15467#ifdef __LITTLE_ENDIAN__
15468__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15469  uint8x8_t __ret;
15470  __ret = __p0 | ~__p1;
15471  return __ret;
15472}
15473#else
15474__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15475  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15476  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15477  uint8x8_t __ret;
15478  __ret = __rev0 | ~__rev1;
15479  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15480  return __ret;
15481}
15482#endif
15483
15484#ifdef __LITTLE_ENDIAN__
15485__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15486  uint32x2_t __ret;
15487  __ret = __p0 | ~__p1;
15488  return __ret;
15489}
15490#else
15491__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15492  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15493  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15494  uint32x2_t __ret;
15495  __ret = __rev0 | ~__rev1;
15496  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15497  return __ret;
15498}
15499#endif
15500
15501#ifdef __LITTLE_ENDIAN__
15502__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15503  uint64x1_t __ret;
15504  __ret = __p0 | ~__p1;
15505  return __ret;
15506}
15507#else
15508__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15509  uint64x1_t __ret;
15510  __ret = __p0 | ~__p1;
15511  return __ret;
15512}
15513#endif
15514
15515#ifdef __LITTLE_ENDIAN__
15516__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15517  uint16x4_t __ret;
15518  __ret = __p0 | ~__p1;
15519  return __ret;
15520}
15521#else
15522__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15523  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15524  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15525  uint16x4_t __ret;
15526  __ret = __rev0 | ~__rev1;
15527  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15528  return __ret;
15529}
15530#endif
15531
15532#ifdef __LITTLE_ENDIAN__
15533__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15534  int8x8_t __ret;
15535  __ret = __p0 | ~__p1;
15536  return __ret;
15537}
15538#else
15539__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15540  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15541  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15542  int8x8_t __ret;
15543  __ret = __rev0 | ~__rev1;
15544  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15545  return __ret;
15546}
15547#endif
15548
15549#ifdef __LITTLE_ENDIAN__
15550__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15551  int32x2_t __ret;
15552  __ret = __p0 | ~__p1;
15553  return __ret;
15554}
15555#else
15556__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15557  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15558  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15559  int32x2_t __ret;
15560  __ret = __rev0 | ~__rev1;
15561  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15562  return __ret;
15563}
15564#endif
15565
15566#ifdef __LITTLE_ENDIAN__
15567__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15568  int64x1_t __ret;
15569  __ret = __p0 | ~__p1;
15570  return __ret;
15571}
15572#else
15573__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15574  int64x1_t __ret;
15575  __ret = __p0 | ~__p1;
15576  return __ret;
15577}
15578#endif
15579
15580#ifdef __LITTLE_ENDIAN__
15581__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15582  int16x4_t __ret;
15583  __ret = __p0 | ~__p1;
15584  return __ret;
15585}
15586#else
15587__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15588  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15589  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15590  int16x4_t __ret;
15591  __ret = __rev0 | ~__rev1;
15592  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15593  return __ret;
15594}
15595#endif
15596
15597#ifdef __LITTLE_ENDIAN__
15598__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15599  uint8x16_t __ret;
15600  __ret = __p0 | __p1;
15601  return __ret;
15602}
15603#else
15604__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15605  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15606  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15607  uint8x16_t __ret;
15608  __ret = __rev0 | __rev1;
15609  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15610  return __ret;
15611}
15612#endif
15613
15614#ifdef __LITTLE_ENDIAN__
15615__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15616  uint32x4_t __ret;
15617  __ret = __p0 | __p1;
15618  return __ret;
15619}
15620#else
15621__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15622  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15623  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15624  uint32x4_t __ret;
15625  __ret = __rev0 | __rev1;
15626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15627  return __ret;
15628}
15629#endif
15630
15631#ifdef __LITTLE_ENDIAN__
15632__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15633  uint64x2_t __ret;
15634  __ret = __p0 | __p1;
15635  return __ret;
15636}
15637#else
15638__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15639  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15640  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15641  uint64x2_t __ret;
15642  __ret = __rev0 | __rev1;
15643  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15644  return __ret;
15645}
15646#endif
15647
15648#ifdef __LITTLE_ENDIAN__
15649__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15650  uint16x8_t __ret;
15651  __ret = __p0 | __p1;
15652  return __ret;
15653}
15654#else
15655__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15656  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15657  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15658  uint16x8_t __ret;
15659  __ret = __rev0 | __rev1;
15660  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15661  return __ret;
15662}
15663#endif
15664
15665#ifdef __LITTLE_ENDIAN__
15666__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15667  int8x16_t __ret;
15668  __ret = __p0 | __p1;
15669  return __ret;
15670}
15671#else
15672__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15673  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15674  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15675  int8x16_t __ret;
15676  __ret = __rev0 | __rev1;
15677  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15678  return __ret;
15679}
15680#endif
15681
15682#ifdef __LITTLE_ENDIAN__
15683__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15684  int32x4_t __ret;
15685  __ret = __p0 | __p1;
15686  return __ret;
15687}
15688#else
15689__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15690  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15691  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15692  int32x4_t __ret;
15693  __ret = __rev0 | __rev1;
15694  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15695  return __ret;
15696}
15697#endif
15698
15699#ifdef __LITTLE_ENDIAN__
15700__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15701  int64x2_t __ret;
15702  __ret = __p0 | __p1;
15703  return __ret;
15704}
15705#else
15706__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15707  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15708  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15709  int64x2_t __ret;
15710  __ret = __rev0 | __rev1;
15711  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15712  return __ret;
15713}
15714#endif
15715
15716#ifdef __LITTLE_ENDIAN__
15717__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15718  int16x8_t __ret;
15719  __ret = __p0 | __p1;
15720  return __ret;
15721}
15722#else
15723__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15724  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15725  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15726  int16x8_t __ret;
15727  __ret = __rev0 | __rev1;
15728  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15729  return __ret;
15730}
15731#endif
15732
15733#ifdef __LITTLE_ENDIAN__
15734__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15735  uint8x8_t __ret;
15736  __ret = __p0 | __p1;
15737  return __ret;
15738}
15739#else
15740__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15741  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15742  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15743  uint8x8_t __ret;
15744  __ret = __rev0 | __rev1;
15745  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15746  return __ret;
15747}
15748#endif
15749
15750#ifdef __LITTLE_ENDIAN__
15751__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15752  uint32x2_t __ret;
15753  __ret = __p0 | __p1;
15754  return __ret;
15755}
15756#else
15757__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15758  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15759  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15760  uint32x2_t __ret;
15761  __ret = __rev0 | __rev1;
15762  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15763  return __ret;
15764}
15765#endif
15766
15767#ifdef __LITTLE_ENDIAN__
15768__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15769  uint64x1_t __ret;
15770  __ret = __p0 | __p1;
15771  return __ret;
15772}
15773#else
15774__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15775  uint64x1_t __ret;
15776  __ret = __p0 | __p1;
15777  return __ret;
15778}
15779#endif
15780
15781#ifdef __LITTLE_ENDIAN__
15782__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15783  uint16x4_t __ret;
15784  __ret = __p0 | __p1;
15785  return __ret;
15786}
15787#else
15788__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15789  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15790  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15791  uint16x4_t __ret;
15792  __ret = __rev0 | __rev1;
15793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15794  return __ret;
15795}
15796#endif
15797
15798#ifdef __LITTLE_ENDIAN__
15799__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15800  int8x8_t __ret;
15801  __ret = __p0 | __p1;
15802  return __ret;
15803}
15804#else
15805__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15806  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15807  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15808  int8x8_t __ret;
15809  __ret = __rev0 | __rev1;
15810  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15811  return __ret;
15812}
15813#endif
15814
15815#ifdef __LITTLE_ENDIAN__
15816__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15817  int32x2_t __ret;
15818  __ret = __p0 | __p1;
15819  return __ret;
15820}
15821#else
15822__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15823  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15824  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15825  int32x2_t __ret;
15826  __ret = __rev0 | __rev1;
15827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15828  return __ret;
15829}
15830#endif
15831
15832#ifdef __LITTLE_ENDIAN__
15833__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15834  int64x1_t __ret;
15835  __ret = __p0 | __p1;
15836  return __ret;
15837}
15838#else
15839__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15840  int64x1_t __ret;
15841  __ret = __p0 | __p1;
15842  return __ret;
15843}
15844#endif
15845
15846#ifdef __LITTLE_ENDIAN__
15847__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15848  int16x4_t __ret;
15849  __ret = __p0 | __p1;
15850  return __ret;
15851}
15852#else
15853__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15854  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15855  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15856  int16x4_t __ret;
15857  __ret = __rev0 | __rev1;
15858  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15859  return __ret;
15860}
15861#endif
15862
15863#ifdef __LITTLE_ENDIAN__
15864__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15865  uint16x8_t __ret;
15866  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
15867  return __ret;
15868}
15869#else
15870__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15871  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15872  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15873  uint16x8_t __ret;
15874  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
15875  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15876  return __ret;
15877}
15878#endif
15879
15880#ifdef __LITTLE_ENDIAN__
15881__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15882  uint64x2_t __ret;
15883  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
15884  return __ret;
15885}
15886#else
15887__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15888  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15889  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15890  uint64x2_t __ret;
15891  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
15892  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15893  return __ret;
15894}
15895#endif
15896
15897#ifdef __LITTLE_ENDIAN__
15898__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15899  uint32x4_t __ret;
15900  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
15901  return __ret;
15902}
15903#else
15904__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15905  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15906  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15907  uint32x4_t __ret;
15908  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
15909  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15910  return __ret;
15911}
15912#endif
15913
15914#ifdef __LITTLE_ENDIAN__
15915__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
15916  int16x8_t __ret;
15917  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
15918  return __ret;
15919}
15920#else
15921__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
15922  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15923  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15924  int16x8_t __ret;
15925  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
15926  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15927  return __ret;
15928}
15929#endif
15930
15931#ifdef __LITTLE_ENDIAN__
15932__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
15933  int64x2_t __ret;
15934  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
15935  return __ret;
15936}
15937#else
15938__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
15939  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15940  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15941  int64x2_t __ret;
15942  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
15943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15944  return __ret;
15945}
15946#endif
15947
15948#ifdef __LITTLE_ENDIAN__
15949__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
15950  int32x4_t __ret;
15951  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
15952  return __ret;
15953}
15954#else
15955__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
15956  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15957  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15958  int32x4_t __ret;
15959  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
15960  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15961  return __ret;
15962}
15963#endif
15964
15965#ifdef __LITTLE_ENDIAN__
15966__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
15967  uint16x4_t __ret;
15968  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
15969  return __ret;
15970}
15971#else
15972__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
15973  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15974  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15975  uint16x4_t __ret;
15976  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
15977  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15978  return __ret;
15979}
15980#endif
15981
15982#ifdef __LITTLE_ENDIAN__
15983__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
15984  uint64x1_t __ret;
15985  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
15986  return __ret;
15987}
15988#else
15989__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
15990  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15991  uint64x1_t __ret;
15992  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
15993  return __ret;
15994}
15995#endif
15996
15997#ifdef __LITTLE_ENDIAN__
15998__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
15999  uint32x2_t __ret;
16000  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16001  return __ret;
16002}
16003#else
16004__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16005  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16006  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16007  uint32x2_t __ret;
16008  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16010  return __ret;
16011}
16012#endif
16013
16014#ifdef __LITTLE_ENDIAN__
16015__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16016  int16x4_t __ret;
16017  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16018  return __ret;
16019}
16020#else
16021__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16022  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16023  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16024  int16x4_t __ret;
16025  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16026  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16027  return __ret;
16028}
16029#endif
16030
16031#ifdef __LITTLE_ENDIAN__
16032__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16033  int64x1_t __ret;
16034  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16035  return __ret;
16036}
16037#else
16038__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16039  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16040  int64x1_t __ret;
16041  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
16042  return __ret;
16043}
16044#endif
16045
16046#ifdef __LITTLE_ENDIAN__
16047__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16048  int32x2_t __ret;
16049  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16050  return __ret;
16051}
16052#else
16053__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16054  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16055  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16056  int32x2_t __ret;
16057  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16058  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16059  return __ret;
16060}
16061#endif
16062
16063#ifdef __LITTLE_ENDIAN__
16064__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16065  uint8x8_t __ret;
16066  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16067  return __ret;
16068}
16069#else
16070__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16071  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16072  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16073  uint8x8_t __ret;
16074  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16075  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16076  return __ret;
16077}
16078#endif
16079
16080#ifdef __LITTLE_ENDIAN__
16081__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16082  uint32x2_t __ret;
16083  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16084  return __ret;
16085}
16086#else
16087__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16088  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16089  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16090  uint32x2_t __ret;
16091  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16093  return __ret;
16094}
16095#endif
16096
16097#ifdef __LITTLE_ENDIAN__
16098__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16099  uint16x4_t __ret;
16100  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16101  return __ret;
16102}
16103#else
16104__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16105  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16106  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16107  uint16x4_t __ret;
16108  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16109  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16110  return __ret;
16111}
16112#endif
16113
16114#ifdef __LITTLE_ENDIAN__
16115__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16116  int8x8_t __ret;
16117  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16118  return __ret;
16119}
16120#else
16121__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16122  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16123  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16124  int8x8_t __ret;
16125  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16126  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16127  return __ret;
16128}
16129#endif
16130
16131#ifdef __LITTLE_ENDIAN__
16132__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16133  float32x2_t __ret;
16134  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16135  return __ret;
16136}
16137#else
16138__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16139  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16140  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16141  float32x2_t __ret;
16142  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16144  return __ret;
16145}
16146#endif
16147
16148#ifdef __LITTLE_ENDIAN__
16149__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16150  int32x2_t __ret;
16151  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16152  return __ret;
16153}
16154#else
16155__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16156  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16157  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16158  int32x2_t __ret;
16159  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16160  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16161  return __ret;
16162}
16163#endif
16164
16165#ifdef __LITTLE_ENDIAN__
16166__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16167  int16x4_t __ret;
16168  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16169  return __ret;
16170}
16171#else
16172__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16173  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16174  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16175  int16x4_t __ret;
16176  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16177  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16178  return __ret;
16179}
16180#endif
16181
16182#ifdef __LITTLE_ENDIAN__
16183__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16184  uint16x8_t __ret;
16185  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
16186  return __ret;
16187}
16188#else
16189__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16190  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16191  uint16x8_t __ret;
16192  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
16193  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16194  return __ret;
16195}
16196#endif
16197
16198#ifdef __LITTLE_ENDIAN__
16199__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16200  uint64x2_t __ret;
16201  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
16202  return __ret;
16203}
16204#else
16205__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16206  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16207  uint64x2_t __ret;
16208  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
16209  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16210  return __ret;
16211}
16212#endif
16213
16214#ifdef __LITTLE_ENDIAN__
16215__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16216  uint32x4_t __ret;
16217  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
16218  return __ret;
16219}
16220#else
16221__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16222  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16223  uint32x4_t __ret;
16224  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
16225  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16226  return __ret;
16227}
16228#endif
16229
16230#ifdef __LITTLE_ENDIAN__
16231__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16232  int16x8_t __ret;
16233  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
16234  return __ret;
16235}
16236#else
16237__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16238  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16239  int16x8_t __ret;
16240  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
16241  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16242  return __ret;
16243}
16244#endif
16245
16246#ifdef __LITTLE_ENDIAN__
16247__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16248  int64x2_t __ret;
16249  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
16250  return __ret;
16251}
16252#else
16253__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16254  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16255  int64x2_t __ret;
16256  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
16257  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16258  return __ret;
16259}
16260#endif
16261
16262#ifdef __LITTLE_ENDIAN__
16263__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16264  int32x4_t __ret;
16265  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
16266  return __ret;
16267}
16268#else
16269__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16270  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16271  int32x4_t __ret;
16272  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
16273  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16274  return __ret;
16275}
16276#endif
16277
16278#ifdef __LITTLE_ENDIAN__
16279__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16280  uint16x4_t __ret;
16281  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
16282  return __ret;
16283}
16284#else
16285__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16286  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16287  uint16x4_t __ret;
16288  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
16289  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16290  return __ret;
16291}
16292#endif
16293
16294#ifdef __LITTLE_ENDIAN__
16295__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16296  uint64x1_t __ret;
16297  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
16298  return __ret;
16299}
16300#else
16301__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16302  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16303  uint64x1_t __ret;
16304  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
16305  return __ret;
16306}
16307#endif
16308
16309#ifdef __LITTLE_ENDIAN__
16310__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16311  uint32x2_t __ret;
16312  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
16313  return __ret;
16314}
16315#else
16316__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16317  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16318  uint32x2_t __ret;
16319  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
16320  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16321  return __ret;
16322}
16323#endif
16324
16325#ifdef __LITTLE_ENDIAN__
16326__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16327  int16x4_t __ret;
16328  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
16329  return __ret;
16330}
16331#else
16332__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16333  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16334  int16x4_t __ret;
16335  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
16336  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16337  return __ret;
16338}
16339#endif
16340
16341#ifdef __LITTLE_ENDIAN__
16342__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16343  int64x1_t __ret;
16344  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
16345  return __ret;
16346}
16347#else
16348__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16349  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16350  int64x1_t __ret;
16351  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
16352  return __ret;
16353}
16354#endif
16355
16356#ifdef __LITTLE_ENDIAN__
16357__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16358  int32x2_t __ret;
16359  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
16360  return __ret;
16361}
16362#else
16363__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16364  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16365  int32x2_t __ret;
16366  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
16367  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16368  return __ret;
16369}
16370#endif
16371
16372#ifdef __LITTLE_ENDIAN__
16373__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16374  uint8x8_t __ret;
16375  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16376  return __ret;
16377}
16378#else
16379__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16380  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16381  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16382  uint8x8_t __ret;
16383  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16384  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16385  return __ret;
16386}
16387#endif
16388
16389#ifdef __LITTLE_ENDIAN__
16390__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16391  uint32x2_t __ret;
16392  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16393  return __ret;
16394}
16395#else
16396__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16397  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16398  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16399  uint32x2_t __ret;
16400  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16401  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16402  return __ret;
16403}
16404#endif
16405
16406#ifdef __LITTLE_ENDIAN__
16407__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16408  uint16x4_t __ret;
16409  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16410  return __ret;
16411}
16412#else
16413__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16414  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16415  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16416  uint16x4_t __ret;
16417  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16419  return __ret;
16420}
16421#endif
16422
16423#ifdef __LITTLE_ENDIAN__
16424__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16425  int8x8_t __ret;
16426  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16427  return __ret;
16428}
16429#else
16430__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16431  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16432  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16433  int8x8_t __ret;
16434  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16435  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16436  return __ret;
16437}
16438#endif
16439
16440#ifdef __LITTLE_ENDIAN__
16441__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16442  float32x2_t __ret;
16443  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16444  return __ret;
16445}
16446#else
16447__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16448  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16449  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16450  float32x2_t __ret;
16451  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16452  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16453  return __ret;
16454}
16455#endif
16456
16457#ifdef __LITTLE_ENDIAN__
16458__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16459  int32x2_t __ret;
16460  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16461  return __ret;
16462}
16463#else
16464__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16465  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16466  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16467  int32x2_t __ret;
16468  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16470  return __ret;
16471}
16472#endif
16473
16474#ifdef __LITTLE_ENDIAN__
16475__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16476  int16x4_t __ret;
16477  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16478  return __ret;
16479}
16480#else
16481__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16482  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16483  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16484  int16x4_t __ret;
16485  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16487  return __ret;
16488}
16489#endif
16490
16491#ifdef __LITTLE_ENDIAN__
16492__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16493  uint8x8_t __ret;
16494  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16495  return __ret;
16496}
16497#else
16498__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16499  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16500  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16501  uint8x8_t __ret;
16502  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16504  return __ret;
16505}
16506#endif
16507
16508#ifdef __LITTLE_ENDIAN__
16509__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16510  uint32x2_t __ret;
16511  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16512  return __ret;
16513}
16514#else
16515__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16516  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16517  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16518  uint32x2_t __ret;
16519  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16520  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16521  return __ret;
16522}
16523#endif
16524
16525#ifdef __LITTLE_ENDIAN__
16526__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16527  uint16x4_t __ret;
16528  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16529  return __ret;
16530}
16531#else
16532__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16533  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16534  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16535  uint16x4_t __ret;
16536  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16537  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16538  return __ret;
16539}
16540#endif
16541
16542#ifdef __LITTLE_ENDIAN__
16543__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16544  int8x8_t __ret;
16545  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16546  return __ret;
16547}
16548#else
16549__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16550  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16551  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16552  int8x8_t __ret;
16553  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16554  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16555  return __ret;
16556}
16557#endif
16558
16559#ifdef __LITTLE_ENDIAN__
16560__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16561  float32x2_t __ret;
16562  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16563  return __ret;
16564}
16565#else
16566__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16567  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16568  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16569  float32x2_t __ret;
16570  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16572  return __ret;
16573}
16574#endif
16575
16576#ifdef __LITTLE_ENDIAN__
16577__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16578  int32x2_t __ret;
16579  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16580  return __ret;
16581}
16582#else
16583__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16584  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16585  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16586  int32x2_t __ret;
16587  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16588  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16589  return __ret;
16590}
16591#endif
16592
16593#ifdef __LITTLE_ENDIAN__
16594__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16595  int16x4_t __ret;
16596  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16597  return __ret;
16598}
16599#else
16600__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16601  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16602  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16603  int16x4_t __ret;
16604  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16605  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16606  return __ret;
16607}
16608#endif
16609
16610#ifdef __LITTLE_ENDIAN__
16611__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16612  int8x16_t __ret;
16613  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
16614  return __ret;
16615}
16616#else
16617__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16618  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16619  int8x16_t __ret;
16620  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
16621  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16622  return __ret;
16623}
16624#endif
16625
16626#ifdef __LITTLE_ENDIAN__
16627__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16628  int32x4_t __ret;
16629  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
16630  return __ret;
16631}
16632#else
16633__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16634  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16635  int32x4_t __ret;
16636  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
16637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16638  return __ret;
16639}
16640#endif
16641
16642#ifdef __LITTLE_ENDIAN__
16643__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16644  int16x8_t __ret;
16645  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
16646  return __ret;
16647}
16648#else
16649__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16650  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16651  int16x8_t __ret;
16652  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
16653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16654  return __ret;
16655}
16656#endif
16657
16658#ifdef __LITTLE_ENDIAN__
16659__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16660  int8x8_t __ret;
16661  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
16662  return __ret;
16663}
16664#else
16665__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16666  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16667  int8x8_t __ret;
16668  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
16669  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16670  return __ret;
16671}
16672#endif
16673
16674#ifdef __LITTLE_ENDIAN__
16675__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16676  int32x2_t __ret;
16677  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
16678  return __ret;
16679}
16680#else
16681__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16682  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16683  int32x2_t __ret;
16684  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
16685  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16686  return __ret;
16687}
16688#endif
16689
16690#ifdef __LITTLE_ENDIAN__
16691__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16692  int16x4_t __ret;
16693  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
16694  return __ret;
16695}
16696#else
16697__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16698  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16699  int16x4_t __ret;
16700  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
16701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16702  return __ret;
16703}
16704#endif
16705
16706#ifdef __LITTLE_ENDIAN__
16707__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16708  uint8x16_t __ret;
16709  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
16710  return __ret;
16711}
16712#else
16713__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16714  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16715  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16716  uint8x16_t __ret;
16717  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
16718  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16719  return __ret;
16720}
16721#endif
16722
16723#ifdef __LITTLE_ENDIAN__
16724__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16725  uint32x4_t __ret;
16726  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
16727  return __ret;
16728}
16729#else
16730__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16731  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16732  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16733  uint32x4_t __ret;
16734  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
16735  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16736  return __ret;
16737}
16738#endif
16739
16740#ifdef __LITTLE_ENDIAN__
16741__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16742  uint64x2_t __ret;
16743  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
16744  return __ret;
16745}
16746#else
16747__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16748  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16749  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16750  uint64x2_t __ret;
16751  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
16752  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16753  return __ret;
16754}
16755#endif
16756
16757#ifdef __LITTLE_ENDIAN__
16758__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16759  uint16x8_t __ret;
16760  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
16761  return __ret;
16762}
16763#else
16764__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16765  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16766  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16767  uint16x8_t __ret;
16768  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
16769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16770  return __ret;
16771}
16772#endif
16773
16774#ifdef __LITTLE_ENDIAN__
16775__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16776  int8x16_t __ret;
16777  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
16778  return __ret;
16779}
16780#else
16781__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16782  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16783  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16784  int8x16_t __ret;
16785  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
16786  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16787  return __ret;
16788}
16789#endif
16790
16791#ifdef __LITTLE_ENDIAN__
16792__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16793  int32x4_t __ret;
16794  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16795  return __ret;
16796}
16797#else
16798__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16799  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16800  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16801  int32x4_t __ret;
16802  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16804  return __ret;
16805}
16806__ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16807  int32x4_t __ret;
16808  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16809  return __ret;
16810}
16811#endif
16812
16813#ifdef __LITTLE_ENDIAN__
16814__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16815  int64x2_t __ret;
16816  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16817  return __ret;
16818}
16819#else
16820__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16821  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16822  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16823  int64x2_t __ret;
16824  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16825  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16826  return __ret;
16827}
16828#endif
16829
16830#ifdef __LITTLE_ENDIAN__
16831__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16832  int16x8_t __ret;
16833  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16834  return __ret;
16835}
16836#else
16837__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16838  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16839  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16840  int16x8_t __ret;
16841  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16842  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16843  return __ret;
16844}
16845__ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16846  int16x8_t __ret;
16847  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16848  return __ret;
16849}
16850#endif
16851
16852#ifdef __LITTLE_ENDIAN__
16853__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16854  uint8x8_t __ret;
16855  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16856  return __ret;
16857}
16858#else
16859__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16860  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16861  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16862  uint8x8_t __ret;
16863  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16864  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16865  return __ret;
16866}
16867#endif
16868
16869#ifdef __LITTLE_ENDIAN__
16870__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16871  uint32x2_t __ret;
16872  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16873  return __ret;
16874}
16875#else
16876__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16877  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16878  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16879  uint32x2_t __ret;
16880  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16881  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16882  return __ret;
16883}
16884#endif
16885
16886#ifdef __LITTLE_ENDIAN__
16887__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16888  uint64x1_t __ret;
16889  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16890  return __ret;
16891}
16892#else
16893__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16894  uint64x1_t __ret;
16895  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16896  return __ret;
16897}
16898#endif
16899
16900#ifdef __LITTLE_ENDIAN__
16901__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16902  uint16x4_t __ret;
16903  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16904  return __ret;
16905}
16906#else
16907__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16908  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16909  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16910  uint16x4_t __ret;
16911  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16913  return __ret;
16914}
16915#endif
16916
16917#ifdef __LITTLE_ENDIAN__
16918__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
16919  int8x8_t __ret;
16920  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16921  return __ret;
16922}
16923#else
16924__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
16925  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16926  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16927  int8x8_t __ret;
16928  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16929  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16930  return __ret;
16931}
16932#endif
16933
16934#ifdef __LITTLE_ENDIAN__
16935__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
16936  int32x2_t __ret;
16937  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16938  return __ret;
16939}
16940#else
16941__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
16942  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16943  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16944  int32x2_t __ret;
16945  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16946  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16947  return __ret;
16948}
16949__ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
16950  int32x2_t __ret;
16951  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16952  return __ret;
16953}
16954#endif
16955
16956#ifdef __LITTLE_ENDIAN__
16957__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
16958  int64x1_t __ret;
16959  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16960  return __ret;
16961}
16962#else
16963__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
16964  int64x1_t __ret;
16965  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16966  return __ret;
16967}
16968#endif
16969
16970#ifdef __LITTLE_ENDIAN__
16971__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
16972  int16x4_t __ret;
16973  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16974  return __ret;
16975}
16976#else
16977__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
16978  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16979  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16980  int16x4_t __ret;
16981  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16982  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16983  return __ret;
16984}
16985__ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
16986  int16x4_t __ret;
16987  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16988  return __ret;
16989}
16990#endif
16991
16992#ifdef __LITTLE_ENDIAN__
16993__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
16994  int64x2_t __ret;
16995  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
16996  return __ret;
16997}
16998#else
16999__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17000  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17001  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17002  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17003  int64x2_t __ret;
17004  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17005  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17006  return __ret;
17007}
17008__ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17009  int64x2_t __ret;
17010  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17011  return __ret;
17012}
17013#endif
17014
17015#ifdef __LITTLE_ENDIAN__
17016__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17017  int32x4_t __ret;
17018  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17019  return __ret;
17020}
17021#else
17022__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17023  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17024  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17025  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17026  int32x4_t __ret;
17027  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17029  return __ret;
17030}
17031__ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17032  int32x4_t __ret;
17033  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17034  return __ret;
17035}
17036#endif
17037
17038#ifdef __LITTLE_ENDIAN__
17039#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17040  int64x2_t __s0 = __p0; \
17041  int32x2_t __s1 = __p1; \
17042  int32x2_t __s2 = __p2; \
17043  int64x2_t __ret; \
17044  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17045  __ret; \
17046})
17047#else
17048#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17049  int64x2_t __s0 = __p0; \
17050  int32x2_t __s1 = __p1; \
17051  int32x2_t __s2 = __p2; \
17052  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17053  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17054  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17055  int64x2_t __ret; \
17056  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17058  __ret; \
17059})
17060#endif
17061
17062#ifdef __LITTLE_ENDIAN__
17063#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17064  int32x4_t __s0 = __p0; \
17065  int16x4_t __s1 = __p1; \
17066  int16x4_t __s2 = __p2; \
17067  int32x4_t __ret; \
17068  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17069  __ret; \
17070})
17071#else
17072#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17073  int32x4_t __s0 = __p0; \
17074  int16x4_t __s1 = __p1; \
17075  int16x4_t __s2 = __p2; \
17076  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17077  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17078  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17079  int32x4_t __ret; \
17080  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17081  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17082  __ret; \
17083})
17084#endif
17085
17086#ifdef __LITTLE_ENDIAN__
17087__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17088  int64x2_t __ret;
17089  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17090  return __ret;
17091}
17092#else
17093__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17094  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17095  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17096  int64x2_t __ret;
17097  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17098  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17099  return __ret;
17100}
17101__ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17102  int64x2_t __ret;
17103  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17104  return __ret;
17105}
17106#endif
17107
17108#ifdef __LITTLE_ENDIAN__
17109__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17110  int32x4_t __ret;
17111  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17112  return __ret;
17113}
17114#else
17115__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17116  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17117  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17118  int32x4_t __ret;
17119  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17120  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17121  return __ret;
17122}
17123__ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17124  int32x4_t __ret;
17125  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17126  return __ret;
17127}
17128#endif
17129
17130#ifdef __LITTLE_ENDIAN__
17131__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17132  int64x2_t __ret;
17133  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17134  return __ret;
17135}
17136#else
17137__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17138  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17139  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17140  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17141  int64x2_t __ret;
17142  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17144  return __ret;
17145}
17146__ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17147  int64x2_t __ret;
17148  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17149  return __ret;
17150}
17151#endif
17152
17153#ifdef __LITTLE_ENDIAN__
17154__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17155  int32x4_t __ret;
17156  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17157  return __ret;
17158}
17159#else
17160__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17161  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17162  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17163  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17164  int32x4_t __ret;
17165  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17166  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17167  return __ret;
17168}
17169__ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17170  int32x4_t __ret;
17171  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17172  return __ret;
17173}
17174#endif
17175
17176#ifdef __LITTLE_ENDIAN__
17177#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17178  int64x2_t __s0 = __p0; \
17179  int32x2_t __s1 = __p1; \
17180  int32x2_t __s2 = __p2; \
17181  int64x2_t __ret; \
17182  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17183  __ret; \
17184})
17185#else
17186#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17187  int64x2_t __s0 = __p0; \
17188  int32x2_t __s1 = __p1; \
17189  int32x2_t __s2 = __p2; \
17190  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17191  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17192  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17193  int64x2_t __ret; \
17194  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17195  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17196  __ret; \
17197})
17198#endif
17199
17200#ifdef __LITTLE_ENDIAN__
17201#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17202  int32x4_t __s0 = __p0; \
17203  int16x4_t __s1 = __p1; \
17204  int16x4_t __s2 = __p2; \
17205  int32x4_t __ret; \
17206  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17207  __ret; \
17208})
17209#else
17210#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17211  int32x4_t __s0 = __p0; \
17212  int16x4_t __s1 = __p1; \
17213  int16x4_t __s2 = __p2; \
17214  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17215  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17216  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17217  int32x4_t __ret; \
17218  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17220  __ret; \
17221})
17222#endif
17223
17224#ifdef __LITTLE_ENDIAN__
17225__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17226  int64x2_t __ret;
17227  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17228  return __ret;
17229}
17230#else
17231__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17232  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17233  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17234  int64x2_t __ret;
17235  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17236  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17237  return __ret;
17238}
17239__ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17240  int64x2_t __ret;
17241  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17242  return __ret;
17243}
17244#endif
17245
17246#ifdef __LITTLE_ENDIAN__
17247__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17248  int32x4_t __ret;
17249  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17250  return __ret;
17251}
17252#else
17253__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17254  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17255  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17256  int32x4_t __ret;
17257  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17258  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17259  return __ret;
17260}
17261__ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17262  int32x4_t __ret;
17263  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17264  return __ret;
17265}
17266#endif
17267
17268#ifdef __LITTLE_ENDIAN__
17269__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17270  int32x4_t __ret;
17271  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17272  return __ret;
17273}
17274#else
17275__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17276  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17277  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17278  int32x4_t __ret;
17279  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17280  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17281  return __ret;
17282}
17283__ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17284  int32x4_t __ret;
17285  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17286  return __ret;
17287}
17288#endif
17289
17290#ifdef __LITTLE_ENDIAN__
17291__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17292  int16x8_t __ret;
17293  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17294  return __ret;
17295}
17296#else
17297__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17298  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17299  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17300  int16x8_t __ret;
17301  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17302  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17303  return __ret;
17304}
17305__ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17306  int16x8_t __ret;
17307  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17308  return __ret;
17309}
17310#endif
17311
17312#ifdef __LITTLE_ENDIAN__
17313__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17314  int32x2_t __ret;
17315  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17316  return __ret;
17317}
17318#else
17319__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17320  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17321  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17322  int32x2_t __ret;
17323  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17324  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17325  return __ret;
17326}
17327__ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17328  int32x2_t __ret;
17329  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17330  return __ret;
17331}
17332#endif
17333
17334#ifdef __LITTLE_ENDIAN__
17335__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17336  int16x4_t __ret;
17337  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17338  return __ret;
17339}
17340#else
17341__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17342  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17343  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17344  int16x4_t __ret;
17345  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17346  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17347  return __ret;
17348}
17349__ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17350  int16x4_t __ret;
17351  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17352  return __ret;
17353}
17354#endif
17355
17356#ifdef __LITTLE_ENDIAN__
17357#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17358  int32x4_t __s0 = __p0; \
17359  int32x2_t __s1 = __p1; \
17360  int32x4_t __ret; \
17361  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17362  __ret; \
17363})
17364#else
17365#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17366  int32x4_t __s0 = __p0; \
17367  int32x2_t __s1 = __p1; \
17368  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17369  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17370  int32x4_t __ret; \
17371  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17373  __ret; \
17374})
17375#endif
17376
17377#ifdef __LITTLE_ENDIAN__
17378#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17379  int16x8_t __s0 = __p0; \
17380  int16x4_t __s1 = __p1; \
17381  int16x8_t __ret; \
17382  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17383  __ret; \
17384})
17385#else
17386#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17387  int16x8_t __s0 = __p0; \
17388  int16x4_t __s1 = __p1; \
17389  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
17390  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17391  int16x8_t __ret; \
17392  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17393  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
17394  __ret; \
17395})
17396#endif
17397
17398#ifdef __LITTLE_ENDIAN__
17399#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17400  int32x2_t __s0 = __p0; \
17401  int32x2_t __s1 = __p1; \
17402  int32x2_t __ret; \
17403  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17404  __ret; \
17405})
17406#else
17407#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17408  int32x2_t __s0 = __p0; \
17409  int32x2_t __s1 = __p1; \
17410  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17411  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17412  int32x2_t __ret; \
17413  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17414  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17415  __ret; \
17416})
17417#endif
17418
17419#ifdef __LITTLE_ENDIAN__
17420#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17421  int16x4_t __s0 = __p0; \
17422  int16x4_t __s1 = __p1; \
17423  int16x4_t __ret; \
17424  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17425  __ret; \
17426})
17427#else
17428#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17429  int16x4_t __s0 = __p0; \
17430  int16x4_t __s1 = __p1; \
17431  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17432  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17433  int16x4_t __ret; \
17434  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17435  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17436  __ret; \
17437})
17438#endif
17439
17440#ifdef __LITTLE_ENDIAN__
17441__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17442  int32x4_t __ret;
17443  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17444  return __ret;
17445}
17446#else
17447__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17448  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17449  int32x4_t __ret;
17450  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17451  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17452  return __ret;
17453}
17454#endif
17455
17456#ifdef __LITTLE_ENDIAN__
17457__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17458  int16x8_t __ret;
17459  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17460  return __ret;
17461}
17462#else
17463__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17464  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17465  int16x8_t __ret;
17466  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17467  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17468  return __ret;
17469}
17470#endif
17471
17472#ifdef __LITTLE_ENDIAN__
17473__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17474  int32x2_t __ret;
17475  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17476  return __ret;
17477}
17478#else
17479__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17480  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17481  int32x2_t __ret;
17482  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17483  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17484  return __ret;
17485}
17486#endif
17487
17488#ifdef __LITTLE_ENDIAN__
17489__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17490  int16x4_t __ret;
17491  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17492  return __ret;
17493}
17494#else
17495__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17496  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17497  int16x4_t __ret;
17498  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17500  return __ret;
17501}
17502#endif
17503
17504#ifdef __LITTLE_ENDIAN__
17505__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17506  int64x2_t __ret;
17507  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17508  return __ret;
17509}
17510#else
17511__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17512  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17513  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17514  int64x2_t __ret;
17515  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
17516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17517  return __ret;
17518}
17519__ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17520  int64x2_t __ret;
17521  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17522  return __ret;
17523}
17524#endif
17525
17526#ifdef __LITTLE_ENDIAN__
17527__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17528  int32x4_t __ret;
17529  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17530  return __ret;
17531}
17532#else
17533__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17534  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17535  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17536  int32x4_t __ret;
17537  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
17538  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17539  return __ret;
17540}
17541__ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17542  int32x4_t __ret;
17543  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17544  return __ret;
17545}
17546#endif
17547
17548#ifdef __LITTLE_ENDIAN__
17549#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17550  int32x2_t __s0 = __p0; \
17551  int32x2_t __s1 = __p1; \
17552  int64x2_t __ret; \
17553  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17554  __ret; \
17555})
17556#else
17557#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17558  int32x2_t __s0 = __p0; \
17559  int32x2_t __s1 = __p1; \
17560  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17561  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17562  int64x2_t __ret; \
17563  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17565  __ret; \
17566})
17567#endif
17568
17569#ifdef __LITTLE_ENDIAN__
17570#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17571  int16x4_t __s0 = __p0; \
17572  int16x4_t __s1 = __p1; \
17573  int32x4_t __ret; \
17574  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17575  __ret; \
17576})
17577#else
17578#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17579  int16x4_t __s0 = __p0; \
17580  int16x4_t __s1 = __p1; \
17581  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17582  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17583  int32x4_t __ret; \
17584  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17585  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17586  __ret; \
17587})
17588#endif
17589
17590#ifdef __LITTLE_ENDIAN__
17591__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17592  int64x2_t __ret;
17593  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17594  return __ret;
17595}
17596#else
17597__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17598  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17599  int64x2_t __ret;
17600  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17601  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17602  return __ret;
17603}
17604__ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17605  int64x2_t __ret;
17606  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17607  return __ret;
17608}
17609#endif
17610
17611#ifdef __LITTLE_ENDIAN__
17612__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17613  int32x4_t __ret;
17614  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17615  return __ret;
17616}
17617#else
17618__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17619  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17620  int32x4_t __ret;
17621  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17623  return __ret;
17624}
17625__ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17626  int32x4_t __ret;
17627  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17628  return __ret;
17629}
17630#endif
17631
17632#ifdef __LITTLE_ENDIAN__
17633__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17634  uint16x4_t __ret;
17635  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17636  return __ret;
17637}
17638#else
17639__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17640  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17641  uint16x4_t __ret;
17642  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
17643  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17644  return __ret;
17645}
17646__ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
17647  uint16x4_t __ret;
17648  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17649  return __ret;
17650}
17651#endif
17652
17653#ifdef __LITTLE_ENDIAN__
17654__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17655  uint32x2_t __ret;
17656  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17657  return __ret;
17658}
17659#else
17660__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17661  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17662  uint32x2_t __ret;
17663  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
17664  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17665  return __ret;
17666}
17667__ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
17668  uint32x2_t __ret;
17669  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17670  return __ret;
17671}
17672#endif
17673
17674#ifdef __LITTLE_ENDIAN__
17675__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17676  uint8x8_t __ret;
17677  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17678  return __ret;
17679}
17680#else
17681__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17682  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17683  uint8x8_t __ret;
17684  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
17685  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17686  return __ret;
17687}
17688__ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
17689  uint8x8_t __ret;
17690  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17691  return __ret;
17692}
17693#endif
17694
17695#ifdef __LITTLE_ENDIAN__
17696__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17697  int16x4_t __ret;
17698  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17699  return __ret;
17700}
17701#else
17702__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17703  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17704  int16x4_t __ret;
17705  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
17706  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17707  return __ret;
17708}
17709__ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
17710  int16x4_t __ret;
17711  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17712  return __ret;
17713}
17714#endif
17715
17716#ifdef __LITTLE_ENDIAN__
17717__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17718  int32x2_t __ret;
17719  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17720  return __ret;
17721}
17722#else
17723__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17724  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17725  int32x2_t __ret;
17726  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
17727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17728  return __ret;
17729}
17730__ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
17731  int32x2_t __ret;
17732  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17733  return __ret;
17734}
17735#endif
17736
17737#ifdef __LITTLE_ENDIAN__
17738__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17739  int8x8_t __ret;
17740  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17741  return __ret;
17742}
17743#else
17744__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17745  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17746  int8x8_t __ret;
17747  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
17748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17749  return __ret;
17750}
17751__ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
17752  int8x8_t __ret;
17753  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17754  return __ret;
17755}
17756#endif
17757
17758#ifdef __LITTLE_ENDIAN__
17759__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17760  uint16x4_t __ret;
17761  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17762  return __ret;
17763}
17764#else
17765__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17766  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17767  uint16x4_t __ret;
17768  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
17769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17770  return __ret;
17771}
17772__ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
17773  uint16x4_t __ret;
17774  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17775  return __ret;
17776}
17777#endif
17778
17779#ifdef __LITTLE_ENDIAN__
17780__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17781  uint32x2_t __ret;
17782  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17783  return __ret;
17784}
17785#else
17786__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17787  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17788  uint32x2_t __ret;
17789  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
17790  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17791  return __ret;
17792}
17793__ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
17794  uint32x2_t __ret;
17795  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17796  return __ret;
17797}
17798#endif
17799
17800#ifdef __LITTLE_ENDIAN__
17801__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17802  uint8x8_t __ret;
17803  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17804  return __ret;
17805}
17806#else
17807__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17808  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17809  uint8x8_t __ret;
17810  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
17811  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17812  return __ret;
17813}
17814__ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
17815  uint8x8_t __ret;
17816  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17817  return __ret;
17818}
17819#endif
17820
17821#ifdef __LITTLE_ENDIAN__
17822__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17823  int8x16_t __ret;
17824  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
17825  return __ret;
17826}
17827#else
17828__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17829  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17830  int8x16_t __ret;
17831  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
17832  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17833  return __ret;
17834}
17835#endif
17836
17837#ifdef __LITTLE_ENDIAN__
17838__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17839  int32x4_t __ret;
17840  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
17841  return __ret;
17842}
17843#else
17844__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17845  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17846  int32x4_t __ret;
17847  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
17848  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17849  return __ret;
17850}
17851#endif
17852
17853#ifdef __LITTLE_ENDIAN__
17854__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17855  int16x8_t __ret;
17856  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
17857  return __ret;
17858}
17859#else
17860__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17861  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17862  int16x8_t __ret;
17863  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
17864  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17865  return __ret;
17866}
17867#endif
17868
17869#ifdef __LITTLE_ENDIAN__
17870__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17871  int8x8_t __ret;
17872  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
17873  return __ret;
17874}
17875#else
17876__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17877  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17878  int8x8_t __ret;
17879  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
17880  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17881  return __ret;
17882}
17883#endif
17884
17885#ifdef __LITTLE_ENDIAN__
17886__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17887  int32x2_t __ret;
17888  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
17889  return __ret;
17890}
17891#else
17892__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17893  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17894  int32x2_t __ret;
17895  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
17896  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17897  return __ret;
17898}
17899#endif
17900
17901#ifdef __LITTLE_ENDIAN__
17902__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17903  int16x4_t __ret;
17904  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
17905  return __ret;
17906}
17907#else
17908__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17909  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17910  int16x4_t __ret;
17911  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
17912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17913  return __ret;
17914}
17915#endif
17916
17917#ifdef __LITTLE_ENDIAN__
17918__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17919  int32x4_t __ret;
17920  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17921  return __ret;
17922}
17923#else
17924__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17925  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17926  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17927  int32x4_t __ret;
17928  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17929  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17930  return __ret;
17931}
17932__ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17933  int32x4_t __ret;
17934  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17935  return __ret;
17936}
17937#endif
17938
17939#ifdef __LITTLE_ENDIAN__
17940__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17941  int16x8_t __ret;
17942  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17943  return __ret;
17944}
17945#else
17946__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17947  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17948  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17949  int16x8_t __ret;
17950  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17951  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17952  return __ret;
17953}
17954__ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17955  int16x8_t __ret;
17956  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17957  return __ret;
17958}
17959#endif
17960
17961#ifdef __LITTLE_ENDIAN__
17962__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17963  int32x2_t __ret;
17964  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17965  return __ret;
17966}
17967#else
17968__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17969  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17970  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17971  int32x2_t __ret;
17972  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17974  return __ret;
17975}
17976__ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17977  int32x2_t __ret;
17978  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17979  return __ret;
17980}
17981#endif
17982
17983#ifdef __LITTLE_ENDIAN__
17984__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17985  int16x4_t __ret;
17986  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17987  return __ret;
17988}
17989#else
17990__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17991  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17992  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17993  int16x4_t __ret;
17994  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17995  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17996  return __ret;
17997}
17998__ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17999  int16x4_t __ret;
18000  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18001  return __ret;
18002}
18003#endif
18004
18005#ifdef __LITTLE_ENDIAN__
18006#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18007  int32x4_t __s0 = __p0; \
18008  int32x2_t __s1 = __p1; \
18009  int32x4_t __ret; \
18010  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18011  __ret; \
18012})
18013#else
18014#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18015  int32x4_t __s0 = __p0; \
18016  int32x2_t __s1 = __p1; \
18017  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18018  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18019  int32x4_t __ret; \
18020  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18022  __ret; \
18023})
18024#endif
18025
18026#ifdef __LITTLE_ENDIAN__
18027#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18028  int16x8_t __s0 = __p0; \
18029  int16x4_t __s1 = __p1; \
18030  int16x8_t __ret; \
18031  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18032  __ret; \
18033})
18034#else
18035#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18036  int16x8_t __s0 = __p0; \
18037  int16x4_t __s1 = __p1; \
18038  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18039  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18040  int16x8_t __ret; \
18041  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18042  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18043  __ret; \
18044})
18045#endif
18046
18047#ifdef __LITTLE_ENDIAN__
18048#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18049  int32x2_t __s0 = __p0; \
18050  int32x2_t __s1 = __p1; \
18051  int32x2_t __ret; \
18052  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
18053  __ret; \
18054})
18055#else
18056#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18057  int32x2_t __s0 = __p0; \
18058  int32x2_t __s1 = __p1; \
18059  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18060  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18061  int32x2_t __ret; \
18062  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
18063  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18064  __ret; \
18065})
18066#endif
18067
18068#ifdef __LITTLE_ENDIAN__
18069#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18070  int16x4_t __s0 = __p0; \
18071  int16x4_t __s1 = __p1; \
18072  int16x4_t __ret; \
18073  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18074  __ret; \
18075})
18076#else
18077#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18078  int16x4_t __s0 = __p0; \
18079  int16x4_t __s1 = __p1; \
18080  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18081  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18082  int16x4_t __ret; \
18083  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18084  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18085  __ret; \
18086})
18087#endif
18088
18089#ifdef __LITTLE_ENDIAN__
18090__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18091  int32x4_t __ret;
18092  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18093  return __ret;
18094}
18095#else
18096__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18097  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18098  int32x4_t __ret;
18099  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18101  return __ret;
18102}
18103#endif
18104
18105#ifdef __LITTLE_ENDIAN__
18106__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18107  int16x8_t __ret;
18108  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18109  return __ret;
18110}
18111#else
18112__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18113  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18114  int16x8_t __ret;
18115  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18116  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18117  return __ret;
18118}
18119#endif
18120
18121#ifdef __LITTLE_ENDIAN__
18122__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18123  int32x2_t __ret;
18124  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18125  return __ret;
18126}
18127#else
18128__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18129  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18130  int32x2_t __ret;
18131  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18132  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18133  return __ret;
18134}
18135#endif
18136
18137#ifdef __LITTLE_ENDIAN__
18138__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18139  int16x4_t __ret;
18140  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18141  return __ret;
18142}
18143#else
18144__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18145  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18146  int16x4_t __ret;
18147  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18148  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18149  return __ret;
18150}
18151#endif
18152
18153#ifdef __LITTLE_ENDIAN__
18154__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18155  uint8x16_t __ret;
18156  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18157  return __ret;
18158}
18159#else
18160__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18161  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18162  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18163  uint8x16_t __ret;
18164  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18165  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18166  return __ret;
18167}
18168#endif
18169
18170#ifdef __LITTLE_ENDIAN__
18171__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18172  uint32x4_t __ret;
18173  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18174  return __ret;
18175}
18176#else
18177__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18178  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18179  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18180  uint32x4_t __ret;
18181  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18182  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18183  return __ret;
18184}
18185#endif
18186
18187#ifdef __LITTLE_ENDIAN__
18188__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18189  uint64x2_t __ret;
18190  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18191  return __ret;
18192}
18193#else
18194__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18195  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18196  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18197  uint64x2_t __ret;
18198  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18199  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18200  return __ret;
18201}
18202#endif
18203
18204#ifdef __LITTLE_ENDIAN__
18205__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18206  uint16x8_t __ret;
18207  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18208  return __ret;
18209}
18210#else
18211__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18212  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18213  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18214  uint16x8_t __ret;
18215  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18217  return __ret;
18218}
18219#endif
18220
18221#ifdef __LITTLE_ENDIAN__
18222__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18223  int8x16_t __ret;
18224  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18225  return __ret;
18226}
18227#else
18228__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18229  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18230  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18231  int8x16_t __ret;
18232  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18233  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18234  return __ret;
18235}
18236#endif
18237
18238#ifdef __LITTLE_ENDIAN__
18239__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18240  int32x4_t __ret;
18241  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18242  return __ret;
18243}
18244#else
18245__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18246  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18247  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18248  int32x4_t __ret;
18249  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18251  return __ret;
18252}
18253#endif
18254
18255#ifdef __LITTLE_ENDIAN__
18256__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18257  int64x2_t __ret;
18258  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18259  return __ret;
18260}
18261#else
18262__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18263  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18264  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18265  int64x2_t __ret;
18266  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18267  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18268  return __ret;
18269}
18270#endif
18271
18272#ifdef __LITTLE_ENDIAN__
18273__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18274  int16x8_t __ret;
18275  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18276  return __ret;
18277}
18278#else
18279__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18280  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18281  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18282  int16x8_t __ret;
18283  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18284  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18285  return __ret;
18286}
18287#endif
18288
18289#ifdef __LITTLE_ENDIAN__
18290__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18291  uint8x8_t __ret;
18292  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18293  return __ret;
18294}
18295#else
18296__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18297  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18298  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18299  uint8x8_t __ret;
18300  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18302  return __ret;
18303}
18304#endif
18305
18306#ifdef __LITTLE_ENDIAN__
18307__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18308  uint32x2_t __ret;
18309  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18310  return __ret;
18311}
18312#else
18313__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18314  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18315  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18316  uint32x2_t __ret;
18317  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18318  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18319  return __ret;
18320}
18321#endif
18322
18323#ifdef __LITTLE_ENDIAN__
18324__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18325  uint64x1_t __ret;
18326  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18327  return __ret;
18328}
18329#else
18330__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18331  uint64x1_t __ret;
18332  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18333  return __ret;
18334}
18335#endif
18336
18337#ifdef __LITTLE_ENDIAN__
18338__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18339  uint16x4_t __ret;
18340  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18341  return __ret;
18342}
18343#else
18344__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18345  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18346  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18347  uint16x4_t __ret;
18348  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18349  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18350  return __ret;
18351}
18352#endif
18353
18354#ifdef __LITTLE_ENDIAN__
18355__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18356  int8x8_t __ret;
18357  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18358  return __ret;
18359}
18360#else
18361__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18362  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18363  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18364  int8x8_t __ret;
18365  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18367  return __ret;
18368}
18369#endif
18370
18371#ifdef __LITTLE_ENDIAN__
18372__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18373  int32x2_t __ret;
18374  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18375  return __ret;
18376}
18377#else
18378__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18379  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18380  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18381  int32x2_t __ret;
18382  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18383  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18384  return __ret;
18385}
18386#endif
18387
18388#ifdef __LITTLE_ENDIAN__
18389__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18390  int64x1_t __ret;
18391  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18392  return __ret;
18393}
18394#else
18395__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18396  int64x1_t __ret;
18397  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18398  return __ret;
18399}
18400#endif
18401
18402#ifdef __LITTLE_ENDIAN__
18403__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18404  int16x4_t __ret;
18405  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18406  return __ret;
18407}
18408#else
18409__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18410  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18411  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18412  int16x4_t __ret;
18413  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18415  return __ret;
18416}
18417#endif
18418
18419#ifdef __LITTLE_ENDIAN__
18420#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18421  uint32x4_t __s0 = __p0; \
18422  uint16x4_t __ret; \
18423  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18424  __ret; \
18425})
18426#else
18427#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18428  uint32x4_t __s0 = __p0; \
18429  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18430  uint16x4_t __ret; \
18431  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
18432  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18433  __ret; \
18434})
18435#define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18436  uint32x4_t __s0 = __p0; \
18437  uint16x4_t __ret; \
18438  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18439  __ret; \
18440})
18441#endif
18442
18443#ifdef __LITTLE_ENDIAN__
18444#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18445  uint64x2_t __s0 = __p0; \
18446  uint32x2_t __ret; \
18447  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18448  __ret; \
18449})
18450#else
18451#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18452  uint64x2_t __s0 = __p0; \
18453  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18454  uint32x2_t __ret; \
18455  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
18456  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18457  __ret; \
18458})
18459#define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18460  uint64x2_t __s0 = __p0; \
18461  uint32x2_t __ret; \
18462  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18463  __ret; \
18464})
18465#endif
18466
18467#ifdef __LITTLE_ENDIAN__
18468#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18469  uint16x8_t __s0 = __p0; \
18470  uint8x8_t __ret; \
18471  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18472  __ret; \
18473})
18474#else
18475#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18476  uint16x8_t __s0 = __p0; \
18477  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18478  uint8x8_t __ret; \
18479  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
18480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18481  __ret; \
18482})
18483#define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18484  uint16x8_t __s0 = __p0; \
18485  uint8x8_t __ret; \
18486  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18487  __ret; \
18488})
18489#endif
18490
18491#ifdef __LITTLE_ENDIAN__
18492#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18493  int32x4_t __s0 = __p0; \
18494  int16x4_t __ret; \
18495  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18496  __ret; \
18497})
18498#else
18499#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18500  int32x4_t __s0 = __p0; \
18501  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18502  int16x4_t __ret; \
18503  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
18504  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18505  __ret; \
18506})
18507#define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18508  int32x4_t __s0 = __p0; \
18509  int16x4_t __ret; \
18510  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18511  __ret; \
18512})
18513#endif
18514
18515#ifdef __LITTLE_ENDIAN__
18516#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18517  int64x2_t __s0 = __p0; \
18518  int32x2_t __ret; \
18519  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18520  __ret; \
18521})
18522#else
18523#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18524  int64x2_t __s0 = __p0; \
18525  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18526  int32x2_t __ret; \
18527  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
18528  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18529  __ret; \
18530})
18531#define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18532  int64x2_t __s0 = __p0; \
18533  int32x2_t __ret; \
18534  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18535  __ret; \
18536})
18537#endif
18538
18539#ifdef __LITTLE_ENDIAN__
18540#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18541  int16x8_t __s0 = __p0; \
18542  int8x8_t __ret; \
18543  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18544  __ret; \
18545})
18546#else
18547#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18548  int16x8_t __s0 = __p0; \
18549  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18550  int8x8_t __ret; \
18551  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
18552  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18553  __ret; \
18554})
18555#define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18556  int16x8_t __s0 = __p0; \
18557  int8x8_t __ret; \
18558  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18559  __ret; \
18560})
18561#endif
18562
18563#ifdef __LITTLE_ENDIAN__
18564#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18565  int32x4_t __s0 = __p0; \
18566  uint16x4_t __ret; \
18567  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18568  __ret; \
18569})
18570#else
18571#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18572  int32x4_t __s0 = __p0; \
18573  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18574  uint16x4_t __ret; \
18575  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
18576  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18577  __ret; \
18578})
18579#define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18580  int32x4_t __s0 = __p0; \
18581  uint16x4_t __ret; \
18582  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18583  __ret; \
18584})
18585#endif
18586
18587#ifdef __LITTLE_ENDIAN__
18588#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18589  int64x2_t __s0 = __p0; \
18590  uint32x2_t __ret; \
18591  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18592  __ret; \
18593})
18594#else
18595#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18596  int64x2_t __s0 = __p0; \
18597  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18598  uint32x2_t __ret; \
18599  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
18600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18601  __ret; \
18602})
18603#define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18604  int64x2_t __s0 = __p0; \
18605  uint32x2_t __ret; \
18606  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18607  __ret; \
18608})
18609#endif
18610
18611#ifdef __LITTLE_ENDIAN__
18612#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18613  int16x8_t __s0 = __p0; \
18614  uint8x8_t __ret; \
18615  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18616  __ret; \
18617})
18618#else
18619#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18620  int16x8_t __s0 = __p0; \
18621  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18622  uint8x8_t __ret; \
18623  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
18624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18625  __ret; \
18626})
18627#define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18628  int16x8_t __s0 = __p0; \
18629  uint8x8_t __ret; \
18630  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18631  __ret; \
18632})
18633#endif
18634
18635#ifdef __LITTLE_ENDIAN__
18636__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18637  uint8x16_t __ret;
18638  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18639  return __ret;
18640}
18641#else
18642__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18643  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18644  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18645  uint8x16_t __ret;
18646  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18648  return __ret;
18649}
18650#endif
18651
18652#ifdef __LITTLE_ENDIAN__
18653__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18654  uint32x4_t __ret;
18655  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18656  return __ret;
18657}
18658#else
18659__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18660  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18661  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18662  uint32x4_t __ret;
18663  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18665  return __ret;
18666}
18667#endif
18668
18669#ifdef __LITTLE_ENDIAN__
18670__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18671  uint64x2_t __ret;
18672  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18673  return __ret;
18674}
18675#else
18676__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18677  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18678  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18679  uint64x2_t __ret;
18680  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18682  return __ret;
18683}
18684#endif
18685
18686#ifdef __LITTLE_ENDIAN__
18687__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18688  uint16x8_t __ret;
18689  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18690  return __ret;
18691}
18692#else
18693__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18694  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18695  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18696  uint16x8_t __ret;
18697  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18698  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18699  return __ret;
18700}
18701#endif
18702
18703#ifdef __LITTLE_ENDIAN__
18704__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18705  int8x16_t __ret;
18706  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18707  return __ret;
18708}
18709#else
18710__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18711  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18712  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18713  int8x16_t __ret;
18714  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18715  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18716  return __ret;
18717}
18718#endif
18719
18720#ifdef __LITTLE_ENDIAN__
18721__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18722  int32x4_t __ret;
18723  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18724  return __ret;
18725}
18726#else
18727__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18728  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18729  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18730  int32x4_t __ret;
18731  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18732  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18733  return __ret;
18734}
18735#endif
18736
18737#ifdef __LITTLE_ENDIAN__
18738__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18739  int64x2_t __ret;
18740  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18741  return __ret;
18742}
18743#else
18744__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18745  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18746  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18747  int64x2_t __ret;
18748  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18750  return __ret;
18751}
18752#endif
18753
18754#ifdef __LITTLE_ENDIAN__
18755__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18756  int16x8_t __ret;
18757  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18758  return __ret;
18759}
18760#else
18761__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18762  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18763  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18764  int16x8_t __ret;
18765  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18766  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18767  return __ret;
18768}
18769#endif
18770
18771#ifdef __LITTLE_ENDIAN__
18772__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18773  uint8x8_t __ret;
18774  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18775  return __ret;
18776}
18777#else
18778__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18779  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18780  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18781  uint8x8_t __ret;
18782  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18783  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18784  return __ret;
18785}
18786#endif
18787
18788#ifdef __LITTLE_ENDIAN__
18789__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18790  uint32x2_t __ret;
18791  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18792  return __ret;
18793}
18794#else
18795__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18796  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18797  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18798  uint32x2_t __ret;
18799  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18800  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18801  return __ret;
18802}
18803#endif
18804
18805#ifdef __LITTLE_ENDIAN__
18806__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18807  uint64x1_t __ret;
18808  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18809  return __ret;
18810}
18811#else
18812__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18813  uint64x1_t __ret;
18814  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18815  return __ret;
18816}
18817#endif
18818
18819#ifdef __LITTLE_ENDIAN__
18820__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18821  uint16x4_t __ret;
18822  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18823  return __ret;
18824}
18825#else
18826__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18827  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18828  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18829  uint16x4_t __ret;
18830  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18831  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18832  return __ret;
18833}
18834#endif
18835
18836#ifdef __LITTLE_ENDIAN__
18837__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18838  int8x8_t __ret;
18839  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18840  return __ret;
18841}
18842#else
18843__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18844  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18845  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18846  int8x8_t __ret;
18847  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18848  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18849  return __ret;
18850}
18851#endif
18852
18853#ifdef __LITTLE_ENDIAN__
18854__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18855  int32x2_t __ret;
18856  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18857  return __ret;
18858}
18859#else
18860__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18861  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18862  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18863  int32x2_t __ret;
18864  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18865  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18866  return __ret;
18867}
18868#endif
18869
18870#ifdef __LITTLE_ENDIAN__
18871__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18872  int64x1_t __ret;
18873  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18874  return __ret;
18875}
18876#else
18877__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18878  int64x1_t __ret;
18879  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18880  return __ret;
18881}
18882#endif
18883
18884#ifdef __LITTLE_ENDIAN__
18885__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18886  int16x4_t __ret;
18887  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18888  return __ret;
18889}
18890#else
18891__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18892  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18893  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18894  int16x4_t __ret;
18895  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18896  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18897  return __ret;
18898}
18899#endif
18900
18901#ifdef __LITTLE_ENDIAN__
18902#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18903  uint8x16_t __s0 = __p0; \
18904  uint8x16_t __ret; \
18905  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
18906  __ret; \
18907})
18908#else
18909#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18910  uint8x16_t __s0 = __p0; \
18911  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18912  uint8x16_t __ret; \
18913  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
18914  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18915  __ret; \
18916})
18917#endif
18918
18919#ifdef __LITTLE_ENDIAN__
18920#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
18921  uint32x4_t __s0 = __p0; \
18922  uint32x4_t __ret; \
18923  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
18924  __ret; \
18925})
18926#else
18927#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
18928  uint32x4_t __s0 = __p0; \
18929  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18930  uint32x4_t __ret; \
18931  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
18932  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18933  __ret; \
18934})
18935#endif
18936
18937#ifdef __LITTLE_ENDIAN__
18938#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
18939  uint64x2_t __s0 = __p0; \
18940  uint64x2_t __ret; \
18941  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
18942  __ret; \
18943})
18944#else
18945#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
18946  uint64x2_t __s0 = __p0; \
18947  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18948  uint64x2_t __ret; \
18949  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
18950  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18951  __ret; \
18952})
18953#endif
18954
18955#ifdef __LITTLE_ENDIAN__
18956#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
18957  uint16x8_t __s0 = __p0; \
18958  uint16x8_t __ret; \
18959  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
18960  __ret; \
18961})
18962#else
18963#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
18964  uint16x8_t __s0 = __p0; \
18965  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18966  uint16x8_t __ret; \
18967  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
18968  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18969  __ret; \
18970})
18971#endif
18972
18973#ifdef __LITTLE_ENDIAN__
18974#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
18975  int8x16_t __s0 = __p0; \
18976  int8x16_t __ret; \
18977  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
18978  __ret; \
18979})
18980#else
18981#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
18982  int8x16_t __s0 = __p0; \
18983  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18984  int8x16_t __ret; \
18985  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
18986  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18987  __ret; \
18988})
18989#endif
18990
18991#ifdef __LITTLE_ENDIAN__
18992#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
18993  int32x4_t __s0 = __p0; \
18994  int32x4_t __ret; \
18995  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
18996  __ret; \
18997})
18998#else
18999#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19000  int32x4_t __s0 = __p0; \
19001  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19002  int32x4_t __ret; \
19003  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
19004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19005  __ret; \
19006})
19007#endif
19008
19009#ifdef __LITTLE_ENDIAN__
19010#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19011  int64x2_t __s0 = __p0; \
19012  int64x2_t __ret; \
19013  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
19014  __ret; \
19015})
19016#else
19017#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19018  int64x2_t __s0 = __p0; \
19019  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19020  int64x2_t __ret; \
19021  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
19022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19023  __ret; \
19024})
19025#endif
19026
19027#ifdef __LITTLE_ENDIAN__
19028#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19029  int16x8_t __s0 = __p0; \
19030  int16x8_t __ret; \
19031  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
19032  __ret; \
19033})
19034#else
19035#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19036  int16x8_t __s0 = __p0; \
19037  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19038  int16x8_t __ret; \
19039  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
19040  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19041  __ret; \
19042})
19043#endif
19044
19045#ifdef __LITTLE_ENDIAN__
19046#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19047  uint8x8_t __s0 = __p0; \
19048  uint8x8_t __ret; \
19049  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
19050  __ret; \
19051})
19052#else
19053#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19054  uint8x8_t __s0 = __p0; \
19055  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19056  uint8x8_t __ret; \
19057  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
19058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19059  __ret; \
19060})
19061#endif
19062
19063#ifdef __LITTLE_ENDIAN__
19064#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19065  uint32x2_t __s0 = __p0; \
19066  uint32x2_t __ret; \
19067  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
19068  __ret; \
19069})
19070#else
19071#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19072  uint32x2_t __s0 = __p0; \
19073  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19074  uint32x2_t __ret; \
19075  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
19076  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19077  __ret; \
19078})
19079#endif
19080
19081#ifdef __LITTLE_ENDIAN__
19082#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19083  uint64x1_t __s0 = __p0; \
19084  uint64x1_t __ret; \
19085  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19086  __ret; \
19087})
19088#else
19089#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19090  uint64x1_t __s0 = __p0; \
19091  uint64x1_t __ret; \
19092  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19093  __ret; \
19094})
19095#endif
19096
19097#ifdef __LITTLE_ENDIAN__
19098#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19099  uint16x4_t __s0 = __p0; \
19100  uint16x4_t __ret; \
19101  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
19102  __ret; \
19103})
19104#else
19105#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19106  uint16x4_t __s0 = __p0; \
19107  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19108  uint16x4_t __ret; \
19109  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
19110  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19111  __ret; \
19112})
19113#endif
19114
19115#ifdef __LITTLE_ENDIAN__
19116#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19117  int8x8_t __s0 = __p0; \
19118  int8x8_t __ret; \
19119  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
19120  __ret; \
19121})
19122#else
19123#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19124  int8x8_t __s0 = __p0; \
19125  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19126  int8x8_t __ret; \
19127  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
19128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19129  __ret; \
19130})
19131#endif
19132
19133#ifdef __LITTLE_ENDIAN__
19134#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19135  int32x2_t __s0 = __p0; \
19136  int32x2_t __ret; \
19137  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
19138  __ret; \
19139})
19140#else
19141#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19142  int32x2_t __s0 = __p0; \
19143  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19144  int32x2_t __ret; \
19145  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
19146  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19147  __ret; \
19148})
19149#endif
19150
19151#ifdef __LITTLE_ENDIAN__
19152#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19153  int64x1_t __s0 = __p0; \
19154  int64x1_t __ret; \
19155  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19156  __ret; \
19157})
19158#else
19159#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19160  int64x1_t __s0 = __p0; \
19161  int64x1_t __ret; \
19162  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19163  __ret; \
19164})
19165#endif
19166
19167#ifdef __LITTLE_ENDIAN__
19168#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19169  int16x4_t __s0 = __p0; \
19170  int16x4_t __ret; \
19171  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
19172  __ret; \
19173})
19174#else
19175#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19176  int16x4_t __s0 = __p0; \
19177  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19178  int16x4_t __ret; \
19179  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
19180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19181  __ret; \
19182})
19183#endif
19184
19185#ifdef __LITTLE_ENDIAN__
19186#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19187  int8x16_t __s0 = __p0; \
19188  uint8x16_t __ret; \
19189  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
19190  __ret; \
19191})
19192#else
19193#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19194  int8x16_t __s0 = __p0; \
19195  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19196  uint8x16_t __ret; \
19197  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
19198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19199  __ret; \
19200})
19201#endif
19202
19203#ifdef __LITTLE_ENDIAN__
19204#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19205  int32x4_t __s0 = __p0; \
19206  uint32x4_t __ret; \
19207  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
19208  __ret; \
19209})
19210#else
19211#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19212  int32x4_t __s0 = __p0; \
19213  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19214  uint32x4_t __ret; \
19215  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
19216  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19217  __ret; \
19218})
19219#endif
19220
19221#ifdef __LITTLE_ENDIAN__
19222#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19223  int64x2_t __s0 = __p0; \
19224  uint64x2_t __ret; \
19225  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
19226  __ret; \
19227})
19228#else
19229#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19230  int64x2_t __s0 = __p0; \
19231  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19232  uint64x2_t __ret; \
19233  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
19234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19235  __ret; \
19236})
19237#endif
19238
19239#ifdef __LITTLE_ENDIAN__
19240#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19241  int16x8_t __s0 = __p0; \
19242  uint16x8_t __ret; \
19243  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
19244  __ret; \
19245})
19246#else
19247#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19248  int16x8_t __s0 = __p0; \
19249  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19250  uint16x8_t __ret; \
19251  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
19252  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19253  __ret; \
19254})
19255#endif
19256
19257#ifdef __LITTLE_ENDIAN__
19258#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19259  int8x8_t __s0 = __p0; \
19260  uint8x8_t __ret; \
19261  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
19262  __ret; \
19263})
19264#else
19265#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19266  int8x8_t __s0 = __p0; \
19267  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19268  uint8x8_t __ret; \
19269  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
19270  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19271  __ret; \
19272})
19273#endif
19274
19275#ifdef __LITTLE_ENDIAN__
19276#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19277  int32x2_t __s0 = __p0; \
19278  uint32x2_t __ret; \
19279  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
19280  __ret; \
19281})
19282#else
19283#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19284  int32x2_t __s0 = __p0; \
19285  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19286  uint32x2_t __ret; \
19287  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
19288  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19289  __ret; \
19290})
19291#endif
19292
19293#ifdef __LITTLE_ENDIAN__
19294#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19295  int64x1_t __s0 = __p0; \
19296  uint64x1_t __ret; \
19297  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19298  __ret; \
19299})
19300#else
19301#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19302  int64x1_t __s0 = __p0; \
19303  uint64x1_t __ret; \
19304  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19305  __ret; \
19306})
19307#endif
19308
19309#ifdef __LITTLE_ENDIAN__
19310#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19311  int16x4_t __s0 = __p0; \
19312  uint16x4_t __ret; \
19313  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
19314  __ret; \
19315})
19316#else
19317#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19318  int16x4_t __s0 = __p0; \
19319  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19320  uint16x4_t __ret; \
19321  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
19322  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19323  __ret; \
19324})
19325#endif
19326
19327#ifdef __LITTLE_ENDIAN__
19328#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19329  uint32x4_t __s0 = __p0; \
19330  uint16x4_t __ret; \
19331  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19332  __ret; \
19333})
19334#else
19335#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19336  uint32x4_t __s0 = __p0; \
19337  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19338  uint16x4_t __ret; \
19339  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
19340  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19341  __ret; \
19342})
19343#define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19344  uint32x4_t __s0 = __p0; \
19345  uint16x4_t __ret; \
19346  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19347  __ret; \
19348})
19349#endif
19350
19351#ifdef __LITTLE_ENDIAN__
19352#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19353  uint64x2_t __s0 = __p0; \
19354  uint32x2_t __ret; \
19355  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19356  __ret; \
19357})
19358#else
19359#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19360  uint64x2_t __s0 = __p0; \
19361  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19362  uint32x2_t __ret; \
19363  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
19364  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19365  __ret; \
19366})
19367#define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19368  uint64x2_t __s0 = __p0; \
19369  uint32x2_t __ret; \
19370  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19371  __ret; \
19372})
19373#endif
19374
19375#ifdef __LITTLE_ENDIAN__
19376#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19377  uint16x8_t __s0 = __p0; \
19378  uint8x8_t __ret; \
19379  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19380  __ret; \
19381})
19382#else
19383#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19384  uint16x8_t __s0 = __p0; \
19385  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19386  uint8x8_t __ret; \
19387  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
19388  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19389  __ret; \
19390})
19391#define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19392  uint16x8_t __s0 = __p0; \
19393  uint8x8_t __ret; \
19394  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19395  __ret; \
19396})
19397#endif
19398
19399#ifdef __LITTLE_ENDIAN__
19400#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19401  int32x4_t __s0 = __p0; \
19402  int16x4_t __ret; \
19403  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19404  __ret; \
19405})
19406#else
19407#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19408  int32x4_t __s0 = __p0; \
19409  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19410  int16x4_t __ret; \
19411  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
19412  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19413  __ret; \
19414})
19415#define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19416  int32x4_t __s0 = __p0; \
19417  int16x4_t __ret; \
19418  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19419  __ret; \
19420})
19421#endif
19422
19423#ifdef __LITTLE_ENDIAN__
19424#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19425  int64x2_t __s0 = __p0; \
19426  int32x2_t __ret; \
19427  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19428  __ret; \
19429})
19430#else
19431#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19432  int64x2_t __s0 = __p0; \
19433  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19434  int32x2_t __ret; \
19435  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
19436  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19437  __ret; \
19438})
19439#define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19440  int64x2_t __s0 = __p0; \
19441  int32x2_t __ret; \
19442  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19443  __ret; \
19444})
19445#endif
19446
19447#ifdef __LITTLE_ENDIAN__
19448#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19449  int16x8_t __s0 = __p0; \
19450  int8x8_t __ret; \
19451  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19452  __ret; \
19453})
19454#else
19455#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19456  int16x8_t __s0 = __p0; \
19457  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19458  int8x8_t __ret; \
19459  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
19460  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19461  __ret; \
19462})
19463#define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19464  int16x8_t __s0 = __p0; \
19465  int8x8_t __ret; \
19466  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19467  __ret; \
19468})
19469#endif
19470
19471#ifdef __LITTLE_ENDIAN__
19472#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19473  int32x4_t __s0 = __p0; \
19474  uint16x4_t __ret; \
19475  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19476  __ret; \
19477})
19478#else
19479#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19480  int32x4_t __s0 = __p0; \
19481  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19482  uint16x4_t __ret; \
19483  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
19484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19485  __ret; \
19486})
19487#define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19488  int32x4_t __s0 = __p0; \
19489  uint16x4_t __ret; \
19490  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19491  __ret; \
19492})
19493#endif
19494
19495#ifdef __LITTLE_ENDIAN__
19496#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19497  int64x2_t __s0 = __p0; \
19498  uint32x2_t __ret; \
19499  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19500  __ret; \
19501})
19502#else
19503#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19504  int64x2_t __s0 = __p0; \
19505  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19506  uint32x2_t __ret; \
19507  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
19508  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19509  __ret; \
19510})
19511#define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19512  int64x2_t __s0 = __p0; \
19513  uint32x2_t __ret; \
19514  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19515  __ret; \
19516})
19517#endif
19518
19519#ifdef __LITTLE_ENDIAN__
19520#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19521  int16x8_t __s0 = __p0; \
19522  uint8x8_t __ret; \
19523  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19524  __ret; \
19525})
19526#else
19527#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19528  int16x8_t __s0 = __p0; \
19529  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19530  uint8x8_t __ret; \
19531  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
19532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19533  __ret; \
19534})
19535#define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19536  int16x8_t __s0 = __p0; \
19537  uint8x8_t __ret; \
19538  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19539  __ret; \
19540})
19541#endif
19542
19543#ifdef __LITTLE_ENDIAN__
19544__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19545  uint8x16_t __ret;
19546  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
19547  return __ret;
19548}
19549#else
19550__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19551  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19552  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19553  uint8x16_t __ret;
19554  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
19555  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19556  return __ret;
19557}
19558#endif
19559
19560#ifdef __LITTLE_ENDIAN__
19561__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19562  uint32x4_t __ret;
19563  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
19564  return __ret;
19565}
19566#else
19567__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19568  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19569  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19570  uint32x4_t __ret;
19571  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
19572  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19573  return __ret;
19574}
19575#endif
19576
19577#ifdef __LITTLE_ENDIAN__
19578__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19579  uint64x2_t __ret;
19580  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
19581  return __ret;
19582}
19583#else
19584__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19585  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19586  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19587  uint64x2_t __ret;
19588  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
19589  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19590  return __ret;
19591}
19592#endif
19593
19594#ifdef __LITTLE_ENDIAN__
19595__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19596  uint16x8_t __ret;
19597  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
19598  return __ret;
19599}
19600#else
19601__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19602  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19603  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19604  uint16x8_t __ret;
19605  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
19606  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19607  return __ret;
19608}
19609#endif
19610
19611#ifdef __LITTLE_ENDIAN__
19612__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19613  int8x16_t __ret;
19614  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
19615  return __ret;
19616}
19617#else
19618__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19619  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19620  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19621  int8x16_t __ret;
19622  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
19623  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19624  return __ret;
19625}
19626#endif
19627
19628#ifdef __LITTLE_ENDIAN__
19629__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19630  int32x4_t __ret;
19631  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19632  return __ret;
19633}
19634#else
19635__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19636  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19637  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19638  int32x4_t __ret;
19639  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19641  return __ret;
19642}
19643__ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19644  int32x4_t __ret;
19645  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19646  return __ret;
19647}
19648#endif
19649
19650#ifdef __LITTLE_ENDIAN__
19651__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19652  int64x2_t __ret;
19653  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
19654  return __ret;
19655}
19656#else
19657__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19658  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19659  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19660  int64x2_t __ret;
19661  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
19662  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19663  return __ret;
19664}
19665#endif
19666
19667#ifdef __LITTLE_ENDIAN__
19668__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19669  int16x8_t __ret;
19670  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19671  return __ret;
19672}
19673#else
19674__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19675  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19676  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19677  int16x8_t __ret;
19678  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19679  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19680  return __ret;
19681}
19682__ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19683  int16x8_t __ret;
19684  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19685  return __ret;
19686}
19687#endif
19688
19689#ifdef __LITTLE_ENDIAN__
19690__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19691  uint8x8_t __ret;
19692  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
19693  return __ret;
19694}
19695#else
19696__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19697  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19698  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19699  uint8x8_t __ret;
19700  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
19701  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19702  return __ret;
19703}
19704#endif
19705
19706#ifdef __LITTLE_ENDIAN__
19707__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19708  uint32x2_t __ret;
19709  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
19710  return __ret;
19711}
19712#else
19713__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19714  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19715  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19716  uint32x2_t __ret;
19717  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
19718  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19719  return __ret;
19720}
19721#endif
19722
19723#ifdef __LITTLE_ENDIAN__
19724__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19725  uint64x1_t __ret;
19726  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19727  return __ret;
19728}
19729#else
19730__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19731  uint64x1_t __ret;
19732  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19733  return __ret;
19734}
19735#endif
19736
19737#ifdef __LITTLE_ENDIAN__
19738__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19739  uint16x4_t __ret;
19740  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
19741  return __ret;
19742}
19743#else
19744__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19745  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19746  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19747  uint16x4_t __ret;
19748  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
19749  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19750  return __ret;
19751}
19752#endif
19753
19754#ifdef __LITTLE_ENDIAN__
19755__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19756  int8x8_t __ret;
19757  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
19758  return __ret;
19759}
19760#else
19761__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19762  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19763  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19764  int8x8_t __ret;
19765  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
19766  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19767  return __ret;
19768}
19769#endif
19770
19771#ifdef __LITTLE_ENDIAN__
19772__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19773  int32x2_t __ret;
19774  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19775  return __ret;
19776}
19777#else
19778__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19779  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19780  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19781  int32x2_t __ret;
19782  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19783  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19784  return __ret;
19785}
19786__ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19787  int32x2_t __ret;
19788  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19789  return __ret;
19790}
19791#endif
19792
19793#ifdef __LITTLE_ENDIAN__
19794__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19795  int64x1_t __ret;
19796  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19797  return __ret;
19798}
19799#else
19800__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19801  int64x1_t __ret;
19802  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19803  return __ret;
19804}
19805#endif
19806
19807#ifdef __LITTLE_ENDIAN__
19808__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19809  int16x4_t __ret;
19810  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19811  return __ret;
19812}
19813#else
19814__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19815  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19816  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19817  int16x4_t __ret;
19818  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19819  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19820  return __ret;
19821}
19822__ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19823  int16x4_t __ret;
19824  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19825  return __ret;
19826}
19827#endif
19828
19829#ifdef __LITTLE_ENDIAN__
19830__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19831  uint16x4_t __ret;
19832  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19833  return __ret;
19834}
19835#else
19836__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19837  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19838  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19839  uint16x4_t __ret;
19840  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
19841  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19842  return __ret;
19843}
19844__ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19845  uint16x4_t __ret;
19846  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19847  return __ret;
19848}
19849#endif
19850
19851#ifdef __LITTLE_ENDIAN__
19852__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19853  uint32x2_t __ret;
19854  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19855  return __ret;
19856}
19857#else
19858__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19859  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19860  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19861  uint32x2_t __ret;
19862  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
19863  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19864  return __ret;
19865}
19866__ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19867  uint32x2_t __ret;
19868  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19869  return __ret;
19870}
19871#endif
19872
19873#ifdef __LITTLE_ENDIAN__
19874__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19875  uint8x8_t __ret;
19876  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19877  return __ret;
19878}
19879#else
19880__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19881  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19882  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19883  uint8x8_t __ret;
19884  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
19885  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19886  return __ret;
19887}
19888__ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19889  uint8x8_t __ret;
19890  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19891  return __ret;
19892}
19893#endif
19894
19895#ifdef __LITTLE_ENDIAN__
19896__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19897  int16x4_t __ret;
19898  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19899  return __ret;
19900}
19901#else
19902__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19903  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19904  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19905  int16x4_t __ret;
19906  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
19907  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19908  return __ret;
19909}
19910__ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19911  int16x4_t __ret;
19912  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19913  return __ret;
19914}
19915#endif
19916
19917#ifdef __LITTLE_ENDIAN__
19918__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19919  int32x2_t __ret;
19920  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
19921  return __ret;
19922}
19923#else
19924__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19925  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19926  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19927  int32x2_t __ret;
19928  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
19929  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19930  return __ret;
19931}
19932__ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19933  int32x2_t __ret;
19934  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
19935  return __ret;
19936}
19937#endif
19938
19939#ifdef __LITTLE_ENDIAN__
19940__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
19941  int8x8_t __ret;
19942  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
19943  return __ret;
19944}
19945#else
19946__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
19947  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19948  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19949  int8x8_t __ret;
19950  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
19951  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19952  return __ret;
19953}
19954__ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
19955  int8x8_t __ret;
19956  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
19957  return __ret;
19958}
19959#endif
19960
19961#ifdef __LITTLE_ENDIAN__
19962__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
19963  uint32x4_t __ret;
19964  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
19965  return __ret;
19966}
19967#else
19968__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
19969  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19970  uint32x4_t __ret;
19971  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
19972  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19973  return __ret;
19974}
19975#endif
19976
19977#ifdef __LITTLE_ENDIAN__
19978__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
19979  float32x4_t __ret;
19980  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
19981  return __ret;
19982}
19983#else
19984__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
19985  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19986  float32x4_t __ret;
19987  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
19988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19989  return __ret;
19990}
19991#endif
19992
19993#ifdef __LITTLE_ENDIAN__
19994__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
19995  uint32x2_t __ret;
19996  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
19997  return __ret;
19998}
19999#else
20000__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20001  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20002  uint32x2_t __ret;
20003  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
20004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20005  return __ret;
20006}
20007#endif
20008
20009#ifdef __LITTLE_ENDIAN__
20010__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20011  float32x2_t __ret;
20012  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
20013  return __ret;
20014}
20015#else
20016__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20017  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20018  float32x2_t __ret;
20019  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
20020  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20021  return __ret;
20022}
20023#endif
20024
20025#ifdef __LITTLE_ENDIAN__
20026__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20027  float32x4_t __ret;
20028  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
20029  return __ret;
20030}
20031#else
20032__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20033  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20034  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20035  float32x4_t __ret;
20036  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
20037  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20038  return __ret;
20039}
20040#endif
20041
20042#ifdef __LITTLE_ENDIAN__
20043__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20044  float32x2_t __ret;
20045  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
20046  return __ret;
20047}
20048#else
20049__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20050  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20051  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20052  float32x2_t __ret;
20053  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
20054  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20055  return __ret;
20056}
20057#endif
20058
20059#ifdef __LITTLE_ENDIAN__
20060__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20061  poly8x8_t __ret;
20062  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20063  return __ret;
20064}
20065#else
20066__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20067  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20068  poly8x8_t __ret;
20069  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20070  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20071  return __ret;
20072}
20073#endif
20074
20075#ifdef __LITTLE_ENDIAN__
20076__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20077  poly8x16_t __ret;
20078  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20079  return __ret;
20080}
20081#else
20082__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20083  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20084  poly8x16_t __ret;
20085  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20086  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20087  return __ret;
20088}
20089#endif
20090
20091#ifdef __LITTLE_ENDIAN__
20092__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20093  uint8x16_t __ret;
20094  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20095  return __ret;
20096}
20097#else
20098__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20099  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20100  uint8x16_t __ret;
20101  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20102  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20103  return __ret;
20104}
20105#endif
20106
20107#ifdef __LITTLE_ENDIAN__
20108__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20109  int8x16_t __ret;
20110  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20111  return __ret;
20112}
20113#else
20114__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20115  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20116  int8x16_t __ret;
20117  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20118  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20119  return __ret;
20120}
20121#endif
20122
20123#ifdef __LITTLE_ENDIAN__
20124__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20125  uint8x8_t __ret;
20126  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20127  return __ret;
20128}
20129#else
20130__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20131  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20132  uint8x8_t __ret;
20133  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20134  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20135  return __ret;
20136}
20137#endif
20138
20139#ifdef __LITTLE_ENDIAN__
20140__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20141  int8x8_t __ret;
20142  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20143  return __ret;
20144}
20145#else
20146__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20147  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20148  int8x8_t __ret;
20149  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20150  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20151  return __ret;
20152}
20153#endif
20154
20155#ifdef __LITTLE_ENDIAN__
20156__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20157  poly8x8_t __ret;
20158  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20159  return __ret;
20160}
20161#else
20162__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20163  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20164  poly8x8_t __ret;
20165  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20166  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20167  return __ret;
20168}
20169#endif
20170
20171#ifdef __LITTLE_ENDIAN__
20172__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20173  poly16x4_t __ret;
20174  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20175  return __ret;
20176}
20177#else
20178__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20179  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20180  poly16x4_t __ret;
20181  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20182  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20183  return __ret;
20184}
20185#endif
20186
20187#ifdef __LITTLE_ENDIAN__
20188__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20189  poly8x16_t __ret;
20190  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20191  return __ret;
20192}
20193#else
20194__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20195  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20196  poly8x16_t __ret;
20197  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20199  return __ret;
20200}
20201#endif
20202
20203#ifdef __LITTLE_ENDIAN__
20204__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20205  poly16x8_t __ret;
20206  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20207  return __ret;
20208}
20209#else
20210__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20211  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20212  poly16x8_t __ret;
20213  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20215  return __ret;
20216}
20217#endif
20218
20219#ifdef __LITTLE_ENDIAN__
20220__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20221  uint8x16_t __ret;
20222  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20223  return __ret;
20224}
20225#else
20226__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20227  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20228  uint8x16_t __ret;
20229  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20230  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20231  return __ret;
20232}
20233#endif
20234
20235#ifdef __LITTLE_ENDIAN__
20236__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20237  uint16x8_t __ret;
20238  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20239  return __ret;
20240}
20241#else
20242__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20243  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20244  uint16x8_t __ret;
20245  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20246  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20247  return __ret;
20248}
20249#endif
20250
20251#ifdef __LITTLE_ENDIAN__
20252__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20253  int8x16_t __ret;
20254  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20255  return __ret;
20256}
20257#else
20258__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20259  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20260  int8x16_t __ret;
20261  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20262  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20263  return __ret;
20264}
20265#endif
20266
20267#ifdef __LITTLE_ENDIAN__
20268__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20269  int16x8_t __ret;
20270  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20271  return __ret;
20272}
20273#else
20274__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20275  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20276  int16x8_t __ret;
20277  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20279  return __ret;
20280}
20281#endif
20282
20283#ifdef __LITTLE_ENDIAN__
20284__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20285  uint8x8_t __ret;
20286  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20287  return __ret;
20288}
20289#else
20290__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20291  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20292  uint8x8_t __ret;
20293  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20295  return __ret;
20296}
20297#endif
20298
20299#ifdef __LITTLE_ENDIAN__
20300__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20301  uint16x4_t __ret;
20302  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20303  return __ret;
20304}
20305#else
20306__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20307  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20308  uint16x4_t __ret;
20309  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20310  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20311  return __ret;
20312}
20313#endif
20314
20315#ifdef __LITTLE_ENDIAN__
20316__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20317  int8x8_t __ret;
20318  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20319  return __ret;
20320}
20321#else
20322__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20323  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20324  int8x8_t __ret;
20325  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20326  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20327  return __ret;
20328}
20329#endif
20330
20331#ifdef __LITTLE_ENDIAN__
20332__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20333  int16x4_t __ret;
20334  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20335  return __ret;
20336}
20337#else
20338__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20339  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20340  int16x4_t __ret;
20341  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20342  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20343  return __ret;
20344}
20345#endif
20346
20347#ifdef __LITTLE_ENDIAN__
20348__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20349  poly8x8_t __ret;
20350  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20351  return __ret;
20352}
20353#else
20354__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20355  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20356  poly8x8_t __ret;
20357  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20359  return __ret;
20360}
20361#endif
20362
20363#ifdef __LITTLE_ENDIAN__
20364__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20365  poly16x4_t __ret;
20366  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20367  return __ret;
20368}
20369#else
20370__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20371  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20372  poly16x4_t __ret;
20373  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20374  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20375  return __ret;
20376}
20377#endif
20378
20379#ifdef __LITTLE_ENDIAN__
20380__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20381  poly8x16_t __ret;
20382  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20383  return __ret;
20384}
20385#else
20386__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20387  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20388  poly8x16_t __ret;
20389  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20390  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20391  return __ret;
20392}
20393#endif
20394
20395#ifdef __LITTLE_ENDIAN__
20396__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20397  poly16x8_t __ret;
20398  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20399  return __ret;
20400}
20401#else
20402__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20403  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20404  poly16x8_t __ret;
20405  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20406  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20407  return __ret;
20408}
20409#endif
20410
20411#ifdef __LITTLE_ENDIAN__
20412__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20413  uint8x16_t __ret;
20414  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20415  return __ret;
20416}
20417#else
20418__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20419  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20420  uint8x16_t __ret;
20421  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20422  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20423  return __ret;
20424}
20425#endif
20426
20427#ifdef __LITTLE_ENDIAN__
20428__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20429  uint32x4_t __ret;
20430  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20431  return __ret;
20432}
20433#else
20434__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20435  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20436  uint32x4_t __ret;
20437  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20438  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20439  return __ret;
20440}
20441#endif
20442
20443#ifdef __LITTLE_ENDIAN__
20444__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20445  uint16x8_t __ret;
20446  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20447  return __ret;
20448}
20449#else
20450__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20451  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20452  uint16x8_t __ret;
20453  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20454  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20455  return __ret;
20456}
20457#endif
20458
20459#ifdef __LITTLE_ENDIAN__
20460__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20461  int8x16_t __ret;
20462  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20463  return __ret;
20464}
20465#else
20466__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20467  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20468  int8x16_t __ret;
20469  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20470  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20471  return __ret;
20472}
20473#endif
20474
20475#ifdef __LITTLE_ENDIAN__
20476__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20477  float32x4_t __ret;
20478  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20479  return __ret;
20480}
20481#else
20482__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20483  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20484  float32x4_t __ret;
20485  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20487  return __ret;
20488}
20489#endif
20490
20491#ifdef __LITTLE_ENDIAN__
20492__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20493  int32x4_t __ret;
20494  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20495  return __ret;
20496}
20497#else
20498__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20499  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20500  int32x4_t __ret;
20501  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20502  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20503  return __ret;
20504}
20505#endif
20506
20507#ifdef __LITTLE_ENDIAN__
20508__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20509  int16x8_t __ret;
20510  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20511  return __ret;
20512}
20513#else
20514__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20515  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20516  int16x8_t __ret;
20517  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20518  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20519  return __ret;
20520}
20521#endif
20522
20523#ifdef __LITTLE_ENDIAN__
20524__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20525  uint8x8_t __ret;
20526  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20527  return __ret;
20528}
20529#else
20530__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20531  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20532  uint8x8_t __ret;
20533  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20535  return __ret;
20536}
20537#endif
20538
20539#ifdef __LITTLE_ENDIAN__
20540__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20541  uint32x2_t __ret;
20542  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20543  return __ret;
20544}
20545#else
20546__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20547  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20548  uint32x2_t __ret;
20549  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20550  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20551  return __ret;
20552}
20553#endif
20554
20555#ifdef __LITTLE_ENDIAN__
20556__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20557  uint16x4_t __ret;
20558  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20559  return __ret;
20560}
20561#else
20562__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20563  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20564  uint16x4_t __ret;
20565  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20567  return __ret;
20568}
20569#endif
20570
20571#ifdef __LITTLE_ENDIAN__
20572__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20573  int8x8_t __ret;
20574  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20575  return __ret;
20576}
20577#else
20578__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20579  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20580  int8x8_t __ret;
20581  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20583  return __ret;
20584}
20585#endif
20586
20587#ifdef __LITTLE_ENDIAN__
20588__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20589  float32x2_t __ret;
20590  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20591  return __ret;
20592}
20593#else
20594__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20595  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20596  float32x2_t __ret;
20597  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20598  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20599  return __ret;
20600}
20601#endif
20602
20603#ifdef __LITTLE_ENDIAN__
20604__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20605  int32x2_t __ret;
20606  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20607  return __ret;
20608}
20609#else
20610__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20611  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20612  int32x2_t __ret;
20613  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20614  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20615  return __ret;
20616}
20617#endif
20618
20619#ifdef __LITTLE_ENDIAN__
20620__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20621  int16x4_t __ret;
20622  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20623  return __ret;
20624}
20625#else
20626__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20627  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20628  int16x4_t __ret;
20629  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20630  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20631  return __ret;
20632}
20633#endif
20634
20635#ifdef __LITTLE_ENDIAN__
20636__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20637  uint8x16_t __ret;
20638  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20639  return __ret;
20640}
20641#else
20642__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20643  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20644  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20645  uint8x16_t __ret;
20646  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20648  return __ret;
20649}
20650#endif
20651
20652#ifdef __LITTLE_ENDIAN__
20653__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20654  uint32x4_t __ret;
20655  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20656  return __ret;
20657}
20658#else
20659__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20660  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20661  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20662  uint32x4_t __ret;
20663  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20665  return __ret;
20666}
20667#endif
20668
20669#ifdef __LITTLE_ENDIAN__
20670__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20671  uint16x8_t __ret;
20672  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20673  return __ret;
20674}
20675#else
20676__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20677  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20678  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20679  uint16x8_t __ret;
20680  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20681  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20682  return __ret;
20683}
20684#endif
20685
20686#ifdef __LITTLE_ENDIAN__
20687__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20688  int8x16_t __ret;
20689  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20690  return __ret;
20691}
20692#else
20693__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20694  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20695  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20696  int8x16_t __ret;
20697  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20698  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20699  return __ret;
20700}
20701#endif
20702
20703#ifdef __LITTLE_ENDIAN__
20704__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20705  int32x4_t __ret;
20706  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20707  return __ret;
20708}
20709#else
20710__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20711  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20712  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20713  int32x4_t __ret;
20714  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20716  return __ret;
20717}
20718#endif
20719
20720#ifdef __LITTLE_ENDIAN__
20721__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20722  int16x8_t __ret;
20723  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20724  return __ret;
20725}
20726#else
20727__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20728  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20729  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20730  int16x8_t __ret;
20731  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20733  return __ret;
20734}
20735#endif
20736
20737#ifdef __LITTLE_ENDIAN__
20738__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20739  uint8x8_t __ret;
20740  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20741  return __ret;
20742}
20743#else
20744__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20745  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20746  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20747  uint8x8_t __ret;
20748  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20750  return __ret;
20751}
20752#endif
20753
20754#ifdef __LITTLE_ENDIAN__
20755__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20756  uint32x2_t __ret;
20757  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20758  return __ret;
20759}
20760#else
20761__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20762  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20763  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20764  uint32x2_t __ret;
20765  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20767  return __ret;
20768}
20769#endif
20770
20771#ifdef __LITTLE_ENDIAN__
20772__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20773  uint16x4_t __ret;
20774  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20775  return __ret;
20776}
20777#else
20778__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20779  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20781  uint16x4_t __ret;
20782  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20784  return __ret;
20785}
20786#endif
20787
20788#ifdef __LITTLE_ENDIAN__
20789__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20790  int8x8_t __ret;
20791  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20792  return __ret;
20793}
20794#else
20795__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20796  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20797  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20798  int8x8_t __ret;
20799  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20801  return __ret;
20802}
20803#endif
20804
20805#ifdef __LITTLE_ENDIAN__
20806__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20807  int32x2_t __ret;
20808  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20809  return __ret;
20810}
20811#else
20812__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20813  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20814  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20815  int32x2_t __ret;
20816  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20818  return __ret;
20819}
20820#endif
20821
20822#ifdef __LITTLE_ENDIAN__
20823__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20824  int16x4_t __ret;
20825  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20826  return __ret;
20827}
20828#else
20829__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20830  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20831  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20832  int16x4_t __ret;
20833  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20834  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20835  return __ret;
20836}
20837#endif
20838
20839#ifdef __LITTLE_ENDIAN__
20840__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20841  uint8x16_t __ret;
20842  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20843  return __ret;
20844}
20845#else
20846__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20847  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20848  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20849  uint8x16_t __ret;
20850  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20851  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20852  return __ret;
20853}
20854#endif
20855
20856#ifdef __LITTLE_ENDIAN__
20857__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20858  uint32x4_t __ret;
20859  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20860  return __ret;
20861}
20862#else
20863__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20864  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20865  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20866  uint32x4_t __ret;
20867  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20869  return __ret;
20870}
20871#endif
20872
20873#ifdef __LITTLE_ENDIAN__
20874__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20875  uint64x2_t __ret;
20876  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20877  return __ret;
20878}
20879#else
20880__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20881  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20882  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20883  uint64x2_t __ret;
20884  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20886  return __ret;
20887}
20888#endif
20889
20890#ifdef __LITTLE_ENDIAN__
20891__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20892  uint16x8_t __ret;
20893  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20894  return __ret;
20895}
20896#else
20897__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20898  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20899  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20900  uint16x8_t __ret;
20901  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20903  return __ret;
20904}
20905#endif
20906
20907#ifdef __LITTLE_ENDIAN__
20908__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20909  int8x16_t __ret;
20910  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20911  return __ret;
20912}
20913#else
20914__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20915  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20916  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20917  int8x16_t __ret;
20918  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20919  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20920  return __ret;
20921}
20922#endif
20923
20924#ifdef __LITTLE_ENDIAN__
20925__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20926  int32x4_t __ret;
20927  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20928  return __ret;
20929}
20930#else
20931__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20932  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20933  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20934  int32x4_t __ret;
20935  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20937  return __ret;
20938}
20939#endif
20940
20941#ifdef __LITTLE_ENDIAN__
20942__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20943  int64x2_t __ret;
20944  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20945  return __ret;
20946}
20947#else
20948__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20949  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20950  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20951  int64x2_t __ret;
20952  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
20953  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20954  return __ret;
20955}
20956#endif
20957
20958#ifdef __LITTLE_ENDIAN__
20959__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20960  int16x8_t __ret;
20961  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20962  return __ret;
20963}
20964#else
20965__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20966  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20967  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20968  int16x8_t __ret;
20969  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20971  return __ret;
20972}
20973#endif
20974
20975#ifdef __LITTLE_ENDIAN__
20976__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20977  uint8x8_t __ret;
20978  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20979  return __ret;
20980}
20981#else
20982__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20983  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20984  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20985  uint8x8_t __ret;
20986  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20987  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20988  return __ret;
20989}
20990#endif
20991
20992#ifdef __LITTLE_ENDIAN__
20993__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20994  uint32x2_t __ret;
20995  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20996  return __ret;
20997}
20998#else
20999__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21001  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21002  uint32x2_t __ret;
21003  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21005  return __ret;
21006}
21007#endif
21008
21009#ifdef __LITTLE_ENDIAN__
21010__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21011  uint64x1_t __ret;
21012  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21013  return __ret;
21014}
21015#else
21016__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21017  uint64x1_t __ret;
21018  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21019  return __ret;
21020}
21021#endif
21022
21023#ifdef __LITTLE_ENDIAN__
21024__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21025  uint16x4_t __ret;
21026  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21027  return __ret;
21028}
21029#else
21030__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21031  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21032  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21033  uint16x4_t __ret;
21034  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21035  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21036  return __ret;
21037}
21038#endif
21039
21040#ifdef __LITTLE_ENDIAN__
21041__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21042  int8x8_t __ret;
21043  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21044  return __ret;
21045}
21046#else
21047__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21048  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21049  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21050  int8x8_t __ret;
21051  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21052  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21053  return __ret;
21054}
21055#endif
21056
21057#ifdef __LITTLE_ENDIAN__
21058__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21059  int32x2_t __ret;
21060  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21061  return __ret;
21062}
21063#else
21064__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21065  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21066  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21067  int32x2_t __ret;
21068  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21069  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21070  return __ret;
21071}
21072#endif
21073
21074#ifdef __LITTLE_ENDIAN__
21075__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21076  int64x1_t __ret;
21077  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21078  return __ret;
21079}
21080#else
21081__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21082  int64x1_t __ret;
21083  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21084  return __ret;
21085}
21086#endif
21087
21088#ifdef __LITTLE_ENDIAN__
21089__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21090  int16x4_t __ret;
21091  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21092  return __ret;
21093}
21094#else
21095__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21096  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21097  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21098  int16x4_t __ret;
21099  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21101  return __ret;
21102}
21103#endif
21104
21105#ifdef __LITTLE_ENDIAN__
21106#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21107  uint8x16_t __s0 = __p0; \
21108  uint8x16_t __ret; \
21109  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
21110  __ret; \
21111})
21112#else
21113#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21114  uint8x16_t __s0 = __p0; \
21115  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21116  uint8x16_t __ret; \
21117  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
21118  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21119  __ret; \
21120})
21121#endif
21122
21123#ifdef __LITTLE_ENDIAN__
21124#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21125  uint32x4_t __s0 = __p0; \
21126  uint32x4_t __ret; \
21127  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
21128  __ret; \
21129})
21130#else
21131#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21132  uint32x4_t __s0 = __p0; \
21133  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21134  uint32x4_t __ret; \
21135  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
21136  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21137  __ret; \
21138})
21139#endif
21140
21141#ifdef __LITTLE_ENDIAN__
21142#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21143  uint64x2_t __s0 = __p0; \
21144  uint64x2_t __ret; \
21145  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
21146  __ret; \
21147})
21148#else
21149#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21150  uint64x2_t __s0 = __p0; \
21151  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21152  uint64x2_t __ret; \
21153  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
21154  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21155  __ret; \
21156})
21157#endif
21158
21159#ifdef __LITTLE_ENDIAN__
21160#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21161  uint16x8_t __s0 = __p0; \
21162  uint16x8_t __ret; \
21163  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
21164  __ret; \
21165})
21166#else
21167#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21168  uint16x8_t __s0 = __p0; \
21169  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21170  uint16x8_t __ret; \
21171  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
21172  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21173  __ret; \
21174})
21175#endif
21176
21177#ifdef __LITTLE_ENDIAN__
21178#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21179  int8x16_t __s0 = __p0; \
21180  int8x16_t __ret; \
21181  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
21182  __ret; \
21183})
21184#else
21185#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21186  int8x16_t __s0 = __p0; \
21187  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21188  int8x16_t __ret; \
21189  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
21190  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21191  __ret; \
21192})
21193#endif
21194
21195#ifdef __LITTLE_ENDIAN__
21196#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21197  int32x4_t __s0 = __p0; \
21198  int32x4_t __ret; \
21199  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
21200  __ret; \
21201})
21202#else
21203#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21204  int32x4_t __s0 = __p0; \
21205  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21206  int32x4_t __ret; \
21207  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
21208  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21209  __ret; \
21210})
21211#endif
21212
21213#ifdef __LITTLE_ENDIAN__
21214#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21215  int64x2_t __s0 = __p0; \
21216  int64x2_t __ret; \
21217  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
21218  __ret; \
21219})
21220#else
21221#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21222  int64x2_t __s0 = __p0; \
21223  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21224  int64x2_t __ret; \
21225  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
21226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21227  __ret; \
21228})
21229#endif
21230
21231#ifdef __LITTLE_ENDIAN__
21232#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21233  int16x8_t __s0 = __p0; \
21234  int16x8_t __ret; \
21235  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
21236  __ret; \
21237})
21238#else
21239#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21240  int16x8_t __s0 = __p0; \
21241  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21242  int16x8_t __ret; \
21243  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
21244  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21245  __ret; \
21246})
21247#endif
21248
21249#ifdef __LITTLE_ENDIAN__
21250#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21251  uint8x8_t __s0 = __p0; \
21252  uint8x8_t __ret; \
21253  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
21254  __ret; \
21255})
21256#else
21257#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21258  uint8x8_t __s0 = __p0; \
21259  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21260  uint8x8_t __ret; \
21261  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
21262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21263  __ret; \
21264})
21265#endif
21266
21267#ifdef __LITTLE_ENDIAN__
21268#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21269  uint32x2_t __s0 = __p0; \
21270  uint32x2_t __ret; \
21271  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
21272  __ret; \
21273})
21274#else
21275#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21276  uint32x2_t __s0 = __p0; \
21277  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21278  uint32x2_t __ret; \
21279  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
21280  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21281  __ret; \
21282})
21283#endif
21284
21285#ifdef __LITTLE_ENDIAN__
21286#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21287  uint64x1_t __s0 = __p0; \
21288  uint64x1_t __ret; \
21289  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21290  __ret; \
21291})
21292#else
21293#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21294  uint64x1_t __s0 = __p0; \
21295  uint64x1_t __ret; \
21296  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21297  __ret; \
21298})
21299#endif
21300
21301#ifdef __LITTLE_ENDIAN__
21302#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21303  uint16x4_t __s0 = __p0; \
21304  uint16x4_t __ret; \
21305  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
21306  __ret; \
21307})
21308#else
21309#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21310  uint16x4_t __s0 = __p0; \
21311  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21312  uint16x4_t __ret; \
21313  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
21314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21315  __ret; \
21316})
21317#endif
21318
21319#ifdef __LITTLE_ENDIAN__
21320#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21321  int8x8_t __s0 = __p0; \
21322  int8x8_t __ret; \
21323  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
21324  __ret; \
21325})
21326#else
21327#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21328  int8x8_t __s0 = __p0; \
21329  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21330  int8x8_t __ret; \
21331  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
21332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21333  __ret; \
21334})
21335#endif
21336
21337#ifdef __LITTLE_ENDIAN__
21338#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21339  int32x2_t __s0 = __p0; \
21340  int32x2_t __ret; \
21341  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
21342  __ret; \
21343})
21344#else
21345#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21346  int32x2_t __s0 = __p0; \
21347  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21348  int32x2_t __ret; \
21349  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
21350  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21351  __ret; \
21352})
21353#endif
21354
21355#ifdef __LITTLE_ENDIAN__
21356#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21357  int64x1_t __s0 = __p0; \
21358  int64x1_t __ret; \
21359  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21360  __ret; \
21361})
21362#else
21363#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21364  int64x1_t __s0 = __p0; \
21365  int64x1_t __ret; \
21366  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21367  __ret; \
21368})
21369#endif
21370
21371#ifdef __LITTLE_ENDIAN__
21372#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21373  int16x4_t __s0 = __p0; \
21374  int16x4_t __ret; \
21375  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
21376  __ret; \
21377})
21378#else
21379#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21380  int16x4_t __s0 = __p0; \
21381  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21382  int16x4_t __ret; \
21383  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
21384  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21385  __ret; \
21386})
21387#endif
21388
21389#ifdef __LITTLE_ENDIAN__
21390#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21391  uint32x4_t __s0 = __p0; \
21392  uint16x4_t __ret; \
21393  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21394  __ret; \
21395})
21396#else
21397#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21398  uint32x4_t __s0 = __p0; \
21399  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21400  uint16x4_t __ret; \
21401  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21402  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21403  __ret; \
21404})
21405#define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21406  uint32x4_t __s0 = __p0; \
21407  uint16x4_t __ret; \
21408  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21409  __ret; \
21410})
21411#endif
21412
21413#ifdef __LITTLE_ENDIAN__
21414#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21415  uint64x2_t __s0 = __p0; \
21416  uint32x2_t __ret; \
21417  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21418  __ret; \
21419})
21420#else
21421#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21422  uint64x2_t __s0 = __p0; \
21423  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21424  uint32x2_t __ret; \
21425  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21427  __ret; \
21428})
21429#define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21430  uint64x2_t __s0 = __p0; \
21431  uint32x2_t __ret; \
21432  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21433  __ret; \
21434})
21435#endif
21436
21437#ifdef __LITTLE_ENDIAN__
21438#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21439  uint16x8_t __s0 = __p0; \
21440  uint8x8_t __ret; \
21441  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21442  __ret; \
21443})
21444#else
21445#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21446  uint16x8_t __s0 = __p0; \
21447  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21448  uint8x8_t __ret; \
21449  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21450  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21451  __ret; \
21452})
21453#define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21454  uint16x8_t __s0 = __p0; \
21455  uint8x8_t __ret; \
21456  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21457  __ret; \
21458})
21459#endif
21460
21461#ifdef __LITTLE_ENDIAN__
21462#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21463  int32x4_t __s0 = __p0; \
21464  int16x4_t __ret; \
21465  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21466  __ret; \
21467})
21468#else
21469#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21470  int32x4_t __s0 = __p0; \
21471  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21472  int16x4_t __ret; \
21473  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21474  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21475  __ret; \
21476})
21477#define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21478  int32x4_t __s0 = __p0; \
21479  int16x4_t __ret; \
21480  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21481  __ret; \
21482})
21483#endif
21484
21485#ifdef __LITTLE_ENDIAN__
21486#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21487  int64x2_t __s0 = __p0; \
21488  int32x2_t __ret; \
21489  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21490  __ret; \
21491})
21492#else
21493#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21494  int64x2_t __s0 = __p0; \
21495  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21496  int32x2_t __ret; \
21497  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21498  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21499  __ret; \
21500})
21501#define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21502  int64x2_t __s0 = __p0; \
21503  int32x2_t __ret; \
21504  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21505  __ret; \
21506})
21507#endif
21508
21509#ifdef __LITTLE_ENDIAN__
21510#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21511  int16x8_t __s0 = __p0; \
21512  int8x8_t __ret; \
21513  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21514  __ret; \
21515})
21516#else
21517#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21518  int16x8_t __s0 = __p0; \
21519  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21520  int8x8_t __ret; \
21521  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21522  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21523  __ret; \
21524})
21525#define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21526  int16x8_t __s0 = __p0; \
21527  int8x8_t __ret; \
21528  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21529  __ret; \
21530})
21531#endif
21532
21533#ifdef __LITTLE_ENDIAN__
21534__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21535  uint32x4_t __ret;
21536  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
21537  return __ret;
21538}
21539#else
21540__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21541  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21542  uint32x4_t __ret;
21543  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
21544  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21545  return __ret;
21546}
21547#endif
21548
21549#ifdef __LITTLE_ENDIAN__
21550__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21551  float32x4_t __ret;
21552  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
21553  return __ret;
21554}
21555#else
21556__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21557  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21558  float32x4_t __ret;
21559  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
21560  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21561  return __ret;
21562}
21563#endif
21564
21565#ifdef __LITTLE_ENDIAN__
21566__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21567  uint32x2_t __ret;
21568  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
21569  return __ret;
21570}
21571#else
21572__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21573  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21574  uint32x2_t __ret;
21575  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
21576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21577  return __ret;
21578}
21579#endif
21580
21581#ifdef __LITTLE_ENDIAN__
21582__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21583  float32x2_t __ret;
21584  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
21585  return __ret;
21586}
21587#else
21588__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21589  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21590  float32x2_t __ret;
21591  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
21592  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21593  return __ret;
21594}
21595#endif
21596
21597#ifdef __LITTLE_ENDIAN__
21598__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21599  float32x4_t __ret;
21600  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21601  return __ret;
21602}
21603#else
21604__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21605  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21606  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21607  float32x4_t __ret;
21608  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21609  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21610  return __ret;
21611}
21612#endif
21613
21614#ifdef __LITTLE_ENDIAN__
21615__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21616  float32x2_t __ret;
21617  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21618  return __ret;
21619}
21620#else
21621__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21622  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21623  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21624  float32x2_t __ret;
21625  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21626  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21627  return __ret;
21628}
21629#endif
21630
21631#ifdef __LITTLE_ENDIAN__
21632#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21633  uint8x16_t __s0 = __p0; \
21634  uint8x16_t __s1 = __p1; \
21635  uint8x16_t __ret; \
21636  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
21637  __ret; \
21638})
21639#else
21640#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21641  uint8x16_t __s0 = __p0; \
21642  uint8x16_t __s1 = __p1; \
21643  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21644  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21645  uint8x16_t __ret; \
21646  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
21647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21648  __ret; \
21649})
21650#endif
21651
21652#ifdef __LITTLE_ENDIAN__
21653#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21654  uint32x4_t __s0 = __p0; \
21655  uint32x4_t __s1 = __p1; \
21656  uint32x4_t __ret; \
21657  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
21658  __ret; \
21659})
21660#else
21661#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21662  uint32x4_t __s0 = __p0; \
21663  uint32x4_t __s1 = __p1; \
21664  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21665  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21666  uint32x4_t __ret; \
21667  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
21668  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21669  __ret; \
21670})
21671#endif
21672
21673#ifdef __LITTLE_ENDIAN__
21674#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21675  uint64x2_t __s0 = __p0; \
21676  uint64x2_t __s1 = __p1; \
21677  uint64x2_t __ret; \
21678  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
21679  __ret; \
21680})
21681#else
21682#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21683  uint64x2_t __s0 = __p0; \
21684  uint64x2_t __s1 = __p1; \
21685  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21686  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21687  uint64x2_t __ret; \
21688  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
21689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21690  __ret; \
21691})
21692#endif
21693
21694#ifdef __LITTLE_ENDIAN__
21695#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21696  uint16x8_t __s0 = __p0; \
21697  uint16x8_t __s1 = __p1; \
21698  uint16x8_t __ret; \
21699  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
21700  __ret; \
21701})
21702#else
21703#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21704  uint16x8_t __s0 = __p0; \
21705  uint16x8_t __s1 = __p1; \
21706  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21707  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21708  uint16x8_t __ret; \
21709  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
21710  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21711  __ret; \
21712})
21713#endif
21714
21715#ifdef __LITTLE_ENDIAN__
21716#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21717  int8x16_t __s0 = __p0; \
21718  int8x16_t __s1 = __p1; \
21719  int8x16_t __ret; \
21720  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
21721  __ret; \
21722})
21723#else
21724#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21725  int8x16_t __s0 = __p0; \
21726  int8x16_t __s1 = __p1; \
21727  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21728  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21729  int8x16_t __ret; \
21730  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
21731  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21732  __ret; \
21733})
21734#endif
21735
21736#ifdef __LITTLE_ENDIAN__
21737#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21738  int32x4_t __s0 = __p0; \
21739  int32x4_t __s1 = __p1; \
21740  int32x4_t __ret; \
21741  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
21742  __ret; \
21743})
21744#else
21745#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21746  int32x4_t __s0 = __p0; \
21747  int32x4_t __s1 = __p1; \
21748  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21749  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21750  int32x4_t __ret; \
21751  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
21752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21753  __ret; \
21754})
21755#endif
21756
21757#ifdef __LITTLE_ENDIAN__
21758#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21759  int64x2_t __s0 = __p0; \
21760  int64x2_t __s1 = __p1; \
21761  int64x2_t __ret; \
21762  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
21763  __ret; \
21764})
21765#else
21766#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21767  int64x2_t __s0 = __p0; \
21768  int64x2_t __s1 = __p1; \
21769  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21770  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21771  int64x2_t __ret; \
21772  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
21773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21774  __ret; \
21775})
21776#endif
21777
21778#ifdef __LITTLE_ENDIAN__
21779#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21780  int16x8_t __s0 = __p0; \
21781  int16x8_t __s1 = __p1; \
21782  int16x8_t __ret; \
21783  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
21784  __ret; \
21785})
21786#else
21787#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21788  int16x8_t __s0 = __p0; \
21789  int16x8_t __s1 = __p1; \
21790  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21791  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21792  int16x8_t __ret; \
21793  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
21794  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21795  __ret; \
21796})
21797#endif
21798
21799#ifdef __LITTLE_ENDIAN__
21800#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21801  uint8x8_t __s0 = __p0; \
21802  uint8x8_t __s1 = __p1; \
21803  uint8x8_t __ret; \
21804  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
21805  __ret; \
21806})
21807#else
21808#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21809  uint8x8_t __s0 = __p0; \
21810  uint8x8_t __s1 = __p1; \
21811  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21812  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21813  uint8x8_t __ret; \
21814  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
21815  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21816  __ret; \
21817})
21818#endif
21819
21820#ifdef __LITTLE_ENDIAN__
21821#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21822  uint32x2_t __s0 = __p0; \
21823  uint32x2_t __s1 = __p1; \
21824  uint32x2_t __ret; \
21825  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
21826  __ret; \
21827})
21828#else
21829#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21830  uint32x2_t __s0 = __p0; \
21831  uint32x2_t __s1 = __p1; \
21832  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21833  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21834  uint32x2_t __ret; \
21835  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
21836  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21837  __ret; \
21838})
21839#endif
21840
21841#ifdef __LITTLE_ENDIAN__
21842#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21843  uint64x1_t __s0 = __p0; \
21844  uint64x1_t __s1 = __p1; \
21845  uint64x1_t __ret; \
21846  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21847  __ret; \
21848})
21849#else
21850#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21851  uint64x1_t __s0 = __p0; \
21852  uint64x1_t __s1 = __p1; \
21853  uint64x1_t __ret; \
21854  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21855  __ret; \
21856})
21857#endif
21858
21859#ifdef __LITTLE_ENDIAN__
21860#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21861  uint16x4_t __s0 = __p0; \
21862  uint16x4_t __s1 = __p1; \
21863  uint16x4_t __ret; \
21864  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
21865  __ret; \
21866})
21867#else
21868#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21869  uint16x4_t __s0 = __p0; \
21870  uint16x4_t __s1 = __p1; \
21871  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21872  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21873  uint16x4_t __ret; \
21874  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
21875  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21876  __ret; \
21877})
21878#endif
21879
21880#ifdef __LITTLE_ENDIAN__
21881#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21882  int8x8_t __s0 = __p0; \
21883  int8x8_t __s1 = __p1; \
21884  int8x8_t __ret; \
21885  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
21886  __ret; \
21887})
21888#else
21889#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21890  int8x8_t __s0 = __p0; \
21891  int8x8_t __s1 = __p1; \
21892  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21893  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21894  int8x8_t __ret; \
21895  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
21896  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21897  __ret; \
21898})
21899#endif
21900
21901#ifdef __LITTLE_ENDIAN__
21902#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21903  int32x2_t __s0 = __p0; \
21904  int32x2_t __s1 = __p1; \
21905  int32x2_t __ret; \
21906  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
21907  __ret; \
21908})
21909#else
21910#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21911  int32x2_t __s0 = __p0; \
21912  int32x2_t __s1 = __p1; \
21913  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21914  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21915  int32x2_t __ret; \
21916  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
21917  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21918  __ret; \
21919})
21920#endif
21921
21922#ifdef __LITTLE_ENDIAN__
21923#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
21924  int64x1_t __s0 = __p0; \
21925  int64x1_t __s1 = __p1; \
21926  int64x1_t __ret; \
21927  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
21928  __ret; \
21929})
21930#else
21931#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
21932  int64x1_t __s0 = __p0; \
21933  int64x1_t __s1 = __p1; \
21934  int64x1_t __ret; \
21935  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
21936  __ret; \
21937})
21938#endif
21939
21940#ifdef __LITTLE_ENDIAN__
21941#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
21942  int16x4_t __s0 = __p0; \
21943  int16x4_t __s1 = __p1; \
21944  int16x4_t __ret; \
21945  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
21946  __ret; \
21947})
21948#else
21949#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
21950  int16x4_t __s0 = __p0; \
21951  int16x4_t __s1 = __p1; \
21952  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21953  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21954  int16x4_t __ret; \
21955  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
21956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21957  __ret; \
21958})
21959#endif
21960
21961#ifdef __LITTLE_ENDIAN__
21962__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21963  uint16x4_t __ret;
21964  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21965  return __ret;
21966}
21967#else
21968__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21969  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21970  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21971  uint16x4_t __ret;
21972  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
21973  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21974  return __ret;
21975}
21976__ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21977  uint16x4_t __ret;
21978  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21979  return __ret;
21980}
21981#endif
21982
21983#ifdef __LITTLE_ENDIAN__
21984__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21985  uint32x2_t __ret;
21986  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
21987  return __ret;
21988}
21989#else
21990__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21991  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21992  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21993  uint32x2_t __ret;
21994  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
21995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21996  return __ret;
21997}
21998__ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21999  uint32x2_t __ret;
22000  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22001  return __ret;
22002}
22003#endif
22004
22005#ifdef __LITTLE_ENDIAN__
22006__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22007  uint8x8_t __ret;
22008  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22009  return __ret;
22010}
22011#else
22012__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22013  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22014  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22015  uint8x8_t __ret;
22016  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
22017  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22018  return __ret;
22019}
22020__ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22021  uint8x8_t __ret;
22022  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22023  return __ret;
22024}
22025#endif
22026
22027#ifdef __LITTLE_ENDIAN__
22028__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22029  int16x4_t __ret;
22030  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22031  return __ret;
22032}
22033#else
22034__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22035  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22036  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22037  int16x4_t __ret;
22038  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
22039  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22040  return __ret;
22041}
22042__ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22043  int16x4_t __ret;
22044  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22045  return __ret;
22046}
22047#endif
22048
22049#ifdef __LITTLE_ENDIAN__
22050__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22051  int32x2_t __ret;
22052  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22053  return __ret;
22054}
22055#else
22056__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22057  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22058  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22059  int32x2_t __ret;
22060  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
22061  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22062  return __ret;
22063}
22064__ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22065  int32x2_t __ret;
22066  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22067  return __ret;
22068}
22069#endif
22070
22071#ifdef __LITTLE_ENDIAN__
22072__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22073  int8x8_t __ret;
22074  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22075  return __ret;
22076}
22077#else
22078__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22079  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22080  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22081  int8x8_t __ret;
22082  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
22083  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22084  return __ret;
22085}
22086__ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22087  int8x8_t __ret;
22088  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22089  return __ret;
22090}
22091#endif
22092
22093#ifdef __LITTLE_ENDIAN__
22094#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22095  poly8_t __s0 = __p0; \
22096  poly8x8_t __s1 = __p1; \
22097  poly8x8_t __ret; \
22098  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22099  __ret; \
22100})
22101#else
22102#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22103  poly8_t __s0 = __p0; \
22104  poly8x8_t __s1 = __p1; \
22105  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22106  poly8x8_t __ret; \
22107  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22108  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22109  __ret; \
22110})
22111#define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22112  poly8_t __s0 = __p0; \
22113  poly8x8_t __s1 = __p1; \
22114  poly8x8_t __ret; \
22115  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22116  __ret; \
22117})
22118#endif
22119
22120#ifdef __LITTLE_ENDIAN__
22121#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22122  poly16_t __s0 = __p0; \
22123  poly16x4_t __s1 = __p1; \
22124  poly16x4_t __ret; \
22125  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22126  __ret; \
22127})
22128#else
22129#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22130  poly16_t __s0 = __p0; \
22131  poly16x4_t __s1 = __p1; \
22132  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22133  poly16x4_t __ret; \
22134  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22135  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22136  __ret; \
22137})
22138#define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22139  poly16_t __s0 = __p0; \
22140  poly16x4_t __s1 = __p1; \
22141  poly16x4_t __ret; \
22142  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22143  __ret; \
22144})
22145#endif
22146
22147#ifdef __LITTLE_ENDIAN__
22148#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22149  poly8_t __s0 = __p0; \
22150  poly8x16_t __s1 = __p1; \
22151  poly8x16_t __ret; \
22152  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22153  __ret; \
22154})
22155#else
22156#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22157  poly8_t __s0 = __p0; \
22158  poly8x16_t __s1 = __p1; \
22159  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22160  poly8x16_t __ret; \
22161  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22162  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22163  __ret; \
22164})
22165#define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22166  poly8_t __s0 = __p0; \
22167  poly8x16_t __s1 = __p1; \
22168  poly8x16_t __ret; \
22169  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22170  __ret; \
22171})
22172#endif
22173
22174#ifdef __LITTLE_ENDIAN__
22175#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22176  poly16_t __s0 = __p0; \
22177  poly16x8_t __s1 = __p1; \
22178  poly16x8_t __ret; \
22179  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22180  __ret; \
22181})
22182#else
22183#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22184  poly16_t __s0 = __p0; \
22185  poly16x8_t __s1 = __p1; \
22186  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22187  poly16x8_t __ret; \
22188  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22189  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22190  __ret; \
22191})
22192#define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22193  poly16_t __s0 = __p0; \
22194  poly16x8_t __s1 = __p1; \
22195  poly16x8_t __ret; \
22196  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22197  __ret; \
22198})
22199#endif
22200
22201#ifdef __LITTLE_ENDIAN__
22202#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22203  uint8_t __s0 = __p0; \
22204  uint8x16_t __s1 = __p1; \
22205  uint8x16_t __ret; \
22206  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22207  __ret; \
22208})
22209#else
22210#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22211  uint8_t __s0 = __p0; \
22212  uint8x16_t __s1 = __p1; \
22213  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22214  uint8x16_t __ret; \
22215  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22216  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22217  __ret; \
22218})
22219#define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22220  uint8_t __s0 = __p0; \
22221  uint8x16_t __s1 = __p1; \
22222  uint8x16_t __ret; \
22223  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22224  __ret; \
22225})
22226#endif
22227
22228#ifdef __LITTLE_ENDIAN__
22229#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22230  uint32_t __s0 = __p0; \
22231  uint32x4_t __s1 = __p1; \
22232  uint32x4_t __ret; \
22233  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22234  __ret; \
22235})
22236#else
22237#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22238  uint32_t __s0 = __p0; \
22239  uint32x4_t __s1 = __p1; \
22240  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22241  uint32x4_t __ret; \
22242  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22244  __ret; \
22245})
22246#define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22247  uint32_t __s0 = __p0; \
22248  uint32x4_t __s1 = __p1; \
22249  uint32x4_t __ret; \
22250  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22251  __ret; \
22252})
22253#endif
22254
22255#ifdef __LITTLE_ENDIAN__
22256#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22257  uint64_t __s0 = __p0; \
22258  uint64x2_t __s1 = __p1; \
22259  uint64x2_t __ret; \
22260  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22261  __ret; \
22262})
22263#else
22264#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22265  uint64_t __s0 = __p0; \
22266  uint64x2_t __s1 = __p1; \
22267  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22268  uint64x2_t __ret; \
22269  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22271  __ret; \
22272})
22273#define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22274  uint64_t __s0 = __p0; \
22275  uint64x2_t __s1 = __p1; \
22276  uint64x2_t __ret; \
22277  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22278  __ret; \
22279})
22280#endif
22281
22282#ifdef __LITTLE_ENDIAN__
22283#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22284  uint16_t __s0 = __p0; \
22285  uint16x8_t __s1 = __p1; \
22286  uint16x8_t __ret; \
22287  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22288  __ret; \
22289})
22290#else
22291#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22292  uint16_t __s0 = __p0; \
22293  uint16x8_t __s1 = __p1; \
22294  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22295  uint16x8_t __ret; \
22296  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22297  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22298  __ret; \
22299})
22300#define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22301  uint16_t __s0 = __p0; \
22302  uint16x8_t __s1 = __p1; \
22303  uint16x8_t __ret; \
22304  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22305  __ret; \
22306})
22307#endif
22308
22309#ifdef __LITTLE_ENDIAN__
22310#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22311  int8_t __s0 = __p0; \
22312  int8x16_t __s1 = __p1; \
22313  int8x16_t __ret; \
22314  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22315  __ret; \
22316})
22317#else
22318#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22319  int8_t __s0 = __p0; \
22320  int8x16_t __s1 = __p1; \
22321  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22322  int8x16_t __ret; \
22323  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22324  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22325  __ret; \
22326})
22327#define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22328  int8_t __s0 = __p0; \
22329  int8x16_t __s1 = __p1; \
22330  int8x16_t __ret; \
22331  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22332  __ret; \
22333})
22334#endif
22335
22336#ifdef __LITTLE_ENDIAN__
22337#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22338  float32_t __s0 = __p0; \
22339  float32x4_t __s1 = __p1; \
22340  float32x4_t __ret; \
22341  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22342  __ret; \
22343})
22344#else
22345#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22346  float32_t __s0 = __p0; \
22347  float32x4_t __s1 = __p1; \
22348  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22349  float32x4_t __ret; \
22350  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__rev1, __p2); \
22351  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22352  __ret; \
22353})
22354#define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22355  float32_t __s0 = __p0; \
22356  float32x4_t __s1 = __p1; \
22357  float32x4_t __ret; \
22358  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22359  __ret; \
22360})
22361#endif
22362
22363#ifdef __LITTLE_ENDIAN__
22364#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22365  int32_t __s0 = __p0; \
22366  int32x4_t __s1 = __p1; \
22367  int32x4_t __ret; \
22368  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22369  __ret; \
22370})
22371#else
22372#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22373  int32_t __s0 = __p0; \
22374  int32x4_t __s1 = __p1; \
22375  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22376  int32x4_t __ret; \
22377  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22378  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22379  __ret; \
22380})
22381#define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22382  int32_t __s0 = __p0; \
22383  int32x4_t __s1 = __p1; \
22384  int32x4_t __ret; \
22385  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22386  __ret; \
22387})
22388#endif
22389
22390#ifdef __LITTLE_ENDIAN__
22391#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22392  int64_t __s0 = __p0; \
22393  int64x2_t __s1 = __p1; \
22394  int64x2_t __ret; \
22395  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22396  __ret; \
22397})
22398#else
22399#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22400  int64_t __s0 = __p0; \
22401  int64x2_t __s1 = __p1; \
22402  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22403  int64x2_t __ret; \
22404  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22406  __ret; \
22407})
22408#define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22409  int64_t __s0 = __p0; \
22410  int64x2_t __s1 = __p1; \
22411  int64x2_t __ret; \
22412  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22413  __ret; \
22414})
22415#endif
22416
22417#ifdef __LITTLE_ENDIAN__
22418#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22419  int16_t __s0 = __p0; \
22420  int16x8_t __s1 = __p1; \
22421  int16x8_t __ret; \
22422  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22423  __ret; \
22424})
22425#else
22426#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22427  int16_t __s0 = __p0; \
22428  int16x8_t __s1 = __p1; \
22429  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22430  int16x8_t __ret; \
22431  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22432  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22433  __ret; \
22434})
22435#define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22436  int16_t __s0 = __p0; \
22437  int16x8_t __s1 = __p1; \
22438  int16x8_t __ret; \
22439  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22440  __ret; \
22441})
22442#endif
22443
22444#ifdef __LITTLE_ENDIAN__
22445#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22446  uint8_t __s0 = __p0; \
22447  uint8x8_t __s1 = __p1; \
22448  uint8x8_t __ret; \
22449  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22450  __ret; \
22451})
22452#else
22453#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22454  uint8_t __s0 = __p0; \
22455  uint8x8_t __s1 = __p1; \
22456  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22457  uint8x8_t __ret; \
22458  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22459  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22460  __ret; \
22461})
22462#define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22463  uint8_t __s0 = __p0; \
22464  uint8x8_t __s1 = __p1; \
22465  uint8x8_t __ret; \
22466  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22467  __ret; \
22468})
22469#endif
22470
22471#ifdef __LITTLE_ENDIAN__
22472#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22473  uint32_t __s0 = __p0; \
22474  uint32x2_t __s1 = __p1; \
22475  uint32x2_t __ret; \
22476  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22477  __ret; \
22478})
22479#else
22480#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22481  uint32_t __s0 = __p0; \
22482  uint32x2_t __s1 = __p1; \
22483  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22484  uint32x2_t __ret; \
22485  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22486  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22487  __ret; \
22488})
22489#define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22490  uint32_t __s0 = __p0; \
22491  uint32x2_t __s1 = __p1; \
22492  uint32x2_t __ret; \
22493  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22494  __ret; \
22495})
22496#endif
22497
22498#ifdef __LITTLE_ENDIAN__
22499#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22500  uint64_t __s0 = __p0; \
22501  uint64x1_t __s1 = __p1; \
22502  uint64x1_t __ret; \
22503  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22504  __ret; \
22505})
22506#else
22507#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22508  uint64_t __s0 = __p0; \
22509  uint64x1_t __s1 = __p1; \
22510  uint64x1_t __ret; \
22511  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22512  __ret; \
22513})
22514#define __noswap_vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22515  uint64_t __s0 = __p0; \
22516  uint64x1_t __s1 = __p1; \
22517  uint64x1_t __ret; \
22518  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22519  __ret; \
22520})
22521#endif
22522
22523#ifdef __LITTLE_ENDIAN__
22524#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22525  uint16_t __s0 = __p0; \
22526  uint16x4_t __s1 = __p1; \
22527  uint16x4_t __ret; \
22528  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22529  __ret; \
22530})
22531#else
22532#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22533  uint16_t __s0 = __p0; \
22534  uint16x4_t __s1 = __p1; \
22535  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22536  uint16x4_t __ret; \
22537  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22538  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22539  __ret; \
22540})
22541#define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22542  uint16_t __s0 = __p0; \
22543  uint16x4_t __s1 = __p1; \
22544  uint16x4_t __ret; \
22545  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22546  __ret; \
22547})
22548#endif
22549
22550#ifdef __LITTLE_ENDIAN__
22551#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22552  int8_t __s0 = __p0; \
22553  int8x8_t __s1 = __p1; \
22554  int8x8_t __ret; \
22555  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22556  __ret; \
22557})
22558#else
22559#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22560  int8_t __s0 = __p0; \
22561  int8x8_t __s1 = __p1; \
22562  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22563  int8x8_t __ret; \
22564  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22565  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22566  __ret; \
22567})
22568#define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22569  int8_t __s0 = __p0; \
22570  int8x8_t __s1 = __p1; \
22571  int8x8_t __ret; \
22572  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22573  __ret; \
22574})
22575#endif
22576
22577#ifdef __LITTLE_ENDIAN__
22578#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22579  float32_t __s0 = __p0; \
22580  float32x2_t __s1 = __p1; \
22581  float32x2_t __ret; \
22582  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22583  __ret; \
22584})
22585#else
22586#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22587  float32_t __s0 = __p0; \
22588  float32x2_t __s1 = __p1; \
22589  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22590  float32x2_t __ret; \
22591  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__rev1, __p2); \
22592  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22593  __ret; \
22594})
22595#define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22596  float32_t __s0 = __p0; \
22597  float32x2_t __s1 = __p1; \
22598  float32x2_t __ret; \
22599  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22600  __ret; \
22601})
22602#endif
22603
22604#ifdef __LITTLE_ENDIAN__
22605#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22606  int32_t __s0 = __p0; \
22607  int32x2_t __s1 = __p1; \
22608  int32x2_t __ret; \
22609  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22610  __ret; \
22611})
22612#else
22613#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22614  int32_t __s0 = __p0; \
22615  int32x2_t __s1 = __p1; \
22616  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22617  int32x2_t __ret; \
22618  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22619  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22620  __ret; \
22621})
22622#define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22623  int32_t __s0 = __p0; \
22624  int32x2_t __s1 = __p1; \
22625  int32x2_t __ret; \
22626  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22627  __ret; \
22628})
22629#endif
22630
22631#ifdef __LITTLE_ENDIAN__
22632#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22633  int64_t __s0 = __p0; \
22634  int64x1_t __s1 = __p1; \
22635  int64x1_t __ret; \
22636  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22637  __ret; \
22638})
22639#else
22640#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22641  int64_t __s0 = __p0; \
22642  int64x1_t __s1 = __p1; \
22643  int64x1_t __ret; \
22644  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22645  __ret; \
22646})
22647#define __noswap_vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22648  int64_t __s0 = __p0; \
22649  int64x1_t __s1 = __p1; \
22650  int64x1_t __ret; \
22651  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22652  __ret; \
22653})
22654#endif
22655
22656#ifdef __LITTLE_ENDIAN__
22657#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22658  int16_t __s0 = __p0; \
22659  int16x4_t __s1 = __p1; \
22660  int16x4_t __ret; \
22661  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22662  __ret; \
22663})
22664#else
22665#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22666  int16_t __s0 = __p0; \
22667  int16x4_t __s1 = __p1; \
22668  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22669  int16x4_t __ret; \
22670  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22671  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22672  __ret; \
22673})
22674#define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22675  int16_t __s0 = __p0; \
22676  int16x4_t __s1 = __p1; \
22677  int16x4_t __ret; \
22678  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22679  __ret; \
22680})
22681#endif
22682
22683#ifdef __LITTLE_ENDIAN__
22684__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22685  uint8x16_t __ret;
22686  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22687  return __ret;
22688}
22689#else
22690__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22691  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22692  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22693  uint8x16_t __ret;
22694  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22695  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22696  return __ret;
22697}
22698#endif
22699
22700#ifdef __LITTLE_ENDIAN__
22701__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22702  uint32x4_t __ret;
22703  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22704  return __ret;
22705}
22706#else
22707__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22708  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22709  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22710  uint32x4_t __ret;
22711  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22712  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22713  return __ret;
22714}
22715#endif
22716
22717#ifdef __LITTLE_ENDIAN__
22718__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22719  uint64x2_t __ret;
22720  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22721  return __ret;
22722}
22723#else
22724__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22725  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22726  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22727  uint64x2_t __ret;
22728  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22729  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22730  return __ret;
22731}
22732#endif
22733
22734#ifdef __LITTLE_ENDIAN__
22735__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22736  uint16x8_t __ret;
22737  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22738  return __ret;
22739}
22740#else
22741__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22742  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22743  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22744  uint16x8_t __ret;
22745  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22746  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22747  return __ret;
22748}
22749#endif
22750
22751#ifdef __LITTLE_ENDIAN__
22752__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22753  int8x16_t __ret;
22754  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22755  return __ret;
22756}
22757#else
22758__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22759  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22760  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22761  int8x16_t __ret;
22762  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22763  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22764  return __ret;
22765}
22766#endif
22767
22768#ifdef __LITTLE_ENDIAN__
22769__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22770  int32x4_t __ret;
22771  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22772  return __ret;
22773}
22774#else
22775__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22776  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22777  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22778  int32x4_t __ret;
22779  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22780  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22781  return __ret;
22782}
22783#endif
22784
22785#ifdef __LITTLE_ENDIAN__
22786__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22787  int64x2_t __ret;
22788  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22789  return __ret;
22790}
22791#else
22792__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22793  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22794  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22795  int64x2_t __ret;
22796  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22797  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22798  return __ret;
22799}
22800#endif
22801
22802#ifdef __LITTLE_ENDIAN__
22803__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22804  int16x8_t __ret;
22805  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22806  return __ret;
22807}
22808#else
22809__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22810  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22811  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22812  int16x8_t __ret;
22813  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22814  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22815  return __ret;
22816}
22817#endif
22818
22819#ifdef __LITTLE_ENDIAN__
22820__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22821  uint8x8_t __ret;
22822  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22823  return __ret;
22824}
22825#else
22826__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22827  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22828  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22829  uint8x8_t __ret;
22830  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22831  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22832  return __ret;
22833}
22834#endif
22835
22836#ifdef __LITTLE_ENDIAN__
22837__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22838  uint32x2_t __ret;
22839  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22840  return __ret;
22841}
22842#else
22843__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22844  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22845  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22846  uint32x2_t __ret;
22847  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22848  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22849  return __ret;
22850}
22851#endif
22852
22853#ifdef __LITTLE_ENDIAN__
22854__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22855  uint64x1_t __ret;
22856  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22857  return __ret;
22858}
22859#else
22860__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22861  uint64x1_t __ret;
22862  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22863  return __ret;
22864}
22865#endif
22866
22867#ifdef __LITTLE_ENDIAN__
22868__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22869  uint16x4_t __ret;
22870  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22871  return __ret;
22872}
22873#else
22874__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22875  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22876  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22877  uint16x4_t __ret;
22878  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22879  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22880  return __ret;
22881}
22882#endif
22883
22884#ifdef __LITTLE_ENDIAN__
22885__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22886  int8x8_t __ret;
22887  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22888  return __ret;
22889}
22890#else
22891__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22892  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22893  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22894  int8x8_t __ret;
22895  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22896  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22897  return __ret;
22898}
22899#endif
22900
22901#ifdef __LITTLE_ENDIAN__
22902__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22903  int32x2_t __ret;
22904  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22905  return __ret;
22906}
22907#else
22908__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22909  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22910  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22911  int32x2_t __ret;
22912  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22914  return __ret;
22915}
22916#endif
22917
22918#ifdef __LITTLE_ENDIAN__
22919__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
22920  int64x1_t __ret;
22921  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22922  return __ret;
22923}
22924#else
22925__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
22926  int64x1_t __ret;
22927  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22928  return __ret;
22929}
22930#endif
22931
22932#ifdef __LITTLE_ENDIAN__
22933__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
22934  int16x4_t __ret;
22935  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22936  return __ret;
22937}
22938#else
22939__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
22940  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22941  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22942  int16x4_t __ret;
22943  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22944  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22945  return __ret;
22946}
22947#endif
22948
22949#ifdef __LITTLE_ENDIAN__
22950#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
22951  uint8x16_t __s0 = __p0; \
22952  uint8x16_t __ret; \
22953  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
22954  __ret; \
22955})
22956#else
22957#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
22958  uint8x16_t __s0 = __p0; \
22959  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22960  uint8x16_t __ret; \
22961  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
22962  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22963  __ret; \
22964})
22965#endif
22966
22967#ifdef __LITTLE_ENDIAN__
22968#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
22969  uint32x4_t __s0 = __p0; \
22970  uint32x4_t __ret; \
22971  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
22972  __ret; \
22973})
22974#else
22975#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
22976  uint32x4_t __s0 = __p0; \
22977  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22978  uint32x4_t __ret; \
22979  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
22980  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22981  __ret; \
22982})
22983#endif
22984
22985#ifdef __LITTLE_ENDIAN__
22986#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
22987  uint64x2_t __s0 = __p0; \
22988  uint64x2_t __ret; \
22989  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
22990  __ret; \
22991})
22992#else
22993#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
22994  uint64x2_t __s0 = __p0; \
22995  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22996  uint64x2_t __ret; \
22997  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
22998  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22999  __ret; \
23000})
23001#endif
23002
23003#ifdef __LITTLE_ENDIAN__
23004#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23005  uint16x8_t __s0 = __p0; \
23006  uint16x8_t __ret; \
23007  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
23008  __ret; \
23009})
23010#else
23011#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23012  uint16x8_t __s0 = __p0; \
23013  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23014  uint16x8_t __ret; \
23015  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
23016  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23017  __ret; \
23018})
23019#endif
23020
23021#ifdef __LITTLE_ENDIAN__
23022#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23023  int8x16_t __s0 = __p0; \
23024  int8x16_t __ret; \
23025  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
23026  __ret; \
23027})
23028#else
23029#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23030  int8x16_t __s0 = __p0; \
23031  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23032  int8x16_t __ret; \
23033  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
23034  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23035  __ret; \
23036})
23037#endif
23038
23039#ifdef __LITTLE_ENDIAN__
23040#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23041  int32x4_t __s0 = __p0; \
23042  int32x4_t __ret; \
23043  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
23044  __ret; \
23045})
23046#else
23047#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23048  int32x4_t __s0 = __p0; \
23049  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23050  int32x4_t __ret; \
23051  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
23052  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23053  __ret; \
23054})
23055#endif
23056
23057#ifdef __LITTLE_ENDIAN__
23058#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23059  int64x2_t __s0 = __p0; \
23060  int64x2_t __ret; \
23061  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
23062  __ret; \
23063})
23064#else
23065#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23066  int64x2_t __s0 = __p0; \
23067  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23068  int64x2_t __ret; \
23069  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
23070  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23071  __ret; \
23072})
23073#endif
23074
23075#ifdef __LITTLE_ENDIAN__
23076#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23077  int16x8_t __s0 = __p0; \
23078  int16x8_t __ret; \
23079  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
23080  __ret; \
23081})
23082#else
23083#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23084  int16x8_t __s0 = __p0; \
23085  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23086  int16x8_t __ret; \
23087  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
23088  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23089  __ret; \
23090})
23091#endif
23092
23093#ifdef __LITTLE_ENDIAN__
23094#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23095  uint8x8_t __s0 = __p0; \
23096  uint8x8_t __ret; \
23097  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
23098  __ret; \
23099})
23100#else
23101#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23102  uint8x8_t __s0 = __p0; \
23103  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23104  uint8x8_t __ret; \
23105  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
23106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23107  __ret; \
23108})
23109#endif
23110
23111#ifdef __LITTLE_ENDIAN__
23112#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23113  uint32x2_t __s0 = __p0; \
23114  uint32x2_t __ret; \
23115  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
23116  __ret; \
23117})
23118#else
23119#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23120  uint32x2_t __s0 = __p0; \
23121  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23122  uint32x2_t __ret; \
23123  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
23124  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23125  __ret; \
23126})
23127#endif
23128
23129#ifdef __LITTLE_ENDIAN__
23130#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23131  uint64x1_t __s0 = __p0; \
23132  uint64x1_t __ret; \
23133  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23134  __ret; \
23135})
23136#else
23137#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23138  uint64x1_t __s0 = __p0; \
23139  uint64x1_t __ret; \
23140  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23141  __ret; \
23142})
23143#endif
23144
23145#ifdef __LITTLE_ENDIAN__
23146#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23147  uint16x4_t __s0 = __p0; \
23148  uint16x4_t __ret; \
23149  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
23150  __ret; \
23151})
23152#else
23153#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23154  uint16x4_t __s0 = __p0; \
23155  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23156  uint16x4_t __ret; \
23157  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
23158  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23159  __ret; \
23160})
23161#endif
23162
23163#ifdef __LITTLE_ENDIAN__
23164#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23165  int8x8_t __s0 = __p0; \
23166  int8x8_t __ret; \
23167  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
23168  __ret; \
23169})
23170#else
23171#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23172  int8x8_t __s0 = __p0; \
23173  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23174  int8x8_t __ret; \
23175  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
23176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23177  __ret; \
23178})
23179#endif
23180
23181#ifdef __LITTLE_ENDIAN__
23182#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23183  int32x2_t __s0 = __p0; \
23184  int32x2_t __ret; \
23185  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
23186  __ret; \
23187})
23188#else
23189#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23190  int32x2_t __s0 = __p0; \
23191  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23192  int32x2_t __ret; \
23193  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
23194  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23195  __ret; \
23196})
23197#endif
23198
23199#ifdef __LITTLE_ENDIAN__
23200#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23201  int64x1_t __s0 = __p0; \
23202  int64x1_t __ret; \
23203  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23204  __ret; \
23205})
23206#else
23207#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23208  int64x1_t __s0 = __p0; \
23209  int64x1_t __ret; \
23210  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23211  __ret; \
23212})
23213#endif
23214
23215#ifdef __LITTLE_ENDIAN__
23216#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23217  int16x4_t __s0 = __p0; \
23218  int16x4_t __ret; \
23219  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
23220  __ret; \
23221})
23222#else
23223#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23224  int16x4_t __s0 = __p0; \
23225  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23226  int16x4_t __ret; \
23227  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
23228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23229  __ret; \
23230})
23231#endif
23232
23233#ifdef __LITTLE_ENDIAN__
23234#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23235  uint8x8_t __s0 = __p0; \
23236  uint16x8_t __ret; \
23237  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23238  __ret; \
23239})
23240#else
23241#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23242  uint8x8_t __s0 = __p0; \
23243  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23244  uint16x8_t __ret; \
23245  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
23246  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23247  __ret; \
23248})
23249#define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
23250  uint8x8_t __s0 = __p0; \
23251  uint16x8_t __ret; \
23252  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23253  __ret; \
23254})
23255#endif
23256
23257#ifdef __LITTLE_ENDIAN__
23258#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23259  uint32x2_t __s0 = __p0; \
23260  uint64x2_t __ret; \
23261  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23262  __ret; \
23263})
23264#else
23265#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23266  uint32x2_t __s0 = __p0; \
23267  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23268  uint64x2_t __ret; \
23269  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
23270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23271  __ret; \
23272})
23273#define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
23274  uint32x2_t __s0 = __p0; \
23275  uint64x2_t __ret; \
23276  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23277  __ret; \
23278})
23279#endif
23280
23281#ifdef __LITTLE_ENDIAN__
23282#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23283  uint16x4_t __s0 = __p0; \
23284  uint32x4_t __ret; \
23285  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23286  __ret; \
23287})
23288#else
23289#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23290  uint16x4_t __s0 = __p0; \
23291  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23292  uint32x4_t __ret; \
23293  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
23294  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23295  __ret; \
23296})
23297#define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
23298  uint16x4_t __s0 = __p0; \
23299  uint32x4_t __ret; \
23300  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23301  __ret; \
23302})
23303#endif
23304
23305#ifdef __LITTLE_ENDIAN__
23306#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23307  int8x8_t __s0 = __p0; \
23308  int16x8_t __ret; \
23309  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23310  __ret; \
23311})
23312#else
23313#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23314  int8x8_t __s0 = __p0; \
23315  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23316  int16x8_t __ret; \
23317  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
23318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23319  __ret; \
23320})
23321#define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
23322  int8x8_t __s0 = __p0; \
23323  int16x8_t __ret; \
23324  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23325  __ret; \
23326})
23327#endif
23328
23329#ifdef __LITTLE_ENDIAN__
23330#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23331  int32x2_t __s0 = __p0; \
23332  int64x2_t __ret; \
23333  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23334  __ret; \
23335})
23336#else
23337#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23338  int32x2_t __s0 = __p0; \
23339  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23340  int64x2_t __ret; \
23341  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
23342  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23343  __ret; \
23344})
23345#define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
23346  int32x2_t __s0 = __p0; \
23347  int64x2_t __ret; \
23348  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23349  __ret; \
23350})
23351#endif
23352
23353#ifdef __LITTLE_ENDIAN__
23354#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23355  int16x4_t __s0 = __p0; \
23356  int32x4_t __ret; \
23357  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23358  __ret; \
23359})
23360#else
23361#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23362  int16x4_t __s0 = __p0; \
23363  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23364  int32x4_t __ret; \
23365  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
23366  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23367  __ret; \
23368})
23369#define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
23370  int16x4_t __s0 = __p0; \
23371  int32x4_t __ret; \
23372  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23373  __ret; \
23374})
23375#endif
23376
23377#ifdef __LITTLE_ENDIAN__
23378#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23379  uint8x16_t __s0 = __p0; \
23380  uint8x16_t __ret; \
23381  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
23382  __ret; \
23383})
23384#else
23385#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23386  uint8x16_t __s0 = __p0; \
23387  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23388  uint8x16_t __ret; \
23389  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
23390  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23391  __ret; \
23392})
23393#endif
23394
23395#ifdef __LITTLE_ENDIAN__
23396#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23397  uint32x4_t __s0 = __p0; \
23398  uint32x4_t __ret; \
23399  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
23400  __ret; \
23401})
23402#else
23403#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23404  uint32x4_t __s0 = __p0; \
23405  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23406  uint32x4_t __ret; \
23407  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
23408  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23409  __ret; \
23410})
23411#endif
23412
23413#ifdef __LITTLE_ENDIAN__
23414#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23415  uint64x2_t __s0 = __p0; \
23416  uint64x2_t __ret; \
23417  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
23418  __ret; \
23419})
23420#else
23421#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23422  uint64x2_t __s0 = __p0; \
23423  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23424  uint64x2_t __ret; \
23425  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
23426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23427  __ret; \
23428})
23429#endif
23430
23431#ifdef __LITTLE_ENDIAN__
23432#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23433  uint16x8_t __s0 = __p0; \
23434  uint16x8_t __ret; \
23435  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
23436  __ret; \
23437})
23438#else
23439#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23440  uint16x8_t __s0 = __p0; \
23441  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23442  uint16x8_t __ret; \
23443  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
23444  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23445  __ret; \
23446})
23447#endif
23448
23449#ifdef __LITTLE_ENDIAN__
23450#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23451  int8x16_t __s0 = __p0; \
23452  int8x16_t __ret; \
23453  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
23454  __ret; \
23455})
23456#else
23457#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23458  int8x16_t __s0 = __p0; \
23459  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23460  int8x16_t __ret; \
23461  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
23462  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23463  __ret; \
23464})
23465#endif
23466
23467#ifdef __LITTLE_ENDIAN__
23468#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23469  int32x4_t __s0 = __p0; \
23470  int32x4_t __ret; \
23471  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
23472  __ret; \
23473})
23474#else
23475#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23476  int32x4_t __s0 = __p0; \
23477  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23478  int32x4_t __ret; \
23479  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
23480  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23481  __ret; \
23482})
23483#endif
23484
23485#ifdef __LITTLE_ENDIAN__
23486#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23487  int64x2_t __s0 = __p0; \
23488  int64x2_t __ret; \
23489  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
23490  __ret; \
23491})
23492#else
23493#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23494  int64x2_t __s0 = __p0; \
23495  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23496  int64x2_t __ret; \
23497  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
23498  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23499  __ret; \
23500})
23501#endif
23502
23503#ifdef __LITTLE_ENDIAN__
23504#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23505  int16x8_t __s0 = __p0; \
23506  int16x8_t __ret; \
23507  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
23508  __ret; \
23509})
23510#else
23511#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23512  int16x8_t __s0 = __p0; \
23513  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23514  int16x8_t __ret; \
23515  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
23516  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23517  __ret; \
23518})
23519#endif
23520
23521#ifdef __LITTLE_ENDIAN__
23522#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23523  uint8x8_t __s0 = __p0; \
23524  uint8x8_t __ret; \
23525  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
23526  __ret; \
23527})
23528#else
23529#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23530  uint8x8_t __s0 = __p0; \
23531  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23532  uint8x8_t __ret; \
23533  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
23534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23535  __ret; \
23536})
23537#endif
23538
23539#ifdef __LITTLE_ENDIAN__
23540#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23541  uint32x2_t __s0 = __p0; \
23542  uint32x2_t __ret; \
23543  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
23544  __ret; \
23545})
23546#else
23547#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23548  uint32x2_t __s0 = __p0; \
23549  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23550  uint32x2_t __ret; \
23551  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
23552  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23553  __ret; \
23554})
23555#endif
23556
23557#ifdef __LITTLE_ENDIAN__
23558#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23559  uint64x1_t __s0 = __p0; \
23560  uint64x1_t __ret; \
23561  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23562  __ret; \
23563})
23564#else
23565#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23566  uint64x1_t __s0 = __p0; \
23567  uint64x1_t __ret; \
23568  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23569  __ret; \
23570})
23571#endif
23572
23573#ifdef __LITTLE_ENDIAN__
23574#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23575  uint16x4_t __s0 = __p0; \
23576  uint16x4_t __ret; \
23577  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
23578  __ret; \
23579})
23580#else
23581#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23582  uint16x4_t __s0 = __p0; \
23583  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23584  uint16x4_t __ret; \
23585  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
23586  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23587  __ret; \
23588})
23589#endif
23590
23591#ifdef __LITTLE_ENDIAN__
23592#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23593  int8x8_t __s0 = __p0; \
23594  int8x8_t __ret; \
23595  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
23596  __ret; \
23597})
23598#else
23599#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23600  int8x8_t __s0 = __p0; \
23601  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23602  int8x8_t __ret; \
23603  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
23604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23605  __ret; \
23606})
23607#endif
23608
23609#ifdef __LITTLE_ENDIAN__
23610#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23611  int32x2_t __s0 = __p0; \
23612  int32x2_t __ret; \
23613  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
23614  __ret; \
23615})
23616#else
23617#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23618  int32x2_t __s0 = __p0; \
23619  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23620  int32x2_t __ret; \
23621  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
23622  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23623  __ret; \
23624})
23625#endif
23626
23627#ifdef __LITTLE_ENDIAN__
23628#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23629  int64x1_t __s0 = __p0; \
23630  int64x1_t __ret; \
23631  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23632  __ret; \
23633})
23634#else
23635#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23636  int64x1_t __s0 = __p0; \
23637  int64x1_t __ret; \
23638  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23639  __ret; \
23640})
23641#endif
23642
23643#ifdef __LITTLE_ENDIAN__
23644#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23645  int16x4_t __s0 = __p0; \
23646  int16x4_t __ret; \
23647  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
23648  __ret; \
23649})
23650#else
23651#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23652  int16x4_t __s0 = __p0; \
23653  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23654  int16x4_t __ret; \
23655  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
23656  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23657  __ret; \
23658})
23659#endif
23660
23661#ifdef __LITTLE_ENDIAN__
23662#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23663  uint32x4_t __s0 = __p0; \
23664  uint16x4_t __ret; \
23665  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23666  __ret; \
23667})
23668#else
23669#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23670  uint32x4_t __s0 = __p0; \
23671  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23672  uint16x4_t __ret; \
23673  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23675  __ret; \
23676})
23677#define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
23678  uint32x4_t __s0 = __p0; \
23679  uint16x4_t __ret; \
23680  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23681  __ret; \
23682})
23683#endif
23684
23685#ifdef __LITTLE_ENDIAN__
23686#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23687  uint64x2_t __s0 = __p0; \
23688  uint32x2_t __ret; \
23689  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23690  __ret; \
23691})
23692#else
23693#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23694  uint64x2_t __s0 = __p0; \
23695  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23696  uint32x2_t __ret; \
23697  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23699  __ret; \
23700})
23701#define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
23702  uint64x2_t __s0 = __p0; \
23703  uint32x2_t __ret; \
23704  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23705  __ret; \
23706})
23707#endif
23708
23709#ifdef __LITTLE_ENDIAN__
23710#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23711  uint16x8_t __s0 = __p0; \
23712  uint8x8_t __ret; \
23713  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23714  __ret; \
23715})
23716#else
23717#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23718  uint16x8_t __s0 = __p0; \
23719  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23720  uint8x8_t __ret; \
23721  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23722  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23723  __ret; \
23724})
23725#define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
23726  uint16x8_t __s0 = __p0; \
23727  uint8x8_t __ret; \
23728  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23729  __ret; \
23730})
23731#endif
23732
23733#ifdef __LITTLE_ENDIAN__
23734#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23735  int32x4_t __s0 = __p0; \
23736  int16x4_t __ret; \
23737  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23738  __ret; \
23739})
23740#else
23741#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23742  int32x4_t __s0 = __p0; \
23743  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23744  int16x4_t __ret; \
23745  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23746  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23747  __ret; \
23748})
23749#define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
23750  int32x4_t __s0 = __p0; \
23751  int16x4_t __ret; \
23752  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23753  __ret; \
23754})
23755#endif
23756
23757#ifdef __LITTLE_ENDIAN__
23758#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23759  int64x2_t __s0 = __p0; \
23760  int32x2_t __ret; \
23761  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23762  __ret; \
23763})
23764#else
23765#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23766  int64x2_t __s0 = __p0; \
23767  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23768  int32x2_t __ret; \
23769  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23770  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23771  __ret; \
23772})
23773#define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
23774  int64x2_t __s0 = __p0; \
23775  int32x2_t __ret; \
23776  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23777  __ret; \
23778})
23779#endif
23780
23781#ifdef __LITTLE_ENDIAN__
23782#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23783  int16x8_t __s0 = __p0; \
23784  int8x8_t __ret; \
23785  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23786  __ret; \
23787})
23788#else
23789#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23790  int16x8_t __s0 = __p0; \
23791  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23792  int8x8_t __ret; \
23793  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23794  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23795  __ret; \
23796})
23797#define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
23798  int16x8_t __s0 = __p0; \
23799  int8x8_t __ret; \
23800  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23801  __ret; \
23802})
23803#endif
23804
23805#ifdef __LITTLE_ENDIAN__
23806#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23807  poly8x8_t __s0 = __p0; \
23808  poly8x8_t __s1 = __p1; \
23809  poly8x8_t __ret; \
23810  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
23811  __ret; \
23812})
23813#else
23814#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23815  poly8x8_t __s0 = __p0; \
23816  poly8x8_t __s1 = __p1; \
23817  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23818  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23819  poly8x8_t __ret; \
23820  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
23821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23822  __ret; \
23823})
23824#endif
23825
23826#ifdef __LITTLE_ENDIAN__
23827#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23828  poly16x4_t __s0 = __p0; \
23829  poly16x4_t __s1 = __p1; \
23830  poly16x4_t __ret; \
23831  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
23832  __ret; \
23833})
23834#else
23835#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23836  poly16x4_t __s0 = __p0; \
23837  poly16x4_t __s1 = __p1; \
23838  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23839  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23840  poly16x4_t __ret; \
23841  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
23842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23843  __ret; \
23844})
23845#endif
23846
23847#ifdef __LITTLE_ENDIAN__
23848#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23849  poly8x16_t __s0 = __p0; \
23850  poly8x16_t __s1 = __p1; \
23851  poly8x16_t __ret; \
23852  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
23853  __ret; \
23854})
23855#else
23856#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23857  poly8x16_t __s0 = __p0; \
23858  poly8x16_t __s1 = __p1; \
23859  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23860  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23861  poly8x16_t __ret; \
23862  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
23863  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23864  __ret; \
23865})
23866#endif
23867
23868#ifdef __LITTLE_ENDIAN__
23869#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23870  poly16x8_t __s0 = __p0; \
23871  poly16x8_t __s1 = __p1; \
23872  poly16x8_t __ret; \
23873  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
23874  __ret; \
23875})
23876#else
23877#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23878  poly16x8_t __s0 = __p0; \
23879  poly16x8_t __s1 = __p1; \
23880  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23881  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23882  poly16x8_t __ret; \
23883  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
23884  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23885  __ret; \
23886})
23887#endif
23888
23889#ifdef __LITTLE_ENDIAN__
23890#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23891  uint8x16_t __s0 = __p0; \
23892  uint8x16_t __s1 = __p1; \
23893  uint8x16_t __ret; \
23894  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23895  __ret; \
23896})
23897#else
23898#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23899  uint8x16_t __s0 = __p0; \
23900  uint8x16_t __s1 = __p1; \
23901  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23902  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23903  uint8x16_t __ret; \
23904  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23905  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23906  __ret; \
23907})
23908#endif
23909
23910#ifdef __LITTLE_ENDIAN__
23911#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23912  uint32x4_t __s0 = __p0; \
23913  uint32x4_t __s1 = __p1; \
23914  uint32x4_t __ret; \
23915  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
23916  __ret; \
23917})
23918#else
23919#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23920  uint32x4_t __s0 = __p0; \
23921  uint32x4_t __s1 = __p1; \
23922  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23923  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23924  uint32x4_t __ret; \
23925  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
23926  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23927  __ret; \
23928})
23929#endif
23930
23931#ifdef __LITTLE_ENDIAN__
23932#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23933  uint64x2_t __s0 = __p0; \
23934  uint64x2_t __s1 = __p1; \
23935  uint64x2_t __ret; \
23936  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
23937  __ret; \
23938})
23939#else
23940#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23941  uint64x2_t __s0 = __p0; \
23942  uint64x2_t __s1 = __p1; \
23943  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23944  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23945  uint64x2_t __ret; \
23946  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
23947  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23948  __ret; \
23949})
23950#endif
23951
23952#ifdef __LITTLE_ENDIAN__
23953#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23954  uint16x8_t __s0 = __p0; \
23955  uint16x8_t __s1 = __p1; \
23956  uint16x8_t __ret; \
23957  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
23958  __ret; \
23959})
23960#else
23961#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23962  uint16x8_t __s0 = __p0; \
23963  uint16x8_t __s1 = __p1; \
23964  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23965  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23966  uint16x8_t __ret; \
23967  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
23968  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23969  __ret; \
23970})
23971#endif
23972
23973#ifdef __LITTLE_ENDIAN__
23974#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23975  int8x16_t __s0 = __p0; \
23976  int8x16_t __s1 = __p1; \
23977  int8x16_t __ret; \
23978  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
23979  __ret; \
23980})
23981#else
23982#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23983  int8x16_t __s0 = __p0; \
23984  int8x16_t __s1 = __p1; \
23985  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23986  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23987  int8x16_t __ret; \
23988  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
23989  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23990  __ret; \
23991})
23992#endif
23993
23994#ifdef __LITTLE_ENDIAN__
23995#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
23996  int32x4_t __s0 = __p0; \
23997  int32x4_t __s1 = __p1; \
23998  int32x4_t __ret; \
23999  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24000  __ret; \
24001})
24002#else
24003#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24004  int32x4_t __s0 = __p0; \
24005  int32x4_t __s1 = __p1; \
24006  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24007  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24008  int32x4_t __ret; \
24009  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24010  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24011  __ret; \
24012})
24013#endif
24014
24015#ifdef __LITTLE_ENDIAN__
24016#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24017  int64x2_t __s0 = __p0; \
24018  int64x2_t __s1 = __p1; \
24019  int64x2_t __ret; \
24020  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24021  __ret; \
24022})
24023#else
24024#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24025  int64x2_t __s0 = __p0; \
24026  int64x2_t __s1 = __p1; \
24027  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24028  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24029  int64x2_t __ret; \
24030  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24032  __ret; \
24033})
24034#endif
24035
24036#ifdef __LITTLE_ENDIAN__
24037#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24038  int16x8_t __s0 = __p0; \
24039  int16x8_t __s1 = __p1; \
24040  int16x8_t __ret; \
24041  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24042  __ret; \
24043})
24044#else
24045#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24046  int16x8_t __s0 = __p0; \
24047  int16x8_t __s1 = __p1; \
24048  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24049  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24050  int16x8_t __ret; \
24051  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24052  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24053  __ret; \
24054})
24055#endif
24056
24057#ifdef __LITTLE_ENDIAN__
24058#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24059  uint8x8_t __s0 = __p0; \
24060  uint8x8_t __s1 = __p1; \
24061  uint8x8_t __ret; \
24062  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24063  __ret; \
24064})
24065#else
24066#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24067  uint8x8_t __s0 = __p0; \
24068  uint8x8_t __s1 = __p1; \
24069  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24070  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24071  uint8x8_t __ret; \
24072  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24073  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24074  __ret; \
24075})
24076#endif
24077
24078#ifdef __LITTLE_ENDIAN__
24079#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24080  uint32x2_t __s0 = __p0; \
24081  uint32x2_t __s1 = __p1; \
24082  uint32x2_t __ret; \
24083  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24084  __ret; \
24085})
24086#else
24087#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24088  uint32x2_t __s0 = __p0; \
24089  uint32x2_t __s1 = __p1; \
24090  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24091  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24092  uint32x2_t __ret; \
24093  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24094  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24095  __ret; \
24096})
24097#endif
24098
24099#ifdef __LITTLE_ENDIAN__
24100#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24101  uint64x1_t __s0 = __p0; \
24102  uint64x1_t __s1 = __p1; \
24103  uint64x1_t __ret; \
24104  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24105  __ret; \
24106})
24107#else
24108#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24109  uint64x1_t __s0 = __p0; \
24110  uint64x1_t __s1 = __p1; \
24111  uint64x1_t __ret; \
24112  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24113  __ret; \
24114})
24115#endif
24116
24117#ifdef __LITTLE_ENDIAN__
24118#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24119  uint16x4_t __s0 = __p0; \
24120  uint16x4_t __s1 = __p1; \
24121  uint16x4_t __ret; \
24122  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24123  __ret; \
24124})
24125#else
24126#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24127  uint16x4_t __s0 = __p0; \
24128  uint16x4_t __s1 = __p1; \
24129  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24130  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24131  uint16x4_t __ret; \
24132  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24133  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24134  __ret; \
24135})
24136#endif
24137
24138#ifdef __LITTLE_ENDIAN__
24139#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24140  int8x8_t __s0 = __p0; \
24141  int8x8_t __s1 = __p1; \
24142  int8x8_t __ret; \
24143  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24144  __ret; \
24145})
24146#else
24147#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24148  int8x8_t __s0 = __p0; \
24149  int8x8_t __s1 = __p1; \
24150  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24151  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24152  int8x8_t __ret; \
24153  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24154  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24155  __ret; \
24156})
24157#endif
24158
24159#ifdef __LITTLE_ENDIAN__
24160#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24161  int32x2_t __s0 = __p0; \
24162  int32x2_t __s1 = __p1; \
24163  int32x2_t __ret; \
24164  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24165  __ret; \
24166})
24167#else
24168#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24169  int32x2_t __s0 = __p0; \
24170  int32x2_t __s1 = __p1; \
24171  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24172  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24173  int32x2_t __ret; \
24174  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24175  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24176  __ret; \
24177})
24178#endif
24179
24180#ifdef __LITTLE_ENDIAN__
24181#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24182  int64x1_t __s0 = __p0; \
24183  int64x1_t __s1 = __p1; \
24184  int64x1_t __ret; \
24185  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24186  __ret; \
24187})
24188#else
24189#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24190  int64x1_t __s0 = __p0; \
24191  int64x1_t __s1 = __p1; \
24192  int64x1_t __ret; \
24193  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24194  __ret; \
24195})
24196#endif
24197
24198#ifdef __LITTLE_ENDIAN__
24199#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24200  int16x4_t __s0 = __p0; \
24201  int16x4_t __s1 = __p1; \
24202  int16x4_t __ret; \
24203  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24204  __ret; \
24205})
24206#else
24207#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24208  int16x4_t __s0 = __p0; \
24209  int16x4_t __s1 = __p1; \
24210  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24211  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24212  int16x4_t __ret; \
24213  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24214  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24215  __ret; \
24216})
24217#endif
24218
24219#ifdef __LITTLE_ENDIAN__
24220#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24221  uint8x16_t __s0 = __p0; \
24222  uint8x16_t __s1 = __p1; \
24223  uint8x16_t __ret; \
24224  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24225  __ret; \
24226})
24227#else
24228#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24229  uint8x16_t __s0 = __p0; \
24230  uint8x16_t __s1 = __p1; \
24231  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24232  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24233  uint8x16_t __ret; \
24234  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24235  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24236  __ret; \
24237})
24238#endif
24239
24240#ifdef __LITTLE_ENDIAN__
24241#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24242  uint32x4_t __s0 = __p0; \
24243  uint32x4_t __s1 = __p1; \
24244  uint32x4_t __ret; \
24245  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24246  __ret; \
24247})
24248#else
24249#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24250  uint32x4_t __s0 = __p0; \
24251  uint32x4_t __s1 = __p1; \
24252  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24253  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24254  uint32x4_t __ret; \
24255  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24257  __ret; \
24258})
24259#endif
24260
24261#ifdef __LITTLE_ENDIAN__
24262#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24263  uint64x2_t __s0 = __p0; \
24264  uint64x2_t __s1 = __p1; \
24265  uint64x2_t __ret; \
24266  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24267  __ret; \
24268})
24269#else
24270#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24271  uint64x2_t __s0 = __p0; \
24272  uint64x2_t __s1 = __p1; \
24273  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24274  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24275  uint64x2_t __ret; \
24276  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24277  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24278  __ret; \
24279})
24280#endif
24281
24282#ifdef __LITTLE_ENDIAN__
24283#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24284  uint16x8_t __s0 = __p0; \
24285  uint16x8_t __s1 = __p1; \
24286  uint16x8_t __ret; \
24287  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24288  __ret; \
24289})
24290#else
24291#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24292  uint16x8_t __s0 = __p0; \
24293  uint16x8_t __s1 = __p1; \
24294  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24295  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24296  uint16x8_t __ret; \
24297  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24299  __ret; \
24300})
24301#endif
24302
24303#ifdef __LITTLE_ENDIAN__
24304#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24305  int8x16_t __s0 = __p0; \
24306  int8x16_t __s1 = __p1; \
24307  int8x16_t __ret; \
24308  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24309  __ret; \
24310})
24311#else
24312#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24313  int8x16_t __s0 = __p0; \
24314  int8x16_t __s1 = __p1; \
24315  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24316  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24317  int8x16_t __ret; \
24318  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24319  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24320  __ret; \
24321})
24322#endif
24323
24324#ifdef __LITTLE_ENDIAN__
24325#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24326  int32x4_t __s0 = __p0; \
24327  int32x4_t __s1 = __p1; \
24328  int32x4_t __ret; \
24329  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24330  __ret; \
24331})
24332#else
24333#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24334  int32x4_t __s0 = __p0; \
24335  int32x4_t __s1 = __p1; \
24336  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24337  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24338  int32x4_t __ret; \
24339  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24340  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24341  __ret; \
24342})
24343#endif
24344
24345#ifdef __LITTLE_ENDIAN__
24346#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24347  int64x2_t __s0 = __p0; \
24348  int64x2_t __s1 = __p1; \
24349  int64x2_t __ret; \
24350  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24351  __ret; \
24352})
24353#else
24354#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24355  int64x2_t __s0 = __p0; \
24356  int64x2_t __s1 = __p1; \
24357  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24358  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24359  int64x2_t __ret; \
24360  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24361  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24362  __ret; \
24363})
24364#endif
24365
24366#ifdef __LITTLE_ENDIAN__
24367#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24368  int16x8_t __s0 = __p0; \
24369  int16x8_t __s1 = __p1; \
24370  int16x8_t __ret; \
24371  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24372  __ret; \
24373})
24374#else
24375#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24376  int16x8_t __s0 = __p0; \
24377  int16x8_t __s1 = __p1; \
24378  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24379  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24380  int16x8_t __ret; \
24381  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24382  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24383  __ret; \
24384})
24385#endif
24386
24387#ifdef __LITTLE_ENDIAN__
24388#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24389  uint8x8_t __s0 = __p0; \
24390  uint8x8_t __s1 = __p1; \
24391  uint8x8_t __ret; \
24392  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24393  __ret; \
24394})
24395#else
24396#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24397  uint8x8_t __s0 = __p0; \
24398  uint8x8_t __s1 = __p1; \
24399  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24400  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24401  uint8x8_t __ret; \
24402  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24403  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24404  __ret; \
24405})
24406#endif
24407
24408#ifdef __LITTLE_ENDIAN__
24409#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24410  uint32x2_t __s0 = __p0; \
24411  uint32x2_t __s1 = __p1; \
24412  uint32x2_t __ret; \
24413  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24414  __ret; \
24415})
24416#else
24417#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24418  uint32x2_t __s0 = __p0; \
24419  uint32x2_t __s1 = __p1; \
24420  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24421  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24422  uint32x2_t __ret; \
24423  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24424  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24425  __ret; \
24426})
24427#endif
24428
24429#ifdef __LITTLE_ENDIAN__
24430#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24431  uint64x1_t __s0 = __p0; \
24432  uint64x1_t __s1 = __p1; \
24433  uint64x1_t __ret; \
24434  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24435  __ret; \
24436})
24437#else
24438#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24439  uint64x1_t __s0 = __p0; \
24440  uint64x1_t __s1 = __p1; \
24441  uint64x1_t __ret; \
24442  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24443  __ret; \
24444})
24445#endif
24446
24447#ifdef __LITTLE_ENDIAN__
24448#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24449  uint16x4_t __s0 = __p0; \
24450  uint16x4_t __s1 = __p1; \
24451  uint16x4_t __ret; \
24452  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24453  __ret; \
24454})
24455#else
24456#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24457  uint16x4_t __s0 = __p0; \
24458  uint16x4_t __s1 = __p1; \
24459  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24460  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24461  uint16x4_t __ret; \
24462  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24463  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24464  __ret; \
24465})
24466#endif
24467
24468#ifdef __LITTLE_ENDIAN__
24469#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24470  int8x8_t __s0 = __p0; \
24471  int8x8_t __s1 = __p1; \
24472  int8x8_t __ret; \
24473  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24474  __ret; \
24475})
24476#else
24477#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24478  int8x8_t __s0 = __p0; \
24479  int8x8_t __s1 = __p1; \
24480  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24481  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24482  int8x8_t __ret; \
24483  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24484  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24485  __ret; \
24486})
24487#endif
24488
24489#ifdef __LITTLE_ENDIAN__
24490#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24491  int32x2_t __s0 = __p0; \
24492  int32x2_t __s1 = __p1; \
24493  int32x2_t __ret; \
24494  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24495  __ret; \
24496})
24497#else
24498#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24499  int32x2_t __s0 = __p0; \
24500  int32x2_t __s1 = __p1; \
24501  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24502  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24503  int32x2_t __ret; \
24504  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24505  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24506  __ret; \
24507})
24508#endif
24509
24510#ifdef __LITTLE_ENDIAN__
24511#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24512  int64x1_t __s0 = __p0; \
24513  int64x1_t __s1 = __p1; \
24514  int64x1_t __ret; \
24515  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24516  __ret; \
24517})
24518#else
24519#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24520  int64x1_t __s0 = __p0; \
24521  int64x1_t __s1 = __p1; \
24522  int64x1_t __ret; \
24523  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24524  __ret; \
24525})
24526#endif
24527
24528#ifdef __LITTLE_ENDIAN__
24529#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24530  int16x4_t __s0 = __p0; \
24531  int16x4_t __s1 = __p1; \
24532  int16x4_t __ret; \
24533  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24534  __ret; \
24535})
24536#else
24537#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24538  int16x4_t __s0 = __p0; \
24539  int16x4_t __s1 = __p1; \
24540  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24541  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24542  int16x4_t __ret; \
24543  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24544  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24545  __ret; \
24546})
24547#endif
24548
24549#ifdef __LITTLE_ENDIAN__
24550#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24551  poly8x8_t __s0 = __p0; \
24552  poly8x8_t __s1 = __p1; \
24553  poly8x8_t __ret; \
24554  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
24555  __ret; \
24556})
24557#else
24558#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24559  poly8x8_t __s0 = __p0; \
24560  poly8x8_t __s1 = __p1; \
24561  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24562  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24563  poly8x8_t __ret; \
24564  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
24565  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24566  __ret; \
24567})
24568#endif
24569
24570#ifdef __LITTLE_ENDIAN__
24571#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24572  poly16x4_t __s0 = __p0; \
24573  poly16x4_t __s1 = __p1; \
24574  poly16x4_t __ret; \
24575  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
24576  __ret; \
24577})
24578#else
24579#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24580  poly16x4_t __s0 = __p0; \
24581  poly16x4_t __s1 = __p1; \
24582  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24583  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24584  poly16x4_t __ret; \
24585  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
24586  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24587  __ret; \
24588})
24589#endif
24590
24591#ifdef __LITTLE_ENDIAN__
24592#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24593  poly8x16_t __s0 = __p0; \
24594  poly8x16_t __s1 = __p1; \
24595  poly8x16_t __ret; \
24596  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
24597  __ret; \
24598})
24599#else
24600#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24601  poly8x16_t __s0 = __p0; \
24602  poly8x16_t __s1 = __p1; \
24603  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24604  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24605  poly8x16_t __ret; \
24606  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
24607  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24608  __ret; \
24609})
24610#endif
24611
24612#ifdef __LITTLE_ENDIAN__
24613#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24614  poly16x8_t __s0 = __p0; \
24615  poly16x8_t __s1 = __p1; \
24616  poly16x8_t __ret; \
24617  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
24618  __ret; \
24619})
24620#else
24621#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24622  poly16x8_t __s0 = __p0; \
24623  poly16x8_t __s1 = __p1; \
24624  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24625  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24626  poly16x8_t __ret; \
24627  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
24628  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24629  __ret; \
24630})
24631#endif
24632
24633#ifdef __LITTLE_ENDIAN__
24634#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24635  uint8x16_t __s0 = __p0; \
24636  uint8x16_t __s1 = __p1; \
24637  uint8x16_t __ret; \
24638  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24639  __ret; \
24640})
24641#else
24642#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24643  uint8x16_t __s0 = __p0; \
24644  uint8x16_t __s1 = __p1; \
24645  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24646  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24647  uint8x16_t __ret; \
24648  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24649  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24650  __ret; \
24651})
24652#endif
24653
24654#ifdef __LITTLE_ENDIAN__
24655#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24656  uint32x4_t __s0 = __p0; \
24657  uint32x4_t __s1 = __p1; \
24658  uint32x4_t __ret; \
24659  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24660  __ret; \
24661})
24662#else
24663#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24664  uint32x4_t __s0 = __p0; \
24665  uint32x4_t __s1 = __p1; \
24666  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24667  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24668  uint32x4_t __ret; \
24669  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24670  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24671  __ret; \
24672})
24673#endif
24674
24675#ifdef __LITTLE_ENDIAN__
24676#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24677  uint64x2_t __s0 = __p0; \
24678  uint64x2_t __s1 = __p1; \
24679  uint64x2_t __ret; \
24680  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24681  __ret; \
24682})
24683#else
24684#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24685  uint64x2_t __s0 = __p0; \
24686  uint64x2_t __s1 = __p1; \
24687  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24688  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24689  uint64x2_t __ret; \
24690  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24691  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24692  __ret; \
24693})
24694#endif
24695
24696#ifdef __LITTLE_ENDIAN__
24697#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24698  uint16x8_t __s0 = __p0; \
24699  uint16x8_t __s1 = __p1; \
24700  uint16x8_t __ret; \
24701  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24702  __ret; \
24703})
24704#else
24705#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24706  uint16x8_t __s0 = __p0; \
24707  uint16x8_t __s1 = __p1; \
24708  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24709  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24710  uint16x8_t __ret; \
24711  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24712  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24713  __ret; \
24714})
24715#endif
24716
24717#ifdef __LITTLE_ENDIAN__
24718#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24719  int8x16_t __s0 = __p0; \
24720  int8x16_t __s1 = __p1; \
24721  int8x16_t __ret; \
24722  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24723  __ret; \
24724})
24725#else
24726#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24727  int8x16_t __s0 = __p0; \
24728  int8x16_t __s1 = __p1; \
24729  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24730  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24731  int8x16_t __ret; \
24732  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24733  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24734  __ret; \
24735})
24736#endif
24737
24738#ifdef __LITTLE_ENDIAN__
24739#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24740  int32x4_t __s0 = __p0; \
24741  int32x4_t __s1 = __p1; \
24742  int32x4_t __ret; \
24743  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24744  __ret; \
24745})
24746#else
24747#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24748  int32x4_t __s0 = __p0; \
24749  int32x4_t __s1 = __p1; \
24750  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24751  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24752  int32x4_t __ret; \
24753  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24754  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24755  __ret; \
24756})
24757#endif
24758
24759#ifdef __LITTLE_ENDIAN__
24760#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24761  int64x2_t __s0 = __p0; \
24762  int64x2_t __s1 = __p1; \
24763  int64x2_t __ret; \
24764  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24765  __ret; \
24766})
24767#else
24768#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24769  int64x2_t __s0 = __p0; \
24770  int64x2_t __s1 = __p1; \
24771  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24772  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24773  int64x2_t __ret; \
24774  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24775  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24776  __ret; \
24777})
24778#endif
24779
24780#ifdef __LITTLE_ENDIAN__
24781#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24782  int16x8_t __s0 = __p0; \
24783  int16x8_t __s1 = __p1; \
24784  int16x8_t __ret; \
24785  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24786  __ret; \
24787})
24788#else
24789#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24790  int16x8_t __s0 = __p0; \
24791  int16x8_t __s1 = __p1; \
24792  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24793  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24794  int16x8_t __ret; \
24795  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24796  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24797  __ret; \
24798})
24799#endif
24800
24801#ifdef __LITTLE_ENDIAN__
24802#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24803  uint8x8_t __s0 = __p0; \
24804  uint8x8_t __s1 = __p1; \
24805  uint8x8_t __ret; \
24806  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24807  __ret; \
24808})
24809#else
24810#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24811  uint8x8_t __s0 = __p0; \
24812  uint8x8_t __s1 = __p1; \
24813  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24814  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24815  uint8x8_t __ret; \
24816  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24817  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24818  __ret; \
24819})
24820#endif
24821
24822#ifdef __LITTLE_ENDIAN__
24823#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24824  uint32x2_t __s0 = __p0; \
24825  uint32x2_t __s1 = __p1; \
24826  uint32x2_t __ret; \
24827  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24828  __ret; \
24829})
24830#else
24831#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24832  uint32x2_t __s0 = __p0; \
24833  uint32x2_t __s1 = __p1; \
24834  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24835  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24836  uint32x2_t __ret; \
24837  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24838  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24839  __ret; \
24840})
24841#endif
24842
24843#ifdef __LITTLE_ENDIAN__
24844#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24845  uint64x1_t __s0 = __p0; \
24846  uint64x1_t __s1 = __p1; \
24847  uint64x1_t __ret; \
24848  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24849  __ret; \
24850})
24851#else
24852#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24853  uint64x1_t __s0 = __p0; \
24854  uint64x1_t __s1 = __p1; \
24855  uint64x1_t __ret; \
24856  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24857  __ret; \
24858})
24859#endif
24860
24861#ifdef __LITTLE_ENDIAN__
24862#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24863  uint16x4_t __s0 = __p0; \
24864  uint16x4_t __s1 = __p1; \
24865  uint16x4_t __ret; \
24866  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24867  __ret; \
24868})
24869#else
24870#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24871  uint16x4_t __s0 = __p0; \
24872  uint16x4_t __s1 = __p1; \
24873  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24874  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24875  uint16x4_t __ret; \
24876  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24877  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24878  __ret; \
24879})
24880#endif
24881
24882#ifdef __LITTLE_ENDIAN__
24883#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24884  int8x8_t __s0 = __p0; \
24885  int8x8_t __s1 = __p1; \
24886  int8x8_t __ret; \
24887  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24888  __ret; \
24889})
24890#else
24891#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24892  int8x8_t __s0 = __p0; \
24893  int8x8_t __s1 = __p1; \
24894  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24895  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24896  int8x8_t __ret; \
24897  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24898  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24899  __ret; \
24900})
24901#endif
24902
24903#ifdef __LITTLE_ENDIAN__
24904#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24905  int32x2_t __s0 = __p0; \
24906  int32x2_t __s1 = __p1; \
24907  int32x2_t __ret; \
24908  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24909  __ret; \
24910})
24911#else
24912#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24913  int32x2_t __s0 = __p0; \
24914  int32x2_t __s1 = __p1; \
24915  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24916  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24917  int32x2_t __ret; \
24918  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24920  __ret; \
24921})
24922#endif
24923
24924#ifdef __LITTLE_ENDIAN__
24925#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
24926  int64x1_t __s0 = __p0; \
24927  int64x1_t __s1 = __p1; \
24928  int64x1_t __ret; \
24929  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24930  __ret; \
24931})
24932#else
24933#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
24934  int64x1_t __s0 = __p0; \
24935  int64x1_t __s1 = __p1; \
24936  int64x1_t __ret; \
24937  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24938  __ret; \
24939})
24940#endif
24941
24942#ifdef __LITTLE_ENDIAN__
24943#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
24944  int16x4_t __s0 = __p0; \
24945  int16x4_t __s1 = __p1; \
24946  int16x4_t __ret; \
24947  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24948  __ret; \
24949})
24950#else
24951#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
24952  int16x4_t __s0 = __p0; \
24953  int16x4_t __s1 = __p1; \
24954  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24955  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24956  int16x4_t __ret; \
24957  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24958  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24959  __ret; \
24960})
24961#endif
24962
24963#ifdef __LITTLE_ENDIAN__
24964#define vst1_p8(__p0, __p1) __extension__ ({ \
24965  poly8x8_t __s1 = __p1; \
24966  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
24967})
24968#else
24969#define vst1_p8(__p0, __p1) __extension__ ({ \
24970  poly8x8_t __s1 = __p1; \
24971  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24972  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
24973})
24974#endif
24975
24976#ifdef __LITTLE_ENDIAN__
24977#define vst1_p16(__p0, __p1) __extension__ ({ \
24978  poly16x4_t __s1 = __p1; \
24979  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
24980})
24981#else
24982#define vst1_p16(__p0, __p1) __extension__ ({ \
24983  poly16x4_t __s1 = __p1; \
24984  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24985  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
24986})
24987#endif
24988
24989#ifdef __LITTLE_ENDIAN__
24990#define vst1q_p8(__p0, __p1) __extension__ ({ \
24991  poly8x16_t __s1 = __p1; \
24992  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
24993})
24994#else
24995#define vst1q_p8(__p0, __p1) __extension__ ({ \
24996  poly8x16_t __s1 = __p1; \
24997  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24998  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
24999})
25000#endif
25001
25002#ifdef __LITTLE_ENDIAN__
25003#define vst1q_p16(__p0, __p1) __extension__ ({ \
25004  poly16x8_t __s1 = __p1; \
25005  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
25006})
25007#else
25008#define vst1q_p16(__p0, __p1) __extension__ ({ \
25009  poly16x8_t __s1 = __p1; \
25010  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25011  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
25012})
25013#endif
25014
25015#ifdef __LITTLE_ENDIAN__
25016#define vst1q_u8(__p0, __p1) __extension__ ({ \
25017  uint8x16_t __s1 = __p1; \
25018  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
25019})
25020#else
25021#define vst1q_u8(__p0, __p1) __extension__ ({ \
25022  uint8x16_t __s1 = __p1; \
25023  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25024  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
25025})
25026#endif
25027
25028#ifdef __LITTLE_ENDIAN__
25029#define vst1q_u32(__p0, __p1) __extension__ ({ \
25030  uint32x4_t __s1 = __p1; \
25031  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
25032})
25033#else
25034#define vst1q_u32(__p0, __p1) __extension__ ({ \
25035  uint32x4_t __s1 = __p1; \
25036  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25037  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
25038})
25039#endif
25040
25041#ifdef __LITTLE_ENDIAN__
25042#define vst1q_u64(__p0, __p1) __extension__ ({ \
25043  uint64x2_t __s1 = __p1; \
25044  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
25045})
25046#else
25047#define vst1q_u64(__p0, __p1) __extension__ ({ \
25048  uint64x2_t __s1 = __p1; \
25049  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25050  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
25051})
25052#endif
25053
25054#ifdef __LITTLE_ENDIAN__
25055#define vst1q_u16(__p0, __p1) __extension__ ({ \
25056  uint16x8_t __s1 = __p1; \
25057  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
25058})
25059#else
25060#define vst1q_u16(__p0, __p1) __extension__ ({ \
25061  uint16x8_t __s1 = __p1; \
25062  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25063  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
25064})
25065#endif
25066
25067#ifdef __LITTLE_ENDIAN__
25068#define vst1q_s8(__p0, __p1) __extension__ ({ \
25069  int8x16_t __s1 = __p1; \
25070  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
25071})
25072#else
25073#define vst1q_s8(__p0, __p1) __extension__ ({ \
25074  int8x16_t __s1 = __p1; \
25075  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25076  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
25077})
25078#endif
25079
25080#ifdef __LITTLE_ENDIAN__
25081#define vst1q_f32(__p0, __p1) __extension__ ({ \
25082  float32x4_t __s1 = __p1; \
25083  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
25084})
25085#else
25086#define vst1q_f32(__p0, __p1) __extension__ ({ \
25087  float32x4_t __s1 = __p1; \
25088  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25089  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
25090})
25091#endif
25092
25093#ifdef __LITTLE_ENDIAN__
25094#define vst1q_f16(__p0, __p1) __extension__ ({ \
25095  float16x8_t __s1 = __p1; \
25096  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
25097})
25098#else
25099#define vst1q_f16(__p0, __p1) __extension__ ({ \
25100  float16x8_t __s1 = __p1; \
25101  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25102  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
25103})
25104#endif
25105
25106#ifdef __LITTLE_ENDIAN__
25107#define vst1q_s32(__p0, __p1) __extension__ ({ \
25108  int32x4_t __s1 = __p1; \
25109  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
25110})
25111#else
25112#define vst1q_s32(__p0, __p1) __extension__ ({ \
25113  int32x4_t __s1 = __p1; \
25114  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25115  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
25116})
25117#endif
25118
25119#ifdef __LITTLE_ENDIAN__
25120#define vst1q_s64(__p0, __p1) __extension__ ({ \
25121  int64x2_t __s1 = __p1; \
25122  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
25123})
25124#else
25125#define vst1q_s64(__p0, __p1) __extension__ ({ \
25126  int64x2_t __s1 = __p1; \
25127  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25128  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
25129})
25130#endif
25131
25132#ifdef __LITTLE_ENDIAN__
25133#define vst1q_s16(__p0, __p1) __extension__ ({ \
25134  int16x8_t __s1 = __p1; \
25135  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
25136})
25137#else
25138#define vst1q_s16(__p0, __p1) __extension__ ({ \
25139  int16x8_t __s1 = __p1; \
25140  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25141  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
25142})
25143#endif
25144
25145#ifdef __LITTLE_ENDIAN__
25146#define vst1_u8(__p0, __p1) __extension__ ({ \
25147  uint8x8_t __s1 = __p1; \
25148  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
25149})
25150#else
25151#define vst1_u8(__p0, __p1) __extension__ ({ \
25152  uint8x8_t __s1 = __p1; \
25153  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25154  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
25155})
25156#endif
25157
25158#ifdef __LITTLE_ENDIAN__
25159#define vst1_u32(__p0, __p1) __extension__ ({ \
25160  uint32x2_t __s1 = __p1; \
25161  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
25162})
25163#else
25164#define vst1_u32(__p0, __p1) __extension__ ({ \
25165  uint32x2_t __s1 = __p1; \
25166  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25167  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
25168})
25169#endif
25170
25171#ifdef __LITTLE_ENDIAN__
25172#define vst1_u64(__p0, __p1) __extension__ ({ \
25173  uint64x1_t __s1 = __p1; \
25174  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25175})
25176#else
25177#define vst1_u64(__p0, __p1) __extension__ ({ \
25178  uint64x1_t __s1 = __p1; \
25179  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25180})
25181#endif
25182
25183#ifdef __LITTLE_ENDIAN__
25184#define vst1_u16(__p0, __p1) __extension__ ({ \
25185  uint16x4_t __s1 = __p1; \
25186  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
25187})
25188#else
25189#define vst1_u16(__p0, __p1) __extension__ ({ \
25190  uint16x4_t __s1 = __p1; \
25191  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25192  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
25193})
25194#endif
25195
25196#ifdef __LITTLE_ENDIAN__
25197#define vst1_s8(__p0, __p1) __extension__ ({ \
25198  int8x8_t __s1 = __p1; \
25199  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
25200})
25201#else
25202#define vst1_s8(__p0, __p1) __extension__ ({ \
25203  int8x8_t __s1 = __p1; \
25204  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25205  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
25206})
25207#endif
25208
25209#ifdef __LITTLE_ENDIAN__
25210#define vst1_f32(__p0, __p1) __extension__ ({ \
25211  float32x2_t __s1 = __p1; \
25212  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
25213})
25214#else
25215#define vst1_f32(__p0, __p1) __extension__ ({ \
25216  float32x2_t __s1 = __p1; \
25217  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25218  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
25219})
25220#endif
25221
25222#ifdef __LITTLE_ENDIAN__
25223#define vst1_f16(__p0, __p1) __extension__ ({ \
25224  float16x4_t __s1 = __p1; \
25225  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
25226})
25227#else
25228#define vst1_f16(__p0, __p1) __extension__ ({ \
25229  float16x4_t __s1 = __p1; \
25230  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25231  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
25232})
25233#endif
25234
25235#ifdef __LITTLE_ENDIAN__
25236#define vst1_s32(__p0, __p1) __extension__ ({ \
25237  int32x2_t __s1 = __p1; \
25238  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
25239})
25240#else
25241#define vst1_s32(__p0, __p1) __extension__ ({ \
25242  int32x2_t __s1 = __p1; \
25243  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25244  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
25245})
25246#endif
25247
25248#ifdef __LITTLE_ENDIAN__
25249#define vst1_s64(__p0, __p1) __extension__ ({ \
25250  int64x1_t __s1 = __p1; \
25251  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25252})
25253#else
25254#define vst1_s64(__p0, __p1) __extension__ ({ \
25255  int64x1_t __s1 = __p1; \
25256  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25257})
25258#endif
25259
25260#ifdef __LITTLE_ENDIAN__
25261#define vst1_s16(__p0, __p1) __extension__ ({ \
25262  int16x4_t __s1 = __p1; \
25263  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
25264})
25265#else
25266#define vst1_s16(__p0, __p1) __extension__ ({ \
25267  int16x4_t __s1 = __p1; \
25268  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25269  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
25270})
25271#endif
25272
25273#ifdef __LITTLE_ENDIAN__
25274#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25275  poly8x8_t __s1 = __p1; \
25276  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
25277})
25278#else
25279#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25280  poly8x8_t __s1 = __p1; \
25281  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25282  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
25283})
25284#endif
25285
25286#ifdef __LITTLE_ENDIAN__
25287#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25288  poly16x4_t __s1 = __p1; \
25289  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
25290})
25291#else
25292#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25293  poly16x4_t __s1 = __p1; \
25294  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25295  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
25296})
25297#endif
25298
25299#ifdef __LITTLE_ENDIAN__
25300#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25301  poly8x16_t __s1 = __p1; \
25302  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
25303})
25304#else
25305#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25306  poly8x16_t __s1 = __p1; \
25307  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25308  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
25309})
25310#endif
25311
25312#ifdef __LITTLE_ENDIAN__
25313#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25314  poly16x8_t __s1 = __p1; \
25315  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
25316})
25317#else
25318#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25319  poly16x8_t __s1 = __p1; \
25320  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25321  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
25322})
25323#endif
25324
25325#ifdef __LITTLE_ENDIAN__
25326#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25327  uint8x16_t __s1 = __p1; \
25328  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
25329})
25330#else
25331#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25332  uint8x16_t __s1 = __p1; \
25333  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25334  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
25335})
25336#endif
25337
25338#ifdef __LITTLE_ENDIAN__
25339#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25340  uint32x4_t __s1 = __p1; \
25341  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
25342})
25343#else
25344#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25345  uint32x4_t __s1 = __p1; \
25346  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25347  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
25348})
25349#endif
25350
25351#ifdef __LITTLE_ENDIAN__
25352#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25353  uint64x2_t __s1 = __p1; \
25354  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
25355})
25356#else
25357#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25358  uint64x2_t __s1 = __p1; \
25359  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25360  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
25361})
25362#endif
25363
25364#ifdef __LITTLE_ENDIAN__
25365#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25366  uint16x8_t __s1 = __p1; \
25367  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
25368})
25369#else
25370#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25371  uint16x8_t __s1 = __p1; \
25372  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25373  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
25374})
25375#endif
25376
25377#ifdef __LITTLE_ENDIAN__
25378#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25379  int8x16_t __s1 = __p1; \
25380  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
25381})
25382#else
25383#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25384  int8x16_t __s1 = __p1; \
25385  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25386  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
25387})
25388#endif
25389
25390#ifdef __LITTLE_ENDIAN__
25391#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25392  float32x4_t __s1 = __p1; \
25393  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
25394})
25395#else
25396#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25397  float32x4_t __s1 = __p1; \
25398  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25399  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
25400})
25401#endif
25402
25403#ifdef __LITTLE_ENDIAN__
25404#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25405  float16x8_t __s1 = __p1; \
25406  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
25407})
25408#else
25409#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25410  float16x8_t __s1 = __p1; \
25411  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25412  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
25413})
25414#endif
25415
25416#ifdef __LITTLE_ENDIAN__
25417#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25418  int32x4_t __s1 = __p1; \
25419  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
25420})
25421#else
25422#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25423  int32x4_t __s1 = __p1; \
25424  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25425  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
25426})
25427#endif
25428
25429#ifdef __LITTLE_ENDIAN__
25430#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25431  int64x2_t __s1 = __p1; \
25432  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
25433})
25434#else
25435#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25436  int64x2_t __s1 = __p1; \
25437  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25438  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
25439})
25440#endif
25441
25442#ifdef __LITTLE_ENDIAN__
25443#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25444  int16x8_t __s1 = __p1; \
25445  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
25446})
25447#else
25448#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25449  int16x8_t __s1 = __p1; \
25450  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25451  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
25452})
25453#endif
25454
25455#ifdef __LITTLE_ENDIAN__
25456#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25457  uint8x8_t __s1 = __p1; \
25458  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
25459})
25460#else
25461#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25462  uint8x8_t __s1 = __p1; \
25463  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25464  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
25465})
25466#endif
25467
25468#ifdef __LITTLE_ENDIAN__
25469#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25470  uint32x2_t __s1 = __p1; \
25471  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
25472})
25473#else
25474#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25475  uint32x2_t __s1 = __p1; \
25476  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25477  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
25478})
25479#endif
25480
25481#ifdef __LITTLE_ENDIAN__
25482#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25483  uint64x1_t __s1 = __p1; \
25484  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25485})
25486#else
25487#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25488  uint64x1_t __s1 = __p1; \
25489  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25490})
25491#endif
25492
25493#ifdef __LITTLE_ENDIAN__
25494#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25495  uint16x4_t __s1 = __p1; \
25496  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
25497})
25498#else
25499#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25500  uint16x4_t __s1 = __p1; \
25501  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25502  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
25503})
25504#endif
25505
25506#ifdef __LITTLE_ENDIAN__
25507#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25508  int8x8_t __s1 = __p1; \
25509  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
25510})
25511#else
25512#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25513  int8x8_t __s1 = __p1; \
25514  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25515  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
25516})
25517#endif
25518
25519#ifdef __LITTLE_ENDIAN__
25520#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25521  float32x2_t __s1 = __p1; \
25522  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
25523})
25524#else
25525#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25526  float32x2_t __s1 = __p1; \
25527  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25528  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
25529})
25530#endif
25531
25532#ifdef __LITTLE_ENDIAN__
25533#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25534  float16x4_t __s1 = __p1; \
25535  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
25536})
25537#else
25538#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25539  float16x4_t __s1 = __p1; \
25540  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25541  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
25542})
25543#endif
25544
25545#ifdef __LITTLE_ENDIAN__
25546#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25547  int32x2_t __s1 = __p1; \
25548  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
25549})
25550#else
25551#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25552  int32x2_t __s1 = __p1; \
25553  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25554  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
25555})
25556#endif
25557
25558#ifdef __LITTLE_ENDIAN__
25559#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25560  int64x1_t __s1 = __p1; \
25561  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25562})
25563#else
25564#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25565  int64x1_t __s1 = __p1; \
25566  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25567})
25568#endif
25569
25570#ifdef __LITTLE_ENDIAN__
25571#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25572  int16x4_t __s1 = __p1; \
25573  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
25574})
25575#else
25576#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25577  int16x4_t __s1 = __p1; \
25578  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25579  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
25580})
25581#endif
25582
25583#ifdef __LITTLE_ENDIAN__
25584#define vst2_p8(__p0, __p1) __extension__ ({ \
25585  poly8x8x2_t __s1 = __p1; \
25586  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
25587})
25588#else
25589#define vst2_p8(__p0, __p1) __extension__ ({ \
25590  poly8x8x2_t __s1 = __p1; \
25591  poly8x8x2_t __rev1; \
25592  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25593  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25594  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
25595})
25596#endif
25597
25598#ifdef __LITTLE_ENDIAN__
25599#define vst2_p16(__p0, __p1) __extension__ ({ \
25600  poly16x4x2_t __s1 = __p1; \
25601  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
25602})
25603#else
25604#define vst2_p16(__p0, __p1) __extension__ ({ \
25605  poly16x4x2_t __s1 = __p1; \
25606  poly16x4x2_t __rev1; \
25607  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25608  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25609  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
25610})
25611#endif
25612
25613#ifdef __LITTLE_ENDIAN__
25614#define vst2q_p8(__p0, __p1) __extension__ ({ \
25615  poly8x16x2_t __s1 = __p1; \
25616  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
25617})
25618#else
25619#define vst2q_p8(__p0, __p1) __extension__ ({ \
25620  poly8x16x2_t __s1 = __p1; \
25621  poly8x16x2_t __rev1; \
25622  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25623  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25624  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
25625})
25626#endif
25627
25628#ifdef __LITTLE_ENDIAN__
25629#define vst2q_p16(__p0, __p1) __extension__ ({ \
25630  poly16x8x2_t __s1 = __p1; \
25631  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
25632})
25633#else
25634#define vst2q_p16(__p0, __p1) __extension__ ({ \
25635  poly16x8x2_t __s1 = __p1; \
25636  poly16x8x2_t __rev1; \
25637  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25638  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25639  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
25640})
25641#endif
25642
25643#ifdef __LITTLE_ENDIAN__
25644#define vst2q_u8(__p0, __p1) __extension__ ({ \
25645  uint8x16x2_t __s1 = __p1; \
25646  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
25647})
25648#else
25649#define vst2q_u8(__p0, __p1) __extension__ ({ \
25650  uint8x16x2_t __s1 = __p1; \
25651  uint8x16x2_t __rev1; \
25652  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25653  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25654  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
25655})
25656#endif
25657
25658#ifdef __LITTLE_ENDIAN__
25659#define vst2q_u32(__p0, __p1) __extension__ ({ \
25660  uint32x4x2_t __s1 = __p1; \
25661  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
25662})
25663#else
25664#define vst2q_u32(__p0, __p1) __extension__ ({ \
25665  uint32x4x2_t __s1 = __p1; \
25666  uint32x4x2_t __rev1; \
25667  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25668  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25669  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
25670})
25671#endif
25672
25673#ifdef __LITTLE_ENDIAN__
25674#define vst2q_u16(__p0, __p1) __extension__ ({ \
25675  uint16x8x2_t __s1 = __p1; \
25676  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
25677})
25678#else
25679#define vst2q_u16(__p0, __p1) __extension__ ({ \
25680  uint16x8x2_t __s1 = __p1; \
25681  uint16x8x2_t __rev1; \
25682  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25683  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25684  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
25685})
25686#endif
25687
25688#ifdef __LITTLE_ENDIAN__
25689#define vst2q_s8(__p0, __p1) __extension__ ({ \
25690  int8x16x2_t __s1 = __p1; \
25691  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
25692})
25693#else
25694#define vst2q_s8(__p0, __p1) __extension__ ({ \
25695  int8x16x2_t __s1 = __p1; \
25696  int8x16x2_t __rev1; \
25697  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25698  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25699  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
25700})
25701#endif
25702
25703#ifdef __LITTLE_ENDIAN__
25704#define vst2q_f32(__p0, __p1) __extension__ ({ \
25705  float32x4x2_t __s1 = __p1; \
25706  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 41); \
25707})
25708#else
25709#define vst2q_f32(__p0, __p1) __extension__ ({ \
25710  float32x4x2_t __s1 = __p1; \
25711  float32x4x2_t __rev1; \
25712  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25713  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25714  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
25715})
25716#endif
25717
25718#ifdef __LITTLE_ENDIAN__
25719#define vst2q_f16(__p0, __p1) __extension__ ({ \
25720  float16x8x2_t __s1 = __p1; \
25721  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 40); \
25722})
25723#else
25724#define vst2q_f16(__p0, __p1) __extension__ ({ \
25725  float16x8x2_t __s1 = __p1; \
25726  float16x8x2_t __rev1; \
25727  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25728  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25729  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
25730})
25731#endif
25732
25733#ifdef __LITTLE_ENDIAN__
25734#define vst2q_s32(__p0, __p1) __extension__ ({ \
25735  int32x4x2_t __s1 = __p1; \
25736  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 34); \
25737})
25738#else
25739#define vst2q_s32(__p0, __p1) __extension__ ({ \
25740  int32x4x2_t __s1 = __p1; \
25741  int32x4x2_t __rev1; \
25742  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25743  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25744  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
25745})
25746#endif
25747
25748#ifdef __LITTLE_ENDIAN__
25749#define vst2q_s16(__p0, __p1) __extension__ ({ \
25750  int16x8x2_t __s1 = __p1; \
25751  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 33); \
25752})
25753#else
25754#define vst2q_s16(__p0, __p1) __extension__ ({ \
25755  int16x8x2_t __s1 = __p1; \
25756  int16x8x2_t __rev1; \
25757  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25758  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25759  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
25760})
25761#endif
25762
25763#ifdef __LITTLE_ENDIAN__
25764#define vst2_u8(__p0, __p1) __extension__ ({ \
25765  uint8x8x2_t __s1 = __p1; \
25766  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
25767})
25768#else
25769#define vst2_u8(__p0, __p1) __extension__ ({ \
25770  uint8x8x2_t __s1 = __p1; \
25771  uint8x8x2_t __rev1; \
25772  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25773  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25774  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
25775})
25776#endif
25777
25778#ifdef __LITTLE_ENDIAN__
25779#define vst2_u32(__p0, __p1) __extension__ ({ \
25780  uint32x2x2_t __s1 = __p1; \
25781  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
25782})
25783#else
25784#define vst2_u32(__p0, __p1) __extension__ ({ \
25785  uint32x2x2_t __s1 = __p1; \
25786  uint32x2x2_t __rev1; \
25787  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25788  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25789  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
25790})
25791#endif
25792
25793#ifdef __LITTLE_ENDIAN__
25794#define vst2_u64(__p0, __p1) __extension__ ({ \
25795  uint64x1x2_t __s1 = __p1; \
25796  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25797})
25798#else
25799#define vst2_u64(__p0, __p1) __extension__ ({ \
25800  uint64x1x2_t __s1 = __p1; \
25801  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25802})
25803#endif
25804
25805#ifdef __LITTLE_ENDIAN__
25806#define vst2_u16(__p0, __p1) __extension__ ({ \
25807  uint16x4x2_t __s1 = __p1; \
25808  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
25809})
25810#else
25811#define vst2_u16(__p0, __p1) __extension__ ({ \
25812  uint16x4x2_t __s1 = __p1; \
25813  uint16x4x2_t __rev1; \
25814  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25815  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25816  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
25817})
25818#endif
25819
25820#ifdef __LITTLE_ENDIAN__
25821#define vst2_s8(__p0, __p1) __extension__ ({ \
25822  int8x8x2_t __s1 = __p1; \
25823  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
25824})
25825#else
25826#define vst2_s8(__p0, __p1) __extension__ ({ \
25827  int8x8x2_t __s1 = __p1; \
25828  int8x8x2_t __rev1; \
25829  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25830  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25831  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
25832})
25833#endif
25834
25835#ifdef __LITTLE_ENDIAN__
25836#define vst2_f32(__p0, __p1) __extension__ ({ \
25837  float32x2x2_t __s1 = __p1; \
25838  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 9); \
25839})
25840#else
25841#define vst2_f32(__p0, __p1) __extension__ ({ \
25842  float32x2x2_t __s1 = __p1; \
25843  float32x2x2_t __rev1; \
25844  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25845  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25846  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
25847})
25848#endif
25849
25850#ifdef __LITTLE_ENDIAN__
25851#define vst2_f16(__p0, __p1) __extension__ ({ \
25852  float16x4x2_t __s1 = __p1; \
25853  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 8); \
25854})
25855#else
25856#define vst2_f16(__p0, __p1) __extension__ ({ \
25857  float16x4x2_t __s1 = __p1; \
25858  float16x4x2_t __rev1; \
25859  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25860  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25861  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
25862})
25863#endif
25864
25865#ifdef __LITTLE_ENDIAN__
25866#define vst2_s32(__p0, __p1) __extension__ ({ \
25867  int32x2x2_t __s1 = __p1; \
25868  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 2); \
25869})
25870#else
25871#define vst2_s32(__p0, __p1) __extension__ ({ \
25872  int32x2x2_t __s1 = __p1; \
25873  int32x2x2_t __rev1; \
25874  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25875  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25876  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
25877})
25878#endif
25879
25880#ifdef __LITTLE_ENDIAN__
25881#define vst2_s64(__p0, __p1) __extension__ ({ \
25882  int64x1x2_t __s1 = __p1; \
25883  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25884})
25885#else
25886#define vst2_s64(__p0, __p1) __extension__ ({ \
25887  int64x1x2_t __s1 = __p1; \
25888  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25889})
25890#endif
25891
25892#ifdef __LITTLE_ENDIAN__
25893#define vst2_s16(__p0, __p1) __extension__ ({ \
25894  int16x4x2_t __s1 = __p1; \
25895  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 1); \
25896})
25897#else
25898#define vst2_s16(__p0, __p1) __extension__ ({ \
25899  int16x4x2_t __s1 = __p1; \
25900  int16x4x2_t __rev1; \
25901  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25902  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25903  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
25904})
25905#endif
25906
25907#ifdef __LITTLE_ENDIAN__
25908#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25909  poly8x8x2_t __s1 = __p1; \
25910  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
25911})
25912#else
25913#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25914  poly8x8x2_t __s1 = __p1; \
25915  poly8x8x2_t __rev1; \
25916  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25917  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25918  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
25919})
25920#endif
25921
25922#ifdef __LITTLE_ENDIAN__
25923#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25924  poly16x4x2_t __s1 = __p1; \
25925  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
25926})
25927#else
25928#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25929  poly16x4x2_t __s1 = __p1; \
25930  poly16x4x2_t __rev1; \
25931  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25932  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25933  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
25934})
25935#endif
25936
25937#ifdef __LITTLE_ENDIAN__
25938#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25939  poly16x8x2_t __s1 = __p1; \
25940  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
25941})
25942#else
25943#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25944  poly16x8x2_t __s1 = __p1; \
25945  poly16x8x2_t __rev1; \
25946  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25947  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25948  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
25949})
25950#endif
25951
25952#ifdef __LITTLE_ENDIAN__
25953#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25954  uint32x4x2_t __s1 = __p1; \
25955  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
25956})
25957#else
25958#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25959  uint32x4x2_t __s1 = __p1; \
25960  uint32x4x2_t __rev1; \
25961  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25962  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25963  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
25964})
25965#endif
25966
25967#ifdef __LITTLE_ENDIAN__
25968#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25969  uint16x8x2_t __s1 = __p1; \
25970  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
25971})
25972#else
25973#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25974  uint16x8x2_t __s1 = __p1; \
25975  uint16x8x2_t __rev1; \
25976  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25977  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25978  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
25979})
25980#endif
25981
25982#ifdef __LITTLE_ENDIAN__
25983#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25984  float32x4x2_t __s1 = __p1; \
25985  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 41); \
25986})
25987#else
25988#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25989  float32x4x2_t __s1 = __p1; \
25990  float32x4x2_t __rev1; \
25991  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25992  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25993  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
25994})
25995#endif
25996
25997#ifdef __LITTLE_ENDIAN__
25998#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25999  float16x8x2_t __s1 = __p1; \
26000  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 40); \
26001})
26002#else
26003#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26004  float16x8x2_t __s1 = __p1; \
26005  float16x8x2_t __rev1; \
26006  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26007  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26008  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
26009})
26010#endif
26011
26012#ifdef __LITTLE_ENDIAN__
26013#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26014  int32x4x2_t __s1 = __p1; \
26015  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 34); \
26016})
26017#else
26018#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26019  int32x4x2_t __s1 = __p1; \
26020  int32x4x2_t __rev1; \
26021  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26022  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26023  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
26024})
26025#endif
26026
26027#ifdef __LITTLE_ENDIAN__
26028#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26029  int16x8x2_t __s1 = __p1; \
26030  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 33); \
26031})
26032#else
26033#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26034  int16x8x2_t __s1 = __p1; \
26035  int16x8x2_t __rev1; \
26036  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26037  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26038  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
26039})
26040#endif
26041
26042#ifdef __LITTLE_ENDIAN__
26043#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26044  uint8x8x2_t __s1 = __p1; \
26045  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
26046})
26047#else
26048#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26049  uint8x8x2_t __s1 = __p1; \
26050  uint8x8x2_t __rev1; \
26051  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26052  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26053  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
26054})
26055#endif
26056
26057#ifdef __LITTLE_ENDIAN__
26058#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26059  uint32x2x2_t __s1 = __p1; \
26060  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
26061})
26062#else
26063#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26064  uint32x2x2_t __s1 = __p1; \
26065  uint32x2x2_t __rev1; \
26066  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26067  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26068  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
26069})
26070#endif
26071
26072#ifdef __LITTLE_ENDIAN__
26073#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26074  uint16x4x2_t __s1 = __p1; \
26075  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
26076})
26077#else
26078#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26079  uint16x4x2_t __s1 = __p1; \
26080  uint16x4x2_t __rev1; \
26081  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26082  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26083  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
26084})
26085#endif
26086
26087#ifdef __LITTLE_ENDIAN__
26088#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26089  int8x8x2_t __s1 = __p1; \
26090  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
26091})
26092#else
26093#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26094  int8x8x2_t __s1 = __p1; \
26095  int8x8x2_t __rev1; \
26096  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26097  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26098  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
26099})
26100#endif
26101
26102#ifdef __LITTLE_ENDIAN__
26103#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26104  float32x2x2_t __s1 = __p1; \
26105  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 9); \
26106})
26107#else
26108#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26109  float32x2x2_t __s1 = __p1; \
26110  float32x2x2_t __rev1; \
26111  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26112  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26113  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
26114})
26115#endif
26116
26117#ifdef __LITTLE_ENDIAN__
26118#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26119  float16x4x2_t __s1 = __p1; \
26120  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 8); \
26121})
26122#else
26123#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26124  float16x4x2_t __s1 = __p1; \
26125  float16x4x2_t __rev1; \
26126  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26127  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26128  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
26129})
26130#endif
26131
26132#ifdef __LITTLE_ENDIAN__
26133#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26134  int32x2x2_t __s1 = __p1; \
26135  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 2); \
26136})
26137#else
26138#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26139  int32x2x2_t __s1 = __p1; \
26140  int32x2x2_t __rev1; \
26141  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26142  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26143  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
26144})
26145#endif
26146
26147#ifdef __LITTLE_ENDIAN__
26148#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26149  int16x4x2_t __s1 = __p1; \
26150  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 1); \
26151})
26152#else
26153#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26154  int16x4x2_t __s1 = __p1; \
26155  int16x4x2_t __rev1; \
26156  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26157  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26158  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
26159})
26160#endif
26161
26162#ifdef __LITTLE_ENDIAN__
26163#define vst3_p8(__p0, __p1) __extension__ ({ \
26164  poly8x8x3_t __s1 = __p1; \
26165  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
26166})
26167#else
26168#define vst3_p8(__p0, __p1) __extension__ ({ \
26169  poly8x8x3_t __s1 = __p1; \
26170  poly8x8x3_t __rev1; \
26171  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26172  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26173  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26174  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
26175})
26176#endif
26177
26178#ifdef __LITTLE_ENDIAN__
26179#define vst3_p16(__p0, __p1) __extension__ ({ \
26180  poly16x4x3_t __s1 = __p1; \
26181  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
26182})
26183#else
26184#define vst3_p16(__p0, __p1) __extension__ ({ \
26185  poly16x4x3_t __s1 = __p1; \
26186  poly16x4x3_t __rev1; \
26187  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26188  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26189  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26190  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
26191})
26192#endif
26193
26194#ifdef __LITTLE_ENDIAN__
26195#define vst3q_p8(__p0, __p1) __extension__ ({ \
26196  poly8x16x3_t __s1 = __p1; \
26197  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
26198})
26199#else
26200#define vst3q_p8(__p0, __p1) __extension__ ({ \
26201  poly8x16x3_t __s1 = __p1; \
26202  poly8x16x3_t __rev1; \
26203  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26204  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26205  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26206  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
26207})
26208#endif
26209
26210#ifdef __LITTLE_ENDIAN__
26211#define vst3q_p16(__p0, __p1) __extension__ ({ \
26212  poly16x8x3_t __s1 = __p1; \
26213  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
26214})
26215#else
26216#define vst3q_p16(__p0, __p1) __extension__ ({ \
26217  poly16x8x3_t __s1 = __p1; \
26218  poly16x8x3_t __rev1; \
26219  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26220  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26221  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26222  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
26223})
26224#endif
26225
26226#ifdef __LITTLE_ENDIAN__
26227#define vst3q_u8(__p0, __p1) __extension__ ({ \
26228  uint8x16x3_t __s1 = __p1; \
26229  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
26230})
26231#else
26232#define vst3q_u8(__p0, __p1) __extension__ ({ \
26233  uint8x16x3_t __s1 = __p1; \
26234  uint8x16x3_t __rev1; \
26235  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26236  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26237  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26238  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
26239})
26240#endif
26241
26242#ifdef __LITTLE_ENDIAN__
26243#define vst3q_u32(__p0, __p1) __extension__ ({ \
26244  uint32x4x3_t __s1 = __p1; \
26245  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
26246})
26247#else
26248#define vst3q_u32(__p0, __p1) __extension__ ({ \
26249  uint32x4x3_t __s1 = __p1; \
26250  uint32x4x3_t __rev1; \
26251  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26252  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26253  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26254  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
26255})
26256#endif
26257
26258#ifdef __LITTLE_ENDIAN__
26259#define vst3q_u16(__p0, __p1) __extension__ ({ \
26260  uint16x8x3_t __s1 = __p1; \
26261  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
26262})
26263#else
26264#define vst3q_u16(__p0, __p1) __extension__ ({ \
26265  uint16x8x3_t __s1 = __p1; \
26266  uint16x8x3_t __rev1; \
26267  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26268  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26269  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26270  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
26271})
26272#endif
26273
26274#ifdef __LITTLE_ENDIAN__
26275#define vst3q_s8(__p0, __p1) __extension__ ({ \
26276  int8x16x3_t __s1 = __p1; \
26277  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
26278})
26279#else
26280#define vst3q_s8(__p0, __p1) __extension__ ({ \
26281  int8x16x3_t __s1 = __p1; \
26282  int8x16x3_t __rev1; \
26283  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26284  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26285  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26286  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
26287})
26288#endif
26289
26290#ifdef __LITTLE_ENDIAN__
26291#define vst3q_f32(__p0, __p1) __extension__ ({ \
26292  float32x4x3_t __s1 = __p1; \
26293  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
26294})
26295#else
26296#define vst3q_f32(__p0, __p1) __extension__ ({ \
26297  float32x4x3_t __s1 = __p1; \
26298  float32x4x3_t __rev1; \
26299  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26300  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26301  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26302  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
26303})
26304#endif
26305
26306#ifdef __LITTLE_ENDIAN__
26307#define vst3q_f16(__p0, __p1) __extension__ ({ \
26308  float16x8x3_t __s1 = __p1; \
26309  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
26310})
26311#else
26312#define vst3q_f16(__p0, __p1) __extension__ ({ \
26313  float16x8x3_t __s1 = __p1; \
26314  float16x8x3_t __rev1; \
26315  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26316  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26317  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26318  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
26319})
26320#endif
26321
26322#ifdef __LITTLE_ENDIAN__
26323#define vst3q_s32(__p0, __p1) __extension__ ({ \
26324  int32x4x3_t __s1 = __p1; \
26325  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
26326})
26327#else
26328#define vst3q_s32(__p0, __p1) __extension__ ({ \
26329  int32x4x3_t __s1 = __p1; \
26330  int32x4x3_t __rev1; \
26331  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26332  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26333  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26334  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
26335})
26336#endif
26337
26338#ifdef __LITTLE_ENDIAN__
26339#define vst3q_s16(__p0, __p1) __extension__ ({ \
26340  int16x8x3_t __s1 = __p1; \
26341  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
26342})
26343#else
26344#define vst3q_s16(__p0, __p1) __extension__ ({ \
26345  int16x8x3_t __s1 = __p1; \
26346  int16x8x3_t __rev1; \
26347  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26348  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26349  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26350  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
26351})
26352#endif
26353
26354#ifdef __LITTLE_ENDIAN__
26355#define vst3_u8(__p0, __p1) __extension__ ({ \
26356  uint8x8x3_t __s1 = __p1; \
26357  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
26358})
26359#else
26360#define vst3_u8(__p0, __p1) __extension__ ({ \
26361  uint8x8x3_t __s1 = __p1; \
26362  uint8x8x3_t __rev1; \
26363  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26364  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26365  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26366  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
26367})
26368#endif
26369
26370#ifdef __LITTLE_ENDIAN__
26371#define vst3_u32(__p0, __p1) __extension__ ({ \
26372  uint32x2x3_t __s1 = __p1; \
26373  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
26374})
26375#else
26376#define vst3_u32(__p0, __p1) __extension__ ({ \
26377  uint32x2x3_t __s1 = __p1; \
26378  uint32x2x3_t __rev1; \
26379  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26380  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26381  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26382  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
26383})
26384#endif
26385
26386#ifdef __LITTLE_ENDIAN__
26387#define vst3_u64(__p0, __p1) __extension__ ({ \
26388  uint64x1x3_t __s1 = __p1; \
26389  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26390})
26391#else
26392#define vst3_u64(__p0, __p1) __extension__ ({ \
26393  uint64x1x3_t __s1 = __p1; \
26394  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26395})
26396#endif
26397
26398#ifdef __LITTLE_ENDIAN__
26399#define vst3_u16(__p0, __p1) __extension__ ({ \
26400  uint16x4x3_t __s1 = __p1; \
26401  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
26402})
26403#else
26404#define vst3_u16(__p0, __p1) __extension__ ({ \
26405  uint16x4x3_t __s1 = __p1; \
26406  uint16x4x3_t __rev1; \
26407  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26408  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26409  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26410  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
26411})
26412#endif
26413
26414#ifdef __LITTLE_ENDIAN__
26415#define vst3_s8(__p0, __p1) __extension__ ({ \
26416  int8x8x3_t __s1 = __p1; \
26417  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
26418})
26419#else
26420#define vst3_s8(__p0, __p1) __extension__ ({ \
26421  int8x8x3_t __s1 = __p1; \
26422  int8x8x3_t __rev1; \
26423  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26424  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26425  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26426  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
26427})
26428#endif
26429
26430#ifdef __LITTLE_ENDIAN__
26431#define vst3_f32(__p0, __p1) __extension__ ({ \
26432  float32x2x3_t __s1 = __p1; \
26433  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
26434})
26435#else
26436#define vst3_f32(__p0, __p1) __extension__ ({ \
26437  float32x2x3_t __s1 = __p1; \
26438  float32x2x3_t __rev1; \
26439  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26440  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26441  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26442  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
26443})
26444#endif
26445
26446#ifdef __LITTLE_ENDIAN__
26447#define vst3_f16(__p0, __p1) __extension__ ({ \
26448  float16x4x3_t __s1 = __p1; \
26449  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
26450})
26451#else
26452#define vst3_f16(__p0, __p1) __extension__ ({ \
26453  float16x4x3_t __s1 = __p1; \
26454  float16x4x3_t __rev1; \
26455  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26456  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26457  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26458  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
26459})
26460#endif
26461
26462#ifdef __LITTLE_ENDIAN__
26463#define vst3_s32(__p0, __p1) __extension__ ({ \
26464  int32x2x3_t __s1 = __p1; \
26465  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
26466})
26467#else
26468#define vst3_s32(__p0, __p1) __extension__ ({ \
26469  int32x2x3_t __s1 = __p1; \
26470  int32x2x3_t __rev1; \
26471  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26472  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26473  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26474  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
26475})
26476#endif
26477
26478#ifdef __LITTLE_ENDIAN__
26479#define vst3_s64(__p0, __p1) __extension__ ({ \
26480  int64x1x3_t __s1 = __p1; \
26481  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26482})
26483#else
26484#define vst3_s64(__p0, __p1) __extension__ ({ \
26485  int64x1x3_t __s1 = __p1; \
26486  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26487})
26488#endif
26489
26490#ifdef __LITTLE_ENDIAN__
26491#define vst3_s16(__p0, __p1) __extension__ ({ \
26492  int16x4x3_t __s1 = __p1; \
26493  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
26494})
26495#else
26496#define vst3_s16(__p0, __p1) __extension__ ({ \
26497  int16x4x3_t __s1 = __p1; \
26498  int16x4x3_t __rev1; \
26499  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26500  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26501  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26502  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
26503})
26504#endif
26505
26506#ifdef __LITTLE_ENDIAN__
26507#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26508  poly8x8x3_t __s1 = __p1; \
26509  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
26510})
26511#else
26512#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26513  poly8x8x3_t __s1 = __p1; \
26514  poly8x8x3_t __rev1; \
26515  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26516  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26517  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26518  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
26519})
26520#endif
26521
26522#ifdef __LITTLE_ENDIAN__
26523#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26524  poly16x4x3_t __s1 = __p1; \
26525  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
26526})
26527#else
26528#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26529  poly16x4x3_t __s1 = __p1; \
26530  poly16x4x3_t __rev1; \
26531  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26532  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26533  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26534  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
26535})
26536#endif
26537
26538#ifdef __LITTLE_ENDIAN__
26539#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26540  poly16x8x3_t __s1 = __p1; \
26541  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
26542})
26543#else
26544#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26545  poly16x8x3_t __s1 = __p1; \
26546  poly16x8x3_t __rev1; \
26547  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26548  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26549  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26550  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
26551})
26552#endif
26553
26554#ifdef __LITTLE_ENDIAN__
26555#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26556  uint32x4x3_t __s1 = __p1; \
26557  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
26558})
26559#else
26560#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26561  uint32x4x3_t __s1 = __p1; \
26562  uint32x4x3_t __rev1; \
26563  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26564  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26565  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26566  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
26567})
26568#endif
26569
26570#ifdef __LITTLE_ENDIAN__
26571#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26572  uint16x8x3_t __s1 = __p1; \
26573  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
26574})
26575#else
26576#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26577  uint16x8x3_t __s1 = __p1; \
26578  uint16x8x3_t __rev1; \
26579  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26580  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26581  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26582  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
26583})
26584#endif
26585
26586#ifdef __LITTLE_ENDIAN__
26587#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26588  float32x4x3_t __s1 = __p1; \
26589  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
26590})
26591#else
26592#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26593  float32x4x3_t __s1 = __p1; \
26594  float32x4x3_t __rev1; \
26595  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26596  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26597  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26598  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
26599})
26600#endif
26601
26602#ifdef __LITTLE_ENDIAN__
26603#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26604  float16x8x3_t __s1 = __p1; \
26605  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
26606})
26607#else
26608#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26609  float16x8x3_t __s1 = __p1; \
26610  float16x8x3_t __rev1; \
26611  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26612  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26613  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26614  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
26615})
26616#endif
26617
26618#ifdef __LITTLE_ENDIAN__
26619#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26620  int32x4x3_t __s1 = __p1; \
26621  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
26622})
26623#else
26624#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26625  int32x4x3_t __s1 = __p1; \
26626  int32x4x3_t __rev1; \
26627  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26628  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26629  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26630  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
26631})
26632#endif
26633
26634#ifdef __LITTLE_ENDIAN__
26635#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26636  int16x8x3_t __s1 = __p1; \
26637  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
26638})
26639#else
26640#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26641  int16x8x3_t __s1 = __p1; \
26642  int16x8x3_t __rev1; \
26643  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26644  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26645  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26646  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
26647})
26648#endif
26649
26650#ifdef __LITTLE_ENDIAN__
26651#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26652  uint8x8x3_t __s1 = __p1; \
26653  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
26654})
26655#else
26656#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26657  uint8x8x3_t __s1 = __p1; \
26658  uint8x8x3_t __rev1; \
26659  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26660  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26661  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26662  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
26663})
26664#endif
26665
26666#ifdef __LITTLE_ENDIAN__
26667#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26668  uint32x2x3_t __s1 = __p1; \
26669  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
26670})
26671#else
26672#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26673  uint32x2x3_t __s1 = __p1; \
26674  uint32x2x3_t __rev1; \
26675  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26676  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26677  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26678  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
26679})
26680#endif
26681
26682#ifdef __LITTLE_ENDIAN__
26683#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26684  uint16x4x3_t __s1 = __p1; \
26685  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
26686})
26687#else
26688#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26689  uint16x4x3_t __s1 = __p1; \
26690  uint16x4x3_t __rev1; \
26691  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26692  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26693  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26694  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
26695})
26696#endif
26697
26698#ifdef __LITTLE_ENDIAN__
26699#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26700  int8x8x3_t __s1 = __p1; \
26701  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
26702})
26703#else
26704#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26705  int8x8x3_t __s1 = __p1; \
26706  int8x8x3_t __rev1; \
26707  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26708  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26709  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26710  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
26711})
26712#endif
26713
26714#ifdef __LITTLE_ENDIAN__
26715#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26716  float32x2x3_t __s1 = __p1; \
26717  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
26718})
26719#else
26720#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26721  float32x2x3_t __s1 = __p1; \
26722  float32x2x3_t __rev1; \
26723  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26724  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26725  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26726  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
26727})
26728#endif
26729
26730#ifdef __LITTLE_ENDIAN__
26731#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26732  float16x4x3_t __s1 = __p1; \
26733  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
26734})
26735#else
26736#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26737  float16x4x3_t __s1 = __p1; \
26738  float16x4x3_t __rev1; \
26739  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26740  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26741  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26742  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
26743})
26744#endif
26745
26746#ifdef __LITTLE_ENDIAN__
26747#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26748  int32x2x3_t __s1 = __p1; \
26749  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
26750})
26751#else
26752#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26753  int32x2x3_t __s1 = __p1; \
26754  int32x2x3_t __rev1; \
26755  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26756  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26757  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26758  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
26759})
26760#endif
26761
26762#ifdef __LITTLE_ENDIAN__
26763#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26764  int16x4x3_t __s1 = __p1; \
26765  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
26766})
26767#else
26768#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26769  int16x4x3_t __s1 = __p1; \
26770  int16x4x3_t __rev1; \
26771  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26772  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26773  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26774  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
26775})
26776#endif
26777
26778#ifdef __LITTLE_ENDIAN__
26779#define vst4_p8(__p0, __p1) __extension__ ({ \
26780  poly8x8x4_t __s1 = __p1; \
26781  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
26782})
26783#else
26784#define vst4_p8(__p0, __p1) __extension__ ({ \
26785  poly8x8x4_t __s1 = __p1; \
26786  poly8x8x4_t __rev1; \
26787  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26788  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26789  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26790  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26791  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
26792})
26793#endif
26794
26795#ifdef __LITTLE_ENDIAN__
26796#define vst4_p16(__p0, __p1) __extension__ ({ \
26797  poly16x4x4_t __s1 = __p1; \
26798  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
26799})
26800#else
26801#define vst4_p16(__p0, __p1) __extension__ ({ \
26802  poly16x4x4_t __s1 = __p1; \
26803  poly16x4x4_t __rev1; \
26804  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26805  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26806  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26807  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26808  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
26809})
26810#endif
26811
26812#ifdef __LITTLE_ENDIAN__
26813#define vst4q_p8(__p0, __p1) __extension__ ({ \
26814  poly8x16x4_t __s1 = __p1; \
26815  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
26816})
26817#else
26818#define vst4q_p8(__p0, __p1) __extension__ ({ \
26819  poly8x16x4_t __s1 = __p1; \
26820  poly8x16x4_t __rev1; \
26821  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26822  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26823  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26824  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26825  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
26826})
26827#endif
26828
26829#ifdef __LITTLE_ENDIAN__
26830#define vst4q_p16(__p0, __p1) __extension__ ({ \
26831  poly16x8x4_t __s1 = __p1; \
26832  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
26833})
26834#else
26835#define vst4q_p16(__p0, __p1) __extension__ ({ \
26836  poly16x8x4_t __s1 = __p1; \
26837  poly16x8x4_t __rev1; \
26838  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26839  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26840  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26841  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26842  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
26843})
26844#endif
26845
26846#ifdef __LITTLE_ENDIAN__
26847#define vst4q_u8(__p0, __p1) __extension__ ({ \
26848  uint8x16x4_t __s1 = __p1; \
26849  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
26850})
26851#else
26852#define vst4q_u8(__p0, __p1) __extension__ ({ \
26853  uint8x16x4_t __s1 = __p1; \
26854  uint8x16x4_t __rev1; \
26855  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26856  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26857  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26858  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26859  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
26860})
26861#endif
26862
26863#ifdef __LITTLE_ENDIAN__
26864#define vst4q_u32(__p0, __p1) __extension__ ({ \
26865  uint32x4x4_t __s1 = __p1; \
26866  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
26867})
26868#else
26869#define vst4q_u32(__p0, __p1) __extension__ ({ \
26870  uint32x4x4_t __s1 = __p1; \
26871  uint32x4x4_t __rev1; \
26872  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26873  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26874  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26875  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26876  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
26877})
26878#endif
26879
26880#ifdef __LITTLE_ENDIAN__
26881#define vst4q_u16(__p0, __p1) __extension__ ({ \
26882  uint16x8x4_t __s1 = __p1; \
26883  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
26884})
26885#else
26886#define vst4q_u16(__p0, __p1) __extension__ ({ \
26887  uint16x8x4_t __s1 = __p1; \
26888  uint16x8x4_t __rev1; \
26889  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26890  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26891  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26892  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26893  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
26894})
26895#endif
26896
26897#ifdef __LITTLE_ENDIAN__
26898#define vst4q_s8(__p0, __p1) __extension__ ({ \
26899  int8x16x4_t __s1 = __p1; \
26900  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
26901})
26902#else
26903#define vst4q_s8(__p0, __p1) __extension__ ({ \
26904  int8x16x4_t __s1 = __p1; \
26905  int8x16x4_t __rev1; \
26906  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26907  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26908  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26909  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26910  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
26911})
26912#endif
26913
26914#ifdef __LITTLE_ENDIAN__
26915#define vst4q_f32(__p0, __p1) __extension__ ({ \
26916  float32x4x4_t __s1 = __p1; \
26917  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
26918})
26919#else
26920#define vst4q_f32(__p0, __p1) __extension__ ({ \
26921  float32x4x4_t __s1 = __p1; \
26922  float32x4x4_t __rev1; \
26923  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26924  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26925  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26926  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26927  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
26928})
26929#endif
26930
26931#ifdef __LITTLE_ENDIAN__
26932#define vst4q_f16(__p0, __p1) __extension__ ({ \
26933  float16x8x4_t __s1 = __p1; \
26934  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
26935})
26936#else
26937#define vst4q_f16(__p0, __p1) __extension__ ({ \
26938  float16x8x4_t __s1 = __p1; \
26939  float16x8x4_t __rev1; \
26940  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26941  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26942  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26943  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26944  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
26945})
26946#endif
26947
26948#ifdef __LITTLE_ENDIAN__
26949#define vst4q_s32(__p0, __p1) __extension__ ({ \
26950  int32x4x4_t __s1 = __p1; \
26951  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
26952})
26953#else
26954#define vst4q_s32(__p0, __p1) __extension__ ({ \
26955  int32x4x4_t __s1 = __p1; \
26956  int32x4x4_t __rev1; \
26957  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26958  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26959  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26960  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26961  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
26962})
26963#endif
26964
26965#ifdef __LITTLE_ENDIAN__
26966#define vst4q_s16(__p0, __p1) __extension__ ({ \
26967  int16x8x4_t __s1 = __p1; \
26968  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
26969})
26970#else
26971#define vst4q_s16(__p0, __p1) __extension__ ({ \
26972  int16x8x4_t __s1 = __p1; \
26973  int16x8x4_t __rev1; \
26974  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26975  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26976  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26977  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26978  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
26979})
26980#endif
26981
26982#ifdef __LITTLE_ENDIAN__
26983#define vst4_u8(__p0, __p1) __extension__ ({ \
26984  uint8x8x4_t __s1 = __p1; \
26985  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
26986})
26987#else
26988#define vst4_u8(__p0, __p1) __extension__ ({ \
26989  uint8x8x4_t __s1 = __p1; \
26990  uint8x8x4_t __rev1; \
26991  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26992  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26993  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26994  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26995  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
26996})
26997#endif
26998
26999#ifdef __LITTLE_ENDIAN__
27000#define vst4_u32(__p0, __p1) __extension__ ({ \
27001  uint32x2x4_t __s1 = __p1; \
27002  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27003})
27004#else
27005#define vst4_u32(__p0, __p1) __extension__ ({ \
27006  uint32x2x4_t __s1 = __p1; \
27007  uint32x2x4_t __rev1; \
27008  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27009  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27010  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27011  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27012  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27013})
27014#endif
27015
27016#ifdef __LITTLE_ENDIAN__
27017#define vst4_u64(__p0, __p1) __extension__ ({ \
27018  uint64x1x4_t __s1 = __p1; \
27019  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27020})
27021#else
27022#define vst4_u64(__p0, __p1) __extension__ ({ \
27023  uint64x1x4_t __s1 = __p1; \
27024  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27025})
27026#endif
27027
27028#ifdef __LITTLE_ENDIAN__
27029#define vst4_u16(__p0, __p1) __extension__ ({ \
27030  uint16x4x4_t __s1 = __p1; \
27031  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
27032})
27033#else
27034#define vst4_u16(__p0, __p1) __extension__ ({ \
27035  uint16x4x4_t __s1 = __p1; \
27036  uint16x4x4_t __rev1; \
27037  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27038  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27039  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27040  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27041  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
27042})
27043#endif
27044
27045#ifdef __LITTLE_ENDIAN__
27046#define vst4_s8(__p0, __p1) __extension__ ({ \
27047  int8x8x4_t __s1 = __p1; \
27048  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
27049})
27050#else
27051#define vst4_s8(__p0, __p1) __extension__ ({ \
27052  int8x8x4_t __s1 = __p1; \
27053  int8x8x4_t __rev1; \
27054  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27055  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27056  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27057  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27058  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
27059})
27060#endif
27061
27062#ifdef __LITTLE_ENDIAN__
27063#define vst4_f32(__p0, __p1) __extension__ ({ \
27064  float32x2x4_t __s1 = __p1; \
27065  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
27066})
27067#else
27068#define vst4_f32(__p0, __p1) __extension__ ({ \
27069  float32x2x4_t __s1 = __p1; \
27070  float32x2x4_t __rev1; \
27071  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27072  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27073  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27074  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27075  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
27076})
27077#endif
27078
27079#ifdef __LITTLE_ENDIAN__
27080#define vst4_f16(__p0, __p1) __extension__ ({ \
27081  float16x4x4_t __s1 = __p1; \
27082  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
27083})
27084#else
27085#define vst4_f16(__p0, __p1) __extension__ ({ \
27086  float16x4x4_t __s1 = __p1; \
27087  float16x4x4_t __rev1; \
27088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27090  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27091  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27092  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
27093})
27094#endif
27095
27096#ifdef __LITTLE_ENDIAN__
27097#define vst4_s32(__p0, __p1) __extension__ ({ \
27098  int32x2x4_t __s1 = __p1; \
27099  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
27100})
27101#else
27102#define vst4_s32(__p0, __p1) __extension__ ({ \
27103  int32x2x4_t __s1 = __p1; \
27104  int32x2x4_t __rev1; \
27105  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27106  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27107  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27108  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27109  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
27110})
27111#endif
27112
27113#ifdef __LITTLE_ENDIAN__
27114#define vst4_s64(__p0, __p1) __extension__ ({ \
27115  int64x1x4_t __s1 = __p1; \
27116  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27117})
27118#else
27119#define vst4_s64(__p0, __p1) __extension__ ({ \
27120  int64x1x4_t __s1 = __p1; \
27121  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27122})
27123#endif
27124
27125#ifdef __LITTLE_ENDIAN__
27126#define vst4_s16(__p0, __p1) __extension__ ({ \
27127  int16x4x4_t __s1 = __p1; \
27128  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
27129})
27130#else
27131#define vst4_s16(__p0, __p1) __extension__ ({ \
27132  int16x4x4_t __s1 = __p1; \
27133  int16x4x4_t __rev1; \
27134  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27135  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27136  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27137  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27138  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
27139})
27140#endif
27141
27142#ifdef __LITTLE_ENDIAN__
27143#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27144  poly8x8x4_t __s1 = __p1; \
27145  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
27146})
27147#else
27148#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27149  poly8x8x4_t __s1 = __p1; \
27150  poly8x8x4_t __rev1; \
27151  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27152  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27153  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27154  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27155  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
27156})
27157#endif
27158
27159#ifdef __LITTLE_ENDIAN__
27160#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27161  poly16x4x4_t __s1 = __p1; \
27162  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
27163})
27164#else
27165#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27166  poly16x4x4_t __s1 = __p1; \
27167  poly16x4x4_t __rev1; \
27168  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27169  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27170  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27171  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27172  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
27173})
27174#endif
27175
27176#ifdef __LITTLE_ENDIAN__
27177#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27178  poly16x8x4_t __s1 = __p1; \
27179  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
27180})
27181#else
27182#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27183  poly16x8x4_t __s1 = __p1; \
27184  poly16x8x4_t __rev1; \
27185  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27186  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27187  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27188  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27189  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
27190})
27191#endif
27192
27193#ifdef __LITTLE_ENDIAN__
27194#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27195  uint32x4x4_t __s1 = __p1; \
27196  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
27197})
27198#else
27199#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27200  uint32x4x4_t __s1 = __p1; \
27201  uint32x4x4_t __rev1; \
27202  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27203  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27204  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27205  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27206  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
27207})
27208#endif
27209
27210#ifdef __LITTLE_ENDIAN__
27211#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27212  uint16x8x4_t __s1 = __p1; \
27213  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
27214})
27215#else
27216#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27217  uint16x8x4_t __s1 = __p1; \
27218  uint16x8x4_t __rev1; \
27219  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27220  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27221  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27222  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27223  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
27224})
27225#endif
27226
27227#ifdef __LITTLE_ENDIAN__
27228#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27229  float32x4x4_t __s1 = __p1; \
27230  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
27231})
27232#else
27233#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27234  float32x4x4_t __s1 = __p1; \
27235  float32x4x4_t __rev1; \
27236  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27237  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27238  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27239  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27240  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
27241})
27242#endif
27243
27244#ifdef __LITTLE_ENDIAN__
27245#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27246  float16x8x4_t __s1 = __p1; \
27247  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
27248})
27249#else
27250#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27251  float16x8x4_t __s1 = __p1; \
27252  float16x8x4_t __rev1; \
27253  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27254  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27255  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27256  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27257  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
27258})
27259#endif
27260
27261#ifdef __LITTLE_ENDIAN__
27262#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27263  int32x4x4_t __s1 = __p1; \
27264  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
27265})
27266#else
27267#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27268  int32x4x4_t __s1 = __p1; \
27269  int32x4x4_t __rev1; \
27270  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27271  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27272  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27273  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27274  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
27275})
27276#endif
27277
27278#ifdef __LITTLE_ENDIAN__
27279#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27280  int16x8x4_t __s1 = __p1; \
27281  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
27282})
27283#else
27284#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27285  int16x8x4_t __s1 = __p1; \
27286  int16x8x4_t __rev1; \
27287  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27288  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27289  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27290  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27291  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
27292})
27293#endif
27294
27295#ifdef __LITTLE_ENDIAN__
27296#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27297  uint8x8x4_t __s1 = __p1; \
27298  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
27299})
27300#else
27301#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27302  uint8x8x4_t __s1 = __p1; \
27303  uint8x8x4_t __rev1; \
27304  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27305  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27306  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27307  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27308  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
27309})
27310#endif
27311
27312#ifdef __LITTLE_ENDIAN__
27313#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27314  uint32x2x4_t __s1 = __p1; \
27315  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
27316})
27317#else
27318#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27319  uint32x2x4_t __s1 = __p1; \
27320  uint32x2x4_t __rev1; \
27321  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27322  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27323  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27324  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27325  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
27326})
27327#endif
27328
27329#ifdef __LITTLE_ENDIAN__
27330#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27331  uint16x4x4_t __s1 = __p1; \
27332  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
27333})
27334#else
27335#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27336  uint16x4x4_t __s1 = __p1; \
27337  uint16x4x4_t __rev1; \
27338  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27339  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27340  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27341  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27342  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
27343})
27344#endif
27345
27346#ifdef __LITTLE_ENDIAN__
27347#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27348  int8x8x4_t __s1 = __p1; \
27349  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
27350})
27351#else
27352#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27353  int8x8x4_t __s1 = __p1; \
27354  int8x8x4_t __rev1; \
27355  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27356  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27357  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27358  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27359  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
27360})
27361#endif
27362
27363#ifdef __LITTLE_ENDIAN__
27364#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27365  float32x2x4_t __s1 = __p1; \
27366  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
27367})
27368#else
27369#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27370  float32x2x4_t __s1 = __p1; \
27371  float32x2x4_t __rev1; \
27372  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27373  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27374  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27375  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27376  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
27377})
27378#endif
27379
27380#ifdef __LITTLE_ENDIAN__
27381#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27382  float16x4x4_t __s1 = __p1; \
27383  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
27384})
27385#else
27386#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27387  float16x4x4_t __s1 = __p1; \
27388  float16x4x4_t __rev1; \
27389  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27390  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27391  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27392  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27393  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
27394})
27395#endif
27396
27397#ifdef __LITTLE_ENDIAN__
27398#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27399  int32x2x4_t __s1 = __p1; \
27400  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
27401})
27402#else
27403#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27404  int32x2x4_t __s1 = __p1; \
27405  int32x2x4_t __rev1; \
27406  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27407  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27408  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27409  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27410  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
27411})
27412#endif
27413
27414#ifdef __LITTLE_ENDIAN__
27415#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27416  int16x4x4_t __s1 = __p1; \
27417  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
27418})
27419#else
27420#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27421  int16x4x4_t __s1 = __p1; \
27422  int16x4x4_t __rev1; \
27423  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27424  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27425  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27426  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27427  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
27428})
27429#endif
27430
27431#ifdef __LITTLE_ENDIAN__
27432__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27433  uint8x16_t __ret;
27434  __ret = __p0 - __p1;
27435  return __ret;
27436}
27437#else
27438__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27439  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27440  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27441  uint8x16_t __ret;
27442  __ret = __rev0 - __rev1;
27443  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27444  return __ret;
27445}
27446#endif
27447
27448#ifdef __LITTLE_ENDIAN__
27449__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27450  uint32x4_t __ret;
27451  __ret = __p0 - __p1;
27452  return __ret;
27453}
27454#else
27455__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27456  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27457  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27458  uint32x4_t __ret;
27459  __ret = __rev0 - __rev1;
27460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27461  return __ret;
27462}
27463#endif
27464
27465#ifdef __LITTLE_ENDIAN__
27466__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27467  uint64x2_t __ret;
27468  __ret = __p0 - __p1;
27469  return __ret;
27470}
27471#else
27472__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27473  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27474  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27475  uint64x2_t __ret;
27476  __ret = __rev0 - __rev1;
27477  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27478  return __ret;
27479}
27480#endif
27481
27482#ifdef __LITTLE_ENDIAN__
27483__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27484  uint16x8_t __ret;
27485  __ret = __p0 - __p1;
27486  return __ret;
27487}
27488#else
27489__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27490  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27491  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27492  uint16x8_t __ret;
27493  __ret = __rev0 - __rev1;
27494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27495  return __ret;
27496}
27497#endif
27498
27499#ifdef __LITTLE_ENDIAN__
27500__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27501  int8x16_t __ret;
27502  __ret = __p0 - __p1;
27503  return __ret;
27504}
27505#else
27506__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27507  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27508  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27509  int8x16_t __ret;
27510  __ret = __rev0 - __rev1;
27511  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27512  return __ret;
27513}
27514#endif
27515
27516#ifdef __LITTLE_ENDIAN__
27517__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27518  float32x4_t __ret;
27519  __ret = __p0 - __p1;
27520  return __ret;
27521}
27522#else
27523__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27524  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27525  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27526  float32x4_t __ret;
27527  __ret = __rev0 - __rev1;
27528  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27529  return __ret;
27530}
27531#endif
27532
27533#ifdef __LITTLE_ENDIAN__
27534__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27535  int32x4_t __ret;
27536  __ret = __p0 - __p1;
27537  return __ret;
27538}
27539#else
27540__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27541  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27542  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27543  int32x4_t __ret;
27544  __ret = __rev0 - __rev1;
27545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27546  return __ret;
27547}
27548#endif
27549
27550#ifdef __LITTLE_ENDIAN__
27551__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27552  int64x2_t __ret;
27553  __ret = __p0 - __p1;
27554  return __ret;
27555}
27556#else
27557__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27558  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27559  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27560  int64x2_t __ret;
27561  __ret = __rev0 - __rev1;
27562  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27563  return __ret;
27564}
27565#endif
27566
27567#ifdef __LITTLE_ENDIAN__
27568__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27569  int16x8_t __ret;
27570  __ret = __p0 - __p1;
27571  return __ret;
27572}
27573#else
27574__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27575  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27576  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27577  int16x8_t __ret;
27578  __ret = __rev0 - __rev1;
27579  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27580  return __ret;
27581}
27582#endif
27583
27584#ifdef __LITTLE_ENDIAN__
27585__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27586  uint8x8_t __ret;
27587  __ret = __p0 - __p1;
27588  return __ret;
27589}
27590#else
27591__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27592  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27593  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27594  uint8x8_t __ret;
27595  __ret = __rev0 - __rev1;
27596  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27597  return __ret;
27598}
27599#endif
27600
27601#ifdef __LITTLE_ENDIAN__
27602__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27603  uint32x2_t __ret;
27604  __ret = __p0 - __p1;
27605  return __ret;
27606}
27607#else
27608__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27609  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27610  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27611  uint32x2_t __ret;
27612  __ret = __rev0 - __rev1;
27613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27614  return __ret;
27615}
27616#endif
27617
27618#ifdef __LITTLE_ENDIAN__
27619__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27620  uint64x1_t __ret;
27621  __ret = __p0 - __p1;
27622  return __ret;
27623}
27624#else
27625__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27626  uint64x1_t __ret;
27627  __ret = __p0 - __p1;
27628  return __ret;
27629}
27630#endif
27631
27632#ifdef __LITTLE_ENDIAN__
27633__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27634  uint16x4_t __ret;
27635  __ret = __p0 - __p1;
27636  return __ret;
27637}
27638#else
27639__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27640  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27641  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27642  uint16x4_t __ret;
27643  __ret = __rev0 - __rev1;
27644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27645  return __ret;
27646}
27647#endif
27648
27649#ifdef __LITTLE_ENDIAN__
27650__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27651  int8x8_t __ret;
27652  __ret = __p0 - __p1;
27653  return __ret;
27654}
27655#else
27656__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27657  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27658  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27659  int8x8_t __ret;
27660  __ret = __rev0 - __rev1;
27661  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27662  return __ret;
27663}
27664#endif
27665
27666#ifdef __LITTLE_ENDIAN__
27667__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27668  float32x2_t __ret;
27669  __ret = __p0 - __p1;
27670  return __ret;
27671}
27672#else
27673__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27674  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27675  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27676  float32x2_t __ret;
27677  __ret = __rev0 - __rev1;
27678  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27679  return __ret;
27680}
27681#endif
27682
27683#ifdef __LITTLE_ENDIAN__
27684__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27685  int32x2_t __ret;
27686  __ret = __p0 - __p1;
27687  return __ret;
27688}
27689#else
27690__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27691  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27692  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27693  int32x2_t __ret;
27694  __ret = __rev0 - __rev1;
27695  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27696  return __ret;
27697}
27698#endif
27699
27700#ifdef __LITTLE_ENDIAN__
27701__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27702  int64x1_t __ret;
27703  __ret = __p0 - __p1;
27704  return __ret;
27705}
27706#else
27707__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27708  int64x1_t __ret;
27709  __ret = __p0 - __p1;
27710  return __ret;
27711}
27712#endif
27713
27714#ifdef __LITTLE_ENDIAN__
27715__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27716  int16x4_t __ret;
27717  __ret = __p0 - __p1;
27718  return __ret;
27719}
27720#else
27721__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27722  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27723  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27724  int16x4_t __ret;
27725  __ret = __rev0 - __rev1;
27726  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27727  return __ret;
27728}
27729#endif
27730
27731#ifdef __LITTLE_ENDIAN__
27732__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27733  uint16x4_t __ret;
27734  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27735  return __ret;
27736}
27737#else
27738__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27739  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27740  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27741  uint16x4_t __ret;
27742  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
27743  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27744  return __ret;
27745}
27746__ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27747  uint16x4_t __ret;
27748  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27749  return __ret;
27750}
27751#endif
27752
27753#ifdef __LITTLE_ENDIAN__
27754__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27755  uint32x2_t __ret;
27756  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27757  return __ret;
27758}
27759#else
27760__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27761  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27762  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27763  uint32x2_t __ret;
27764  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
27765  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27766  return __ret;
27767}
27768__ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27769  uint32x2_t __ret;
27770  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27771  return __ret;
27772}
27773#endif
27774
27775#ifdef __LITTLE_ENDIAN__
27776__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27777  uint8x8_t __ret;
27778  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27779  return __ret;
27780}
27781#else
27782__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27783  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27784  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27785  uint8x8_t __ret;
27786  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
27787  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27788  return __ret;
27789}
27790__ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27791  uint8x8_t __ret;
27792  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27793  return __ret;
27794}
27795#endif
27796
27797#ifdef __LITTLE_ENDIAN__
27798__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27799  int16x4_t __ret;
27800  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27801  return __ret;
27802}
27803#else
27804__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27805  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27806  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27807  int16x4_t __ret;
27808  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
27809  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27810  return __ret;
27811}
27812__ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27813  int16x4_t __ret;
27814  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27815  return __ret;
27816}
27817#endif
27818
27819#ifdef __LITTLE_ENDIAN__
27820__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27821  int32x2_t __ret;
27822  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27823  return __ret;
27824}
27825#else
27826__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27827  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27828  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27829  int32x2_t __ret;
27830  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
27831  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27832  return __ret;
27833}
27834__ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27835  int32x2_t __ret;
27836  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27837  return __ret;
27838}
27839#endif
27840
27841#ifdef __LITTLE_ENDIAN__
27842__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27843  int8x8_t __ret;
27844  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27845  return __ret;
27846}
27847#else
27848__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27849  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27850  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27851  int8x8_t __ret;
27852  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
27853  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27854  return __ret;
27855}
27856__ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27857  int8x8_t __ret;
27858  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27859  return __ret;
27860}
27861#endif
27862
27863#ifdef __LITTLE_ENDIAN__
27864__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27865  uint16x8_t __ret;
27866  __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
27867  return __ret;
27868}
27869#else
27870__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27871  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27872  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27873  uint16x8_t __ret;
27874  __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
27875  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27876  return __ret;
27877}
27878#endif
27879
27880#ifdef __LITTLE_ENDIAN__
27881__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27882  uint64x2_t __ret;
27883  __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
27884  return __ret;
27885}
27886#else
27887__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27888  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27889  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27890  uint64x2_t __ret;
27891  __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
27892  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27893  return __ret;
27894}
27895#endif
27896
27897#ifdef __LITTLE_ENDIAN__
27898__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27899  uint32x4_t __ret;
27900  __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
27901  return __ret;
27902}
27903#else
27904__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27905  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27906  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27907  uint32x4_t __ret;
27908  __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
27909  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27910  return __ret;
27911}
27912#endif
27913
27914#ifdef __LITTLE_ENDIAN__
27915__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
27916  int16x8_t __ret;
27917  __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
27918  return __ret;
27919}
27920#else
27921__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
27922  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27923  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27924  int16x8_t __ret;
27925  __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
27926  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27927  return __ret;
27928}
27929#endif
27930
27931#ifdef __LITTLE_ENDIAN__
27932__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
27933  int64x2_t __ret;
27934  __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
27935  return __ret;
27936}
27937#else
27938__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
27939  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27940  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27941  int64x2_t __ret;
27942  __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
27943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27944  return __ret;
27945}
27946#endif
27947
27948#ifdef __LITTLE_ENDIAN__
27949__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
27950  int32x4_t __ret;
27951  __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
27952  return __ret;
27953}
27954#else
27955__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
27956  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27957  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27958  int32x4_t __ret;
27959  __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
27960  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27961  return __ret;
27962}
27963#endif
27964
27965#ifdef __LITTLE_ENDIAN__
27966__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
27967  uint16x8_t __ret;
27968  __ret = __p0 - vmovl_u8(__p1);
27969  return __ret;
27970}
27971#else
27972__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
27973  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27974  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27975  uint16x8_t __ret;
27976  __ret = __rev0 - __noswap_vmovl_u8(__rev1);
27977  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27978  return __ret;
27979}
27980#endif
27981
27982#ifdef __LITTLE_ENDIAN__
27983__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
27984  uint64x2_t __ret;
27985  __ret = __p0 - vmovl_u32(__p1);
27986  return __ret;
27987}
27988#else
27989__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
27990  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27991  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27992  uint64x2_t __ret;
27993  __ret = __rev0 - __noswap_vmovl_u32(__rev1);
27994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27995  return __ret;
27996}
27997#endif
27998
27999#ifdef __LITTLE_ENDIAN__
28000__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28001  uint32x4_t __ret;
28002  __ret = __p0 - vmovl_u16(__p1);
28003  return __ret;
28004}
28005#else
28006__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28007  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28008  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28009  uint32x4_t __ret;
28010  __ret = __rev0 - __noswap_vmovl_u16(__rev1);
28011  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28012  return __ret;
28013}
28014#endif
28015
28016#ifdef __LITTLE_ENDIAN__
28017__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28018  int16x8_t __ret;
28019  __ret = __p0 - vmovl_s8(__p1);
28020  return __ret;
28021}
28022#else
28023__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28024  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28025  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28026  int16x8_t __ret;
28027  __ret = __rev0 - __noswap_vmovl_s8(__rev1);
28028  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28029  return __ret;
28030}
28031#endif
28032
28033#ifdef __LITTLE_ENDIAN__
28034__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28035  int64x2_t __ret;
28036  __ret = __p0 - vmovl_s32(__p1);
28037  return __ret;
28038}
28039#else
28040__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28041  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28042  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28043  int64x2_t __ret;
28044  __ret = __rev0 - __noswap_vmovl_s32(__rev1);
28045  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28046  return __ret;
28047}
28048#endif
28049
28050#ifdef __LITTLE_ENDIAN__
28051__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28052  int32x4_t __ret;
28053  __ret = __p0 - vmovl_s16(__p1);
28054  return __ret;
28055}
28056#else
28057__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28058  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28059  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28060  int32x4_t __ret;
28061  __ret = __rev0 - __noswap_vmovl_s16(__rev1);
28062  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28063  return __ret;
28064}
28065#endif
28066
28067#ifdef __LITTLE_ENDIAN__
28068__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28069  poly8x8_t __ret;
28070  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
28071  return __ret;
28072}
28073#else
28074__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28075  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28076  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28077  poly8x8_t __ret;
28078  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28079  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28080  return __ret;
28081}
28082#endif
28083
28084#ifdef __LITTLE_ENDIAN__
28085__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28086  uint8x8_t __ret;
28087  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28088  return __ret;
28089}
28090#else
28091__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28092  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28093  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28094  uint8x8_t __ret;
28095  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28097  return __ret;
28098}
28099#endif
28100
28101#ifdef __LITTLE_ENDIAN__
28102__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28103  int8x8_t __ret;
28104  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
28105  return __ret;
28106}
28107#else
28108__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28109  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28110  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28111  int8x8_t __ret;
28112  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28113  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28114  return __ret;
28115}
28116#endif
28117
28118#ifdef __LITTLE_ENDIAN__
28119__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28120  poly8x8_t __ret;
28121  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
28122  return __ret;
28123}
28124#else
28125__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28126  poly8x8x2_t __rev0;
28127  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28128  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28129  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28130  poly8x8_t __ret;
28131  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
28132  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28133  return __ret;
28134}
28135#endif
28136
28137#ifdef __LITTLE_ENDIAN__
28138__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28139  uint8x8_t __ret;
28140  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
28141  return __ret;
28142}
28143#else
28144__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28145  uint8x8x2_t __rev0;
28146  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28147  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28148  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28149  uint8x8_t __ret;
28150  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
28151  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28152  return __ret;
28153}
28154#endif
28155
28156#ifdef __LITTLE_ENDIAN__
28157__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28158  int8x8_t __ret;
28159  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
28160  return __ret;
28161}
28162#else
28163__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28164  int8x8x2_t __rev0;
28165  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28166  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28167  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28168  int8x8_t __ret;
28169  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
28170  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28171  return __ret;
28172}
28173#endif
28174
28175#ifdef __LITTLE_ENDIAN__
28176__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28177  poly8x8_t __ret;
28178  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
28179  return __ret;
28180}
28181#else
28182__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28183  poly8x8x3_t __rev0;
28184  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28185  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28186  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28187  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28188  poly8x8_t __ret;
28189  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
28190  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28191  return __ret;
28192}
28193#endif
28194
28195#ifdef __LITTLE_ENDIAN__
28196__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28197  uint8x8_t __ret;
28198  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
28199  return __ret;
28200}
28201#else
28202__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28203  uint8x8x3_t __rev0;
28204  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28205  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28206  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28207  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28208  uint8x8_t __ret;
28209  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
28210  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28211  return __ret;
28212}
28213#endif
28214
28215#ifdef __LITTLE_ENDIAN__
28216__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28217  int8x8_t __ret;
28218  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
28219  return __ret;
28220}
28221#else
28222__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28223  int8x8x3_t __rev0;
28224  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28225  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28226  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28227  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28228  int8x8_t __ret;
28229  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
28230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28231  return __ret;
28232}
28233#endif
28234
28235#ifdef __LITTLE_ENDIAN__
28236__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28237  poly8x8_t __ret;
28238  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
28239  return __ret;
28240}
28241#else
28242__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28243  poly8x8x4_t __rev0;
28244  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28245  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28246  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28247  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28248  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28249  poly8x8_t __ret;
28250  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
28251  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28252  return __ret;
28253}
28254#endif
28255
28256#ifdef __LITTLE_ENDIAN__
28257__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28258  uint8x8_t __ret;
28259  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
28260  return __ret;
28261}
28262#else
28263__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28264  uint8x8x4_t __rev0;
28265  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28266  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28267  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28268  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28269  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28270  uint8x8_t __ret;
28271  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
28272  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28273  return __ret;
28274}
28275#endif
28276
28277#ifdef __LITTLE_ENDIAN__
28278__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28279  int8x8_t __ret;
28280  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
28281  return __ret;
28282}
28283#else
28284__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28285  int8x8x4_t __rev0;
28286  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28287  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28288  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28289  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28290  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28291  int8x8_t __ret;
28292  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
28293  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28294  return __ret;
28295}
28296#endif
28297
28298#ifdef __LITTLE_ENDIAN__
28299__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28300  poly8x8_t __ret;
28301  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
28302  return __ret;
28303}
28304#else
28305__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28306  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28307  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28308  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28309  poly8x8_t __ret;
28310  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
28311  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28312  return __ret;
28313}
28314#endif
28315
28316#ifdef __LITTLE_ENDIAN__
28317__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28318  uint8x8_t __ret;
28319  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
28320  return __ret;
28321}
28322#else
28323__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28324  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28325  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28326  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28327  uint8x8_t __ret;
28328  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
28329  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28330  return __ret;
28331}
28332#endif
28333
28334#ifdef __LITTLE_ENDIAN__
28335__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28336  int8x8_t __ret;
28337  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
28338  return __ret;
28339}
28340#else
28341__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28342  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28343  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28344  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28345  int8x8_t __ret;
28346  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
28347  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28348  return __ret;
28349}
28350#endif
28351
28352#ifdef __LITTLE_ENDIAN__
28353__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28354  poly8x8_t __ret;
28355  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
28356  return __ret;
28357}
28358#else
28359__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28360  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28361  poly8x8x2_t __rev1;
28362  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28363  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28364  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28365  poly8x8_t __ret;
28366  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
28367  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28368  return __ret;
28369}
28370#endif
28371
28372#ifdef __LITTLE_ENDIAN__
28373__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28374  uint8x8_t __ret;
28375  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
28376  return __ret;
28377}
28378#else
28379__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28380  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28381  uint8x8x2_t __rev1;
28382  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28383  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28384  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28385  uint8x8_t __ret;
28386  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
28387  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28388  return __ret;
28389}
28390#endif
28391
28392#ifdef __LITTLE_ENDIAN__
28393__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28394  int8x8_t __ret;
28395  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
28396  return __ret;
28397}
28398#else
28399__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28400  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28401  int8x8x2_t __rev1;
28402  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28403  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28404  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28405  int8x8_t __ret;
28406  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
28407  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28408  return __ret;
28409}
28410#endif
28411
28412#ifdef __LITTLE_ENDIAN__
28413__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28414  poly8x8_t __ret;
28415  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
28416  return __ret;
28417}
28418#else
28419__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28420  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28421  poly8x8x3_t __rev1;
28422  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28423  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28424  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28425  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28426  poly8x8_t __ret;
28427  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
28428  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28429  return __ret;
28430}
28431#endif
28432
28433#ifdef __LITTLE_ENDIAN__
28434__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28435  uint8x8_t __ret;
28436  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
28437  return __ret;
28438}
28439#else
28440__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28441  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28442  uint8x8x3_t __rev1;
28443  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28444  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28445  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28446  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28447  uint8x8_t __ret;
28448  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
28449  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28450  return __ret;
28451}
28452#endif
28453
28454#ifdef __LITTLE_ENDIAN__
28455__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28456  int8x8_t __ret;
28457  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
28458  return __ret;
28459}
28460#else
28461__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28462  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28463  int8x8x3_t __rev1;
28464  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28465  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28466  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28467  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28468  int8x8_t __ret;
28469  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
28470  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28471  return __ret;
28472}
28473#endif
28474
28475#ifdef __LITTLE_ENDIAN__
28476__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28477  poly8x8_t __ret;
28478  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
28479  return __ret;
28480}
28481#else
28482__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28483  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28484  poly8x8x4_t __rev1;
28485  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28486  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28487  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28488  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28489  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28490  poly8x8_t __ret;
28491  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
28492  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28493  return __ret;
28494}
28495#endif
28496
28497#ifdef __LITTLE_ENDIAN__
28498__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28499  uint8x8_t __ret;
28500  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
28501  return __ret;
28502}
28503#else
28504__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28505  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28506  uint8x8x4_t __rev1;
28507  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28508  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28509  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28510  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28511  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28512  uint8x8_t __ret;
28513  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
28514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28515  return __ret;
28516}
28517#endif
28518
28519#ifdef __LITTLE_ENDIAN__
28520__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28521  int8x8_t __ret;
28522  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
28523  return __ret;
28524}
28525#else
28526__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28527  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28528  int8x8x4_t __rev1;
28529  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28530  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28531  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28532  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28533  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28534  int8x8_t __ret;
28535  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
28536  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28537  return __ret;
28538}
28539#endif
28540
28541#ifdef __LITTLE_ENDIAN__
28542__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28543  poly8x8x2_t __ret;
28544  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
28545  return __ret;
28546}
28547#else
28548__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28549  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28550  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28551  poly8x8x2_t __ret;
28552  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28553
28554  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28555  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28556  return __ret;
28557}
28558#endif
28559
28560#ifdef __LITTLE_ENDIAN__
28561__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28562  poly16x4x2_t __ret;
28563  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
28564  return __ret;
28565}
28566#else
28567__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28568  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28569  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28570  poly16x4x2_t __ret;
28571  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
28572
28573  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28574  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28575  return __ret;
28576}
28577#endif
28578
28579#ifdef __LITTLE_ENDIAN__
28580__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28581  poly8x16x2_t __ret;
28582  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
28583  return __ret;
28584}
28585#else
28586__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28587  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28588  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28589  poly8x16x2_t __ret;
28590  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
28591
28592  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28593  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28594  return __ret;
28595}
28596#endif
28597
28598#ifdef __LITTLE_ENDIAN__
28599__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28600  poly16x8x2_t __ret;
28601  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
28602  return __ret;
28603}
28604#else
28605__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28606  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28607  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28608  poly16x8x2_t __ret;
28609  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
28610
28611  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28612  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28613  return __ret;
28614}
28615#endif
28616
28617#ifdef __LITTLE_ENDIAN__
28618__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28619  uint8x16x2_t __ret;
28620  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
28621  return __ret;
28622}
28623#else
28624__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28625  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28626  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28627  uint8x16x2_t __ret;
28628  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28629
28630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28632  return __ret;
28633}
28634#endif
28635
28636#ifdef __LITTLE_ENDIAN__
28637__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28638  uint32x4x2_t __ret;
28639  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
28640  return __ret;
28641}
28642#else
28643__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28644  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28645  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28646  uint32x4x2_t __ret;
28647  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
28648
28649  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28650  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28651  return __ret;
28652}
28653#endif
28654
28655#ifdef __LITTLE_ENDIAN__
28656__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28657  uint16x8x2_t __ret;
28658  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
28659  return __ret;
28660}
28661#else
28662__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28663  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28664  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28665  uint16x8x2_t __ret;
28666  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28667
28668  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28669  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28670  return __ret;
28671}
28672#endif
28673
28674#ifdef __LITTLE_ENDIAN__
28675__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28676  int8x16x2_t __ret;
28677  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
28678  return __ret;
28679}
28680#else
28681__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28682  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28683  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28684  int8x16x2_t __ret;
28685  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
28686
28687  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28688  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28689  return __ret;
28690}
28691#endif
28692
28693#ifdef __LITTLE_ENDIAN__
28694__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28695  float32x4x2_t __ret;
28696  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
28697  return __ret;
28698}
28699#else
28700__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28701  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28702  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28703  float32x4x2_t __ret;
28704  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
28705
28706  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28707  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28708  return __ret;
28709}
28710#endif
28711
28712#ifdef __LITTLE_ENDIAN__
28713__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28714  int32x4x2_t __ret;
28715  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
28716  return __ret;
28717}
28718#else
28719__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28720  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28721  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28722  int32x4x2_t __ret;
28723  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
28724
28725  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28726  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28727  return __ret;
28728}
28729#endif
28730
28731#ifdef __LITTLE_ENDIAN__
28732__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28733  int16x8x2_t __ret;
28734  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
28735  return __ret;
28736}
28737#else
28738__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28739  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28740  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28741  int16x8x2_t __ret;
28742  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
28743
28744  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28745  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28746  return __ret;
28747}
28748#endif
28749
28750#ifdef __LITTLE_ENDIAN__
28751__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28752  uint8x8x2_t __ret;
28753  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
28754  return __ret;
28755}
28756#else
28757__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28758  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28759  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28760  uint8x8x2_t __ret;
28761  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28762
28763  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28764  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28765  return __ret;
28766}
28767#endif
28768
28769#ifdef __LITTLE_ENDIAN__
28770__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28771  uint32x2x2_t __ret;
28772  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
28773  return __ret;
28774}
28775#else
28776__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28777  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28778  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28779  uint32x2x2_t __ret;
28780  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
28781
28782  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28783  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28784  return __ret;
28785}
28786#endif
28787
28788#ifdef __LITTLE_ENDIAN__
28789__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28790  uint16x4x2_t __ret;
28791  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
28792  return __ret;
28793}
28794#else
28795__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28796  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28797  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28798  uint16x4x2_t __ret;
28799  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28800
28801  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28802  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28803  return __ret;
28804}
28805#endif
28806
28807#ifdef __LITTLE_ENDIAN__
28808__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28809  int8x8x2_t __ret;
28810  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
28811  return __ret;
28812}
28813#else
28814__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28815  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28816  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28817  int8x8x2_t __ret;
28818  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28819
28820  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28821  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28822  return __ret;
28823}
28824#endif
28825
28826#ifdef __LITTLE_ENDIAN__
28827__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28828  float32x2x2_t __ret;
28829  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
28830  return __ret;
28831}
28832#else
28833__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28834  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28835  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28836  float32x2x2_t __ret;
28837  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
28838
28839  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28840  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28841  return __ret;
28842}
28843#endif
28844
28845#ifdef __LITTLE_ENDIAN__
28846__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28847  int32x2x2_t __ret;
28848  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
28849  return __ret;
28850}
28851#else
28852__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28853  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28854  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28855  int32x2x2_t __ret;
28856  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
28857
28858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28860  return __ret;
28861}
28862#endif
28863
28864#ifdef __LITTLE_ENDIAN__
28865__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28866  int16x4x2_t __ret;
28867  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
28868  return __ret;
28869}
28870#else
28871__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28872  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28873  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28874  int16x4x2_t __ret;
28875  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
28876
28877  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28878  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28879  return __ret;
28880}
28881#endif
28882
28883#ifdef __LITTLE_ENDIAN__
28884__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28885  uint8x8_t __ret;
28886  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28887  return __ret;
28888}
28889#else
28890__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28891  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28892  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28893  uint8x8_t __ret;
28894  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28895  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28896  return __ret;
28897}
28898#endif
28899
28900#ifdef __LITTLE_ENDIAN__
28901__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28902  uint16x4_t __ret;
28903  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
28904  return __ret;
28905}
28906#else
28907__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28908  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28909  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28910  uint16x4_t __ret;
28911  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28913  return __ret;
28914}
28915#endif
28916
28917#ifdef __LITTLE_ENDIAN__
28918__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28919  uint8x16_t __ret;
28920  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
28921  return __ret;
28922}
28923#else
28924__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28925  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28926  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28927  uint8x16_t __ret;
28928  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28929  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28930  return __ret;
28931}
28932#endif
28933
28934#ifdef __LITTLE_ENDIAN__
28935__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28936  uint16x8_t __ret;
28937  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
28938  return __ret;
28939}
28940#else
28941__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28942  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28943  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28944  uint16x8_t __ret;
28945  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28947  return __ret;
28948}
28949#endif
28950
28951#ifdef __LITTLE_ENDIAN__
28952__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28953  uint8x16_t __ret;
28954  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
28955  return __ret;
28956}
28957#else
28958__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28959  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28960  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28961  uint8x16_t __ret;
28962  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28963  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28964  return __ret;
28965}
28966#endif
28967
28968#ifdef __LITTLE_ENDIAN__
28969__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28970  uint32x4_t __ret;
28971  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
28972  return __ret;
28973}
28974#else
28975__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28976  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28977  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28978  uint32x4_t __ret;
28979  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
28980  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28981  return __ret;
28982}
28983#endif
28984
28985#ifdef __LITTLE_ENDIAN__
28986__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28987  uint16x8_t __ret;
28988  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
28989  return __ret;
28990}
28991#else
28992__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28993  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28994  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28995  uint16x8_t __ret;
28996  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28997  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28998  return __ret;
28999}
29000#endif
29001
29002#ifdef __LITTLE_ENDIAN__
29003__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29004  uint8x16_t __ret;
29005  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29006  return __ret;
29007}
29008#else
29009__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29010  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29011  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29012  uint8x16_t __ret;
29013  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29014  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29015  return __ret;
29016}
29017#endif
29018
29019#ifdef __LITTLE_ENDIAN__
29020__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29021  uint32x4_t __ret;
29022  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29023  return __ret;
29024}
29025#else
29026__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29027  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29028  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29029  uint32x4_t __ret;
29030  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29031  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29032  return __ret;
29033}
29034#endif
29035
29036#ifdef __LITTLE_ENDIAN__
29037__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29038  uint16x8_t __ret;
29039  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29040  return __ret;
29041}
29042#else
29043__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29044  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29045  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29046  uint16x8_t __ret;
29047  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29048  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29049  return __ret;
29050}
29051#endif
29052
29053#ifdef __LITTLE_ENDIAN__
29054__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29055  uint8x8_t __ret;
29056  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29057  return __ret;
29058}
29059#else
29060__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29061  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29062  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29063  uint8x8_t __ret;
29064  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29065  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29066  return __ret;
29067}
29068#endif
29069
29070#ifdef __LITTLE_ENDIAN__
29071__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29072  uint32x2_t __ret;
29073  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29074  return __ret;
29075}
29076#else
29077__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29078  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29079  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29080  uint32x2_t __ret;
29081  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29082  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29083  return __ret;
29084}
29085#endif
29086
29087#ifdef __LITTLE_ENDIAN__
29088__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29089  uint16x4_t __ret;
29090  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29091  return __ret;
29092}
29093#else
29094__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29095  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29096  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29097  uint16x4_t __ret;
29098  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29099  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29100  return __ret;
29101}
29102#endif
29103
29104#ifdef __LITTLE_ENDIAN__
29105__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29106  uint8x8_t __ret;
29107  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29108  return __ret;
29109}
29110#else
29111__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29112  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29113  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29114  uint8x8_t __ret;
29115  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29116  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29117  return __ret;
29118}
29119#endif
29120
29121#ifdef __LITTLE_ENDIAN__
29122__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29123  uint32x2_t __ret;
29124  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29125  return __ret;
29126}
29127#else
29128__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29129  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29130  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29131  uint32x2_t __ret;
29132  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29133  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29134  return __ret;
29135}
29136#endif
29137
29138#ifdef __LITTLE_ENDIAN__
29139__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29140  uint16x4_t __ret;
29141  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29142  return __ret;
29143}
29144#else
29145__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29146  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29147  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29148  uint16x4_t __ret;
29149  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29150  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29151  return __ret;
29152}
29153#endif
29154
29155#ifdef __LITTLE_ENDIAN__
29156__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29157  poly8x8x2_t __ret;
29158  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29159  return __ret;
29160}
29161#else
29162__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29163  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29164  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29165  poly8x8x2_t __ret;
29166  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29167
29168  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29169  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29170  return __ret;
29171}
29172#endif
29173
29174#ifdef __LITTLE_ENDIAN__
29175__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29176  poly16x4x2_t __ret;
29177  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29178  return __ret;
29179}
29180#else
29181__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29182  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29183  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29184  poly16x4x2_t __ret;
29185  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29186
29187  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29188  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29189  return __ret;
29190}
29191#endif
29192
29193#ifdef __LITTLE_ENDIAN__
29194__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29195  poly8x16x2_t __ret;
29196  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29197  return __ret;
29198}
29199#else
29200__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29201  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29202  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29203  poly8x16x2_t __ret;
29204  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29205
29206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29208  return __ret;
29209}
29210#endif
29211
29212#ifdef __LITTLE_ENDIAN__
29213__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29214  poly16x8x2_t __ret;
29215  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29216  return __ret;
29217}
29218#else
29219__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29220  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29221  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29222  poly16x8x2_t __ret;
29223  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29224
29225  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29226  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29227  return __ret;
29228}
29229#endif
29230
29231#ifdef __LITTLE_ENDIAN__
29232__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29233  uint8x16x2_t __ret;
29234  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29235  return __ret;
29236}
29237#else
29238__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29239  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29240  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29241  uint8x16x2_t __ret;
29242  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29243
29244  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29245  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29246  return __ret;
29247}
29248#endif
29249
29250#ifdef __LITTLE_ENDIAN__
29251__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29252  uint32x4x2_t __ret;
29253  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29254  return __ret;
29255}
29256#else
29257__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29258  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29259  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29260  uint32x4x2_t __ret;
29261  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29262
29263  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29264  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29265  return __ret;
29266}
29267#endif
29268
29269#ifdef __LITTLE_ENDIAN__
29270__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29271  uint16x8x2_t __ret;
29272  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29273  return __ret;
29274}
29275#else
29276__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29277  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29278  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29279  uint16x8x2_t __ret;
29280  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29281
29282  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29283  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29284  return __ret;
29285}
29286#endif
29287
29288#ifdef __LITTLE_ENDIAN__
29289__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29290  int8x16x2_t __ret;
29291  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29292  return __ret;
29293}
29294#else
29295__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29296  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29297  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29298  int8x16x2_t __ret;
29299  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29300
29301  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29302  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29303  return __ret;
29304}
29305#endif
29306
29307#ifdef __LITTLE_ENDIAN__
29308__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29309  float32x4x2_t __ret;
29310  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29311  return __ret;
29312}
29313#else
29314__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29315  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29316  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29317  float32x4x2_t __ret;
29318  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29319
29320  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29321  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29322  return __ret;
29323}
29324#endif
29325
29326#ifdef __LITTLE_ENDIAN__
29327__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29328  int32x4x2_t __ret;
29329  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29330  return __ret;
29331}
29332#else
29333__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29334  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29335  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29336  int32x4x2_t __ret;
29337  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29338
29339  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29340  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29341  return __ret;
29342}
29343#endif
29344
29345#ifdef __LITTLE_ENDIAN__
29346__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29347  int16x8x2_t __ret;
29348  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29349  return __ret;
29350}
29351#else
29352__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29353  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29354  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29355  int16x8x2_t __ret;
29356  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29357
29358  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29359  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29360  return __ret;
29361}
29362#endif
29363
29364#ifdef __LITTLE_ENDIAN__
29365__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29366  uint8x8x2_t __ret;
29367  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29368  return __ret;
29369}
29370#else
29371__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29372  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29373  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29374  uint8x8x2_t __ret;
29375  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29376
29377  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29378  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29379  return __ret;
29380}
29381#endif
29382
29383#ifdef __LITTLE_ENDIAN__
29384__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29385  uint32x2x2_t __ret;
29386  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29387  return __ret;
29388}
29389#else
29390__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29391  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29392  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29393  uint32x2x2_t __ret;
29394  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29395
29396  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29397  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29398  return __ret;
29399}
29400#endif
29401
29402#ifdef __LITTLE_ENDIAN__
29403__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29404  uint16x4x2_t __ret;
29405  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29406  return __ret;
29407}
29408#else
29409__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29410  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29411  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29412  uint16x4x2_t __ret;
29413  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29414
29415  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29416  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29417  return __ret;
29418}
29419#endif
29420
29421#ifdef __LITTLE_ENDIAN__
29422__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29423  int8x8x2_t __ret;
29424  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29425  return __ret;
29426}
29427#else
29428__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29429  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29430  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29431  int8x8x2_t __ret;
29432  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29433
29434  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29435  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29436  return __ret;
29437}
29438#endif
29439
29440#ifdef __LITTLE_ENDIAN__
29441__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29442  float32x2x2_t __ret;
29443  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29444  return __ret;
29445}
29446#else
29447__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29448  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29449  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29450  float32x2x2_t __ret;
29451  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29452
29453  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29454  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29455  return __ret;
29456}
29457#endif
29458
29459#ifdef __LITTLE_ENDIAN__
29460__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29461  int32x2x2_t __ret;
29462  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29463  return __ret;
29464}
29465#else
29466__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29467  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29468  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29469  int32x2x2_t __ret;
29470  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29471
29472  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29473  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29474  return __ret;
29475}
29476#endif
29477
29478#ifdef __LITTLE_ENDIAN__
29479__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29480  int16x4x2_t __ret;
29481  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29482  return __ret;
29483}
29484#else
29485__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29486  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29487  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29488  int16x4x2_t __ret;
29489  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29490
29491  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29492  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29493  return __ret;
29494}
29495#endif
29496
29497#ifdef __LITTLE_ENDIAN__
29498__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29499  poly8x8x2_t __ret;
29500  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29501  return __ret;
29502}
29503#else
29504__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29505  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29506  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29507  poly8x8x2_t __ret;
29508  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29509
29510  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29511  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29512  return __ret;
29513}
29514#endif
29515
29516#ifdef __LITTLE_ENDIAN__
29517__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29518  poly16x4x2_t __ret;
29519  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29520  return __ret;
29521}
29522#else
29523__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29524  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29525  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29526  poly16x4x2_t __ret;
29527  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29528
29529  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29530  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29531  return __ret;
29532}
29533#endif
29534
29535#ifdef __LITTLE_ENDIAN__
29536__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29537  poly8x16x2_t __ret;
29538  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29539  return __ret;
29540}
29541#else
29542__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29543  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29544  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29545  poly8x16x2_t __ret;
29546  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29547
29548  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29549  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29550  return __ret;
29551}
29552#endif
29553
29554#ifdef __LITTLE_ENDIAN__
29555__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29556  poly16x8x2_t __ret;
29557  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29558  return __ret;
29559}
29560#else
29561__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29562  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29563  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29564  poly16x8x2_t __ret;
29565  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29566
29567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29569  return __ret;
29570}
29571#endif
29572
29573#ifdef __LITTLE_ENDIAN__
29574__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29575  uint8x16x2_t __ret;
29576  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29577  return __ret;
29578}
29579#else
29580__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29581  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29582  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29583  uint8x16x2_t __ret;
29584  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29585
29586  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29587  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29588  return __ret;
29589}
29590#endif
29591
29592#ifdef __LITTLE_ENDIAN__
29593__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29594  uint32x4x2_t __ret;
29595  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29596  return __ret;
29597}
29598#else
29599__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29600  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29601  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29602  uint32x4x2_t __ret;
29603  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29604
29605  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29606  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29607  return __ret;
29608}
29609#endif
29610
29611#ifdef __LITTLE_ENDIAN__
29612__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29613  uint16x8x2_t __ret;
29614  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29615  return __ret;
29616}
29617#else
29618__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29619  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29620  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29621  uint16x8x2_t __ret;
29622  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29623
29624  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29625  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29626  return __ret;
29627}
29628#endif
29629
29630#ifdef __LITTLE_ENDIAN__
29631__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29632  int8x16x2_t __ret;
29633  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29634  return __ret;
29635}
29636#else
29637__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29638  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29639  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29640  int8x16x2_t __ret;
29641  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29642
29643  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29644  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29645  return __ret;
29646}
29647#endif
29648
29649#ifdef __LITTLE_ENDIAN__
29650__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29651  float32x4x2_t __ret;
29652  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29653  return __ret;
29654}
29655#else
29656__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29657  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29658  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29659  float32x4x2_t __ret;
29660  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29661
29662  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29663  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29664  return __ret;
29665}
29666#endif
29667
29668#ifdef __LITTLE_ENDIAN__
29669__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29670  int32x4x2_t __ret;
29671  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29672  return __ret;
29673}
29674#else
29675__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29676  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29677  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29678  int32x4x2_t __ret;
29679  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29680
29681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29683  return __ret;
29684}
29685#endif
29686
29687#ifdef __LITTLE_ENDIAN__
29688__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29689  int16x8x2_t __ret;
29690  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29691  return __ret;
29692}
29693#else
29694__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29695  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29696  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29697  int16x8x2_t __ret;
29698  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29699
29700  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29701  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29702  return __ret;
29703}
29704#endif
29705
29706#ifdef __LITTLE_ENDIAN__
29707__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29708  uint8x8x2_t __ret;
29709  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29710  return __ret;
29711}
29712#else
29713__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29714  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29715  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29716  uint8x8x2_t __ret;
29717  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29718
29719  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29720  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29721  return __ret;
29722}
29723#endif
29724
29725#ifdef __LITTLE_ENDIAN__
29726__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29727  uint32x2x2_t __ret;
29728  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29729  return __ret;
29730}
29731#else
29732__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29733  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29734  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29735  uint32x2x2_t __ret;
29736  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29737
29738  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29739  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29740  return __ret;
29741}
29742#endif
29743
29744#ifdef __LITTLE_ENDIAN__
29745__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29746  uint16x4x2_t __ret;
29747  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29748  return __ret;
29749}
29750#else
29751__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29752  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29753  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29754  uint16x4x2_t __ret;
29755  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29756
29757  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29758  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29759  return __ret;
29760}
29761#endif
29762
29763#ifdef __LITTLE_ENDIAN__
29764__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29765  int8x8x2_t __ret;
29766  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29767  return __ret;
29768}
29769#else
29770__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29771  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29772  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29773  int8x8x2_t __ret;
29774  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29775
29776  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29777  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29778  return __ret;
29779}
29780#endif
29781
29782#ifdef __LITTLE_ENDIAN__
29783__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29784  float32x2x2_t __ret;
29785  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29786  return __ret;
29787}
29788#else
29789__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29790  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29791  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29792  float32x2x2_t __ret;
29793  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29794
29795  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29796  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29797  return __ret;
29798}
29799#endif
29800
29801#ifdef __LITTLE_ENDIAN__
29802__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29803  int32x2x2_t __ret;
29804  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29805  return __ret;
29806}
29807#else
29808__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29809  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29810  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29811  int32x2x2_t __ret;
29812  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29813
29814  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29815  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29816  return __ret;
29817}
29818#endif
29819
29820#ifdef __LITTLE_ENDIAN__
29821__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29822  int16x4x2_t __ret;
29823  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29824  return __ret;
29825}
29826#else
29827__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29828  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29829  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29830  int16x4x2_t __ret;
29831  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29832
29833  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29834  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29835  return __ret;
29836}
29837#endif
29838
29839#if !defined(__aarch64__)
29840#ifdef __LITTLE_ENDIAN__
29841__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29842  poly8x8_t __ret;
29843  __ret = (poly8x8_t)(__p0);
29844  return __ret;
29845}
29846#else
29847__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29848  poly8x8_t __ret;
29849  __ret = (poly8x8_t)(__p0);
29850  return __ret;
29851}
29852#endif
29853
29854#ifdef __LITTLE_ENDIAN__
29855__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29856  poly8x8_t __ret;
29857  __ret = (poly8x8_t)(__p0);
29858  return __ret;
29859}
29860#else
29861__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29862  poly8x8_t __ret;
29863  __ret = (poly8x8_t)(__p0);
29864  return __ret;
29865}
29866#endif
29867
29868#ifdef __LITTLE_ENDIAN__
29869__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29870  poly8x8_t __ret;
29871  __ret = (poly8x8_t)(__p0);
29872  return __ret;
29873}
29874#else
29875__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29876  poly8x8_t __ret;
29877  __ret = (poly8x8_t)(__p0);
29878  return __ret;
29879}
29880#endif
29881
29882#ifdef __LITTLE_ENDIAN__
29883__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29884  poly8x8_t __ret;
29885  __ret = (poly8x8_t)(__p0);
29886  return __ret;
29887}
29888#else
29889__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29890  poly8x8_t __ret;
29891  __ret = (poly8x8_t)(__p0);
29892  return __ret;
29893}
29894#endif
29895
29896#ifdef __LITTLE_ENDIAN__
29897__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29898  poly8x8_t __ret;
29899  __ret = (poly8x8_t)(__p0);
29900  return __ret;
29901}
29902#else
29903__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29904  poly8x8_t __ret;
29905  __ret = (poly8x8_t)(__p0);
29906  return __ret;
29907}
29908#endif
29909
29910#ifdef __LITTLE_ENDIAN__
29911__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
29912  poly8x8_t __ret;
29913  __ret = (poly8x8_t)(__p0);
29914  return __ret;
29915}
29916#else
29917__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
29918  poly8x8_t __ret;
29919  __ret = (poly8x8_t)(__p0);
29920  return __ret;
29921}
29922#endif
29923
29924#ifdef __LITTLE_ENDIAN__
29925__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
29926  poly8x8_t __ret;
29927  __ret = (poly8x8_t)(__p0);
29928  return __ret;
29929}
29930#else
29931__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
29932  poly8x8_t __ret;
29933  __ret = (poly8x8_t)(__p0);
29934  return __ret;
29935}
29936#endif
29937
29938#ifdef __LITTLE_ENDIAN__
29939__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
29940  poly8x8_t __ret;
29941  __ret = (poly8x8_t)(__p0);
29942  return __ret;
29943}
29944#else
29945__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
29946  poly8x8_t __ret;
29947  __ret = (poly8x8_t)(__p0);
29948  return __ret;
29949}
29950#endif
29951
29952#ifdef __LITTLE_ENDIAN__
29953__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
29954  poly8x8_t __ret;
29955  __ret = (poly8x8_t)(__p0);
29956  return __ret;
29957}
29958#else
29959__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
29960  poly8x8_t __ret;
29961  __ret = (poly8x8_t)(__p0);
29962  return __ret;
29963}
29964#endif
29965
29966#ifdef __LITTLE_ENDIAN__
29967__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
29968  poly8x8_t __ret;
29969  __ret = (poly8x8_t)(__p0);
29970  return __ret;
29971}
29972#else
29973__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
29974  poly8x8_t __ret;
29975  __ret = (poly8x8_t)(__p0);
29976  return __ret;
29977}
29978#endif
29979
29980#ifdef __LITTLE_ENDIAN__
29981__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
29982  poly8x8_t __ret;
29983  __ret = (poly8x8_t)(__p0);
29984  return __ret;
29985}
29986#else
29987__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
29988  poly8x8_t __ret;
29989  __ret = (poly8x8_t)(__p0);
29990  return __ret;
29991}
29992#endif
29993
29994#ifdef __LITTLE_ENDIAN__
29995__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
29996  poly16x4_t __ret;
29997  __ret = (poly16x4_t)(__p0);
29998  return __ret;
29999}
30000#else
30001__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30002  poly16x4_t __ret;
30003  __ret = (poly16x4_t)(__p0);
30004  return __ret;
30005}
30006#endif
30007
30008#ifdef __LITTLE_ENDIAN__
30009__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30010  poly16x4_t __ret;
30011  __ret = (poly16x4_t)(__p0);
30012  return __ret;
30013}
30014#else
30015__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30016  poly16x4_t __ret;
30017  __ret = (poly16x4_t)(__p0);
30018  return __ret;
30019}
30020#endif
30021
30022#ifdef __LITTLE_ENDIAN__
30023__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30024  poly16x4_t __ret;
30025  __ret = (poly16x4_t)(__p0);
30026  return __ret;
30027}
30028#else
30029__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30030  poly16x4_t __ret;
30031  __ret = (poly16x4_t)(__p0);
30032  return __ret;
30033}
30034#endif
30035
30036#ifdef __LITTLE_ENDIAN__
30037__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30038  poly16x4_t __ret;
30039  __ret = (poly16x4_t)(__p0);
30040  return __ret;
30041}
30042#else
30043__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30044  poly16x4_t __ret;
30045  __ret = (poly16x4_t)(__p0);
30046  return __ret;
30047}
30048#endif
30049
30050#ifdef __LITTLE_ENDIAN__
30051__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30052  poly16x4_t __ret;
30053  __ret = (poly16x4_t)(__p0);
30054  return __ret;
30055}
30056#else
30057__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30058  poly16x4_t __ret;
30059  __ret = (poly16x4_t)(__p0);
30060  return __ret;
30061}
30062#endif
30063
30064#ifdef __LITTLE_ENDIAN__
30065__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30066  poly16x4_t __ret;
30067  __ret = (poly16x4_t)(__p0);
30068  return __ret;
30069}
30070#else
30071__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30072  poly16x4_t __ret;
30073  __ret = (poly16x4_t)(__p0);
30074  return __ret;
30075}
30076#endif
30077
30078#ifdef __LITTLE_ENDIAN__
30079__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30080  poly16x4_t __ret;
30081  __ret = (poly16x4_t)(__p0);
30082  return __ret;
30083}
30084#else
30085__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30086  poly16x4_t __ret;
30087  __ret = (poly16x4_t)(__p0);
30088  return __ret;
30089}
30090#endif
30091
30092#ifdef __LITTLE_ENDIAN__
30093__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30094  poly16x4_t __ret;
30095  __ret = (poly16x4_t)(__p0);
30096  return __ret;
30097}
30098#else
30099__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30100  poly16x4_t __ret;
30101  __ret = (poly16x4_t)(__p0);
30102  return __ret;
30103}
30104#endif
30105
30106#ifdef __LITTLE_ENDIAN__
30107__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30108  poly16x4_t __ret;
30109  __ret = (poly16x4_t)(__p0);
30110  return __ret;
30111}
30112#else
30113__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30114  poly16x4_t __ret;
30115  __ret = (poly16x4_t)(__p0);
30116  return __ret;
30117}
30118#endif
30119
30120#ifdef __LITTLE_ENDIAN__
30121__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30122  poly16x4_t __ret;
30123  __ret = (poly16x4_t)(__p0);
30124  return __ret;
30125}
30126#else
30127__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30128  poly16x4_t __ret;
30129  __ret = (poly16x4_t)(__p0);
30130  return __ret;
30131}
30132#endif
30133
30134#ifdef __LITTLE_ENDIAN__
30135__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30136  poly16x4_t __ret;
30137  __ret = (poly16x4_t)(__p0);
30138  return __ret;
30139}
30140#else
30141__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30142  poly16x4_t __ret;
30143  __ret = (poly16x4_t)(__p0);
30144  return __ret;
30145}
30146#endif
30147
30148#ifdef __LITTLE_ENDIAN__
30149__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30150  poly8x16_t __ret;
30151  __ret = (poly8x16_t)(__p0);
30152  return __ret;
30153}
30154#else
30155__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30156  poly8x16_t __ret;
30157  __ret = (poly8x16_t)(__p0);
30158  return __ret;
30159}
30160#endif
30161
30162#ifdef __LITTLE_ENDIAN__
30163__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30164  poly8x16_t __ret;
30165  __ret = (poly8x16_t)(__p0);
30166  return __ret;
30167}
30168#else
30169__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30170  poly8x16_t __ret;
30171  __ret = (poly8x16_t)(__p0);
30172  return __ret;
30173}
30174#endif
30175
30176#ifdef __LITTLE_ENDIAN__
30177__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30178  poly8x16_t __ret;
30179  __ret = (poly8x16_t)(__p0);
30180  return __ret;
30181}
30182#else
30183__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30184  poly8x16_t __ret;
30185  __ret = (poly8x16_t)(__p0);
30186  return __ret;
30187}
30188#endif
30189
30190#ifdef __LITTLE_ENDIAN__
30191__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30192  poly8x16_t __ret;
30193  __ret = (poly8x16_t)(__p0);
30194  return __ret;
30195}
30196#else
30197__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30198  poly8x16_t __ret;
30199  __ret = (poly8x16_t)(__p0);
30200  return __ret;
30201}
30202#endif
30203
30204#ifdef __LITTLE_ENDIAN__
30205__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30206  poly8x16_t __ret;
30207  __ret = (poly8x16_t)(__p0);
30208  return __ret;
30209}
30210#else
30211__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30212  poly8x16_t __ret;
30213  __ret = (poly8x16_t)(__p0);
30214  return __ret;
30215}
30216#endif
30217
30218#ifdef __LITTLE_ENDIAN__
30219__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30220  poly8x16_t __ret;
30221  __ret = (poly8x16_t)(__p0);
30222  return __ret;
30223}
30224#else
30225__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30226  poly8x16_t __ret;
30227  __ret = (poly8x16_t)(__p0);
30228  return __ret;
30229}
30230#endif
30231
30232#ifdef __LITTLE_ENDIAN__
30233__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30234  poly8x16_t __ret;
30235  __ret = (poly8x16_t)(__p0);
30236  return __ret;
30237}
30238#else
30239__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30240  poly8x16_t __ret;
30241  __ret = (poly8x16_t)(__p0);
30242  return __ret;
30243}
30244#endif
30245
30246#ifdef __LITTLE_ENDIAN__
30247__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30248  poly8x16_t __ret;
30249  __ret = (poly8x16_t)(__p0);
30250  return __ret;
30251}
30252#else
30253__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30254  poly8x16_t __ret;
30255  __ret = (poly8x16_t)(__p0);
30256  return __ret;
30257}
30258#endif
30259
30260#ifdef __LITTLE_ENDIAN__
30261__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30262  poly8x16_t __ret;
30263  __ret = (poly8x16_t)(__p0);
30264  return __ret;
30265}
30266#else
30267__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30268  poly8x16_t __ret;
30269  __ret = (poly8x16_t)(__p0);
30270  return __ret;
30271}
30272#endif
30273
30274#ifdef __LITTLE_ENDIAN__
30275__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30276  poly8x16_t __ret;
30277  __ret = (poly8x16_t)(__p0);
30278  return __ret;
30279}
30280#else
30281__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30282  poly8x16_t __ret;
30283  __ret = (poly8x16_t)(__p0);
30284  return __ret;
30285}
30286#endif
30287
30288#ifdef __LITTLE_ENDIAN__
30289__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30290  poly8x16_t __ret;
30291  __ret = (poly8x16_t)(__p0);
30292  return __ret;
30293}
30294#else
30295__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30296  poly8x16_t __ret;
30297  __ret = (poly8x16_t)(__p0);
30298  return __ret;
30299}
30300#endif
30301
30302#ifdef __LITTLE_ENDIAN__
30303__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30304  poly16x8_t __ret;
30305  __ret = (poly16x8_t)(__p0);
30306  return __ret;
30307}
30308#else
30309__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30310  poly16x8_t __ret;
30311  __ret = (poly16x8_t)(__p0);
30312  return __ret;
30313}
30314#endif
30315
30316#ifdef __LITTLE_ENDIAN__
30317__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30318  poly16x8_t __ret;
30319  __ret = (poly16x8_t)(__p0);
30320  return __ret;
30321}
30322#else
30323__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30324  poly16x8_t __ret;
30325  __ret = (poly16x8_t)(__p0);
30326  return __ret;
30327}
30328#endif
30329
30330#ifdef __LITTLE_ENDIAN__
30331__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30332  poly16x8_t __ret;
30333  __ret = (poly16x8_t)(__p0);
30334  return __ret;
30335}
30336#else
30337__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30338  poly16x8_t __ret;
30339  __ret = (poly16x8_t)(__p0);
30340  return __ret;
30341}
30342#endif
30343
30344#ifdef __LITTLE_ENDIAN__
30345__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30346  poly16x8_t __ret;
30347  __ret = (poly16x8_t)(__p0);
30348  return __ret;
30349}
30350#else
30351__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30352  poly16x8_t __ret;
30353  __ret = (poly16x8_t)(__p0);
30354  return __ret;
30355}
30356#endif
30357
30358#ifdef __LITTLE_ENDIAN__
30359__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30360  poly16x8_t __ret;
30361  __ret = (poly16x8_t)(__p0);
30362  return __ret;
30363}
30364#else
30365__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30366  poly16x8_t __ret;
30367  __ret = (poly16x8_t)(__p0);
30368  return __ret;
30369}
30370#endif
30371
30372#ifdef __LITTLE_ENDIAN__
30373__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30374  poly16x8_t __ret;
30375  __ret = (poly16x8_t)(__p0);
30376  return __ret;
30377}
30378#else
30379__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30380  poly16x8_t __ret;
30381  __ret = (poly16x8_t)(__p0);
30382  return __ret;
30383}
30384#endif
30385
30386#ifdef __LITTLE_ENDIAN__
30387__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30388  poly16x8_t __ret;
30389  __ret = (poly16x8_t)(__p0);
30390  return __ret;
30391}
30392#else
30393__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30394  poly16x8_t __ret;
30395  __ret = (poly16x8_t)(__p0);
30396  return __ret;
30397}
30398#endif
30399
30400#ifdef __LITTLE_ENDIAN__
30401__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30402  poly16x8_t __ret;
30403  __ret = (poly16x8_t)(__p0);
30404  return __ret;
30405}
30406#else
30407__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30408  poly16x8_t __ret;
30409  __ret = (poly16x8_t)(__p0);
30410  return __ret;
30411}
30412#endif
30413
30414#ifdef __LITTLE_ENDIAN__
30415__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30416  poly16x8_t __ret;
30417  __ret = (poly16x8_t)(__p0);
30418  return __ret;
30419}
30420#else
30421__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30422  poly16x8_t __ret;
30423  __ret = (poly16x8_t)(__p0);
30424  return __ret;
30425}
30426#endif
30427
30428#ifdef __LITTLE_ENDIAN__
30429__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30430  poly16x8_t __ret;
30431  __ret = (poly16x8_t)(__p0);
30432  return __ret;
30433}
30434#else
30435__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30436  poly16x8_t __ret;
30437  __ret = (poly16x8_t)(__p0);
30438  return __ret;
30439}
30440#endif
30441
30442#ifdef __LITTLE_ENDIAN__
30443__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30444  poly16x8_t __ret;
30445  __ret = (poly16x8_t)(__p0);
30446  return __ret;
30447}
30448#else
30449__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30450  poly16x8_t __ret;
30451  __ret = (poly16x8_t)(__p0);
30452  return __ret;
30453}
30454#endif
30455
30456#ifdef __LITTLE_ENDIAN__
30457__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30458  uint8x16_t __ret;
30459  __ret = (uint8x16_t)(__p0);
30460  return __ret;
30461}
30462#else
30463__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30464  uint8x16_t __ret;
30465  __ret = (uint8x16_t)(__p0);
30466  return __ret;
30467}
30468#endif
30469
30470#ifdef __LITTLE_ENDIAN__
30471__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30472  uint8x16_t __ret;
30473  __ret = (uint8x16_t)(__p0);
30474  return __ret;
30475}
30476#else
30477__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30478  uint8x16_t __ret;
30479  __ret = (uint8x16_t)(__p0);
30480  return __ret;
30481}
30482#endif
30483
30484#ifdef __LITTLE_ENDIAN__
30485__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30486  uint8x16_t __ret;
30487  __ret = (uint8x16_t)(__p0);
30488  return __ret;
30489}
30490#else
30491__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30492  uint8x16_t __ret;
30493  __ret = (uint8x16_t)(__p0);
30494  return __ret;
30495}
30496#endif
30497
30498#ifdef __LITTLE_ENDIAN__
30499__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30500  uint8x16_t __ret;
30501  __ret = (uint8x16_t)(__p0);
30502  return __ret;
30503}
30504#else
30505__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30506  uint8x16_t __ret;
30507  __ret = (uint8x16_t)(__p0);
30508  return __ret;
30509}
30510#endif
30511
30512#ifdef __LITTLE_ENDIAN__
30513__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30514  uint8x16_t __ret;
30515  __ret = (uint8x16_t)(__p0);
30516  return __ret;
30517}
30518#else
30519__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30520  uint8x16_t __ret;
30521  __ret = (uint8x16_t)(__p0);
30522  return __ret;
30523}
30524#endif
30525
30526#ifdef __LITTLE_ENDIAN__
30527__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30528  uint8x16_t __ret;
30529  __ret = (uint8x16_t)(__p0);
30530  return __ret;
30531}
30532#else
30533__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30534  uint8x16_t __ret;
30535  __ret = (uint8x16_t)(__p0);
30536  return __ret;
30537}
30538#endif
30539
30540#ifdef __LITTLE_ENDIAN__
30541__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30542  uint8x16_t __ret;
30543  __ret = (uint8x16_t)(__p0);
30544  return __ret;
30545}
30546#else
30547__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30548  uint8x16_t __ret;
30549  __ret = (uint8x16_t)(__p0);
30550  return __ret;
30551}
30552#endif
30553
30554#ifdef __LITTLE_ENDIAN__
30555__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30556  uint8x16_t __ret;
30557  __ret = (uint8x16_t)(__p0);
30558  return __ret;
30559}
30560#else
30561__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30562  uint8x16_t __ret;
30563  __ret = (uint8x16_t)(__p0);
30564  return __ret;
30565}
30566#endif
30567
30568#ifdef __LITTLE_ENDIAN__
30569__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30570  uint8x16_t __ret;
30571  __ret = (uint8x16_t)(__p0);
30572  return __ret;
30573}
30574#else
30575__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30576  uint8x16_t __ret;
30577  __ret = (uint8x16_t)(__p0);
30578  return __ret;
30579}
30580#endif
30581
30582#ifdef __LITTLE_ENDIAN__
30583__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30584  uint8x16_t __ret;
30585  __ret = (uint8x16_t)(__p0);
30586  return __ret;
30587}
30588#else
30589__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30590  uint8x16_t __ret;
30591  __ret = (uint8x16_t)(__p0);
30592  return __ret;
30593}
30594#endif
30595
30596#ifdef __LITTLE_ENDIAN__
30597__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30598  uint8x16_t __ret;
30599  __ret = (uint8x16_t)(__p0);
30600  return __ret;
30601}
30602#else
30603__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30604  uint8x16_t __ret;
30605  __ret = (uint8x16_t)(__p0);
30606  return __ret;
30607}
30608#endif
30609
30610#ifdef __LITTLE_ENDIAN__
30611__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30612  uint32x4_t __ret;
30613  __ret = (uint32x4_t)(__p0);
30614  return __ret;
30615}
30616#else
30617__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30618  uint32x4_t __ret;
30619  __ret = (uint32x4_t)(__p0);
30620  return __ret;
30621}
30622#endif
30623
30624#ifdef __LITTLE_ENDIAN__
30625__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30626  uint32x4_t __ret;
30627  __ret = (uint32x4_t)(__p0);
30628  return __ret;
30629}
30630#else
30631__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30632  uint32x4_t __ret;
30633  __ret = (uint32x4_t)(__p0);
30634  return __ret;
30635}
30636#endif
30637
30638#ifdef __LITTLE_ENDIAN__
30639__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30640  uint32x4_t __ret;
30641  __ret = (uint32x4_t)(__p0);
30642  return __ret;
30643}
30644#else
30645__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30646  uint32x4_t __ret;
30647  __ret = (uint32x4_t)(__p0);
30648  return __ret;
30649}
30650#endif
30651
30652#ifdef __LITTLE_ENDIAN__
30653__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30654  uint32x4_t __ret;
30655  __ret = (uint32x4_t)(__p0);
30656  return __ret;
30657}
30658#else
30659__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30660  uint32x4_t __ret;
30661  __ret = (uint32x4_t)(__p0);
30662  return __ret;
30663}
30664#endif
30665
30666#ifdef __LITTLE_ENDIAN__
30667__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30668  uint32x4_t __ret;
30669  __ret = (uint32x4_t)(__p0);
30670  return __ret;
30671}
30672#else
30673__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30674  uint32x4_t __ret;
30675  __ret = (uint32x4_t)(__p0);
30676  return __ret;
30677}
30678#endif
30679
30680#ifdef __LITTLE_ENDIAN__
30681__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30682  uint32x4_t __ret;
30683  __ret = (uint32x4_t)(__p0);
30684  return __ret;
30685}
30686#else
30687__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30688  uint32x4_t __ret;
30689  __ret = (uint32x4_t)(__p0);
30690  return __ret;
30691}
30692#endif
30693
30694#ifdef __LITTLE_ENDIAN__
30695__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30696  uint32x4_t __ret;
30697  __ret = (uint32x4_t)(__p0);
30698  return __ret;
30699}
30700#else
30701__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30702  uint32x4_t __ret;
30703  __ret = (uint32x4_t)(__p0);
30704  return __ret;
30705}
30706#endif
30707
30708#ifdef __LITTLE_ENDIAN__
30709__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30710  uint32x4_t __ret;
30711  __ret = (uint32x4_t)(__p0);
30712  return __ret;
30713}
30714#else
30715__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30716  uint32x4_t __ret;
30717  __ret = (uint32x4_t)(__p0);
30718  return __ret;
30719}
30720#endif
30721
30722#ifdef __LITTLE_ENDIAN__
30723__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30724  uint32x4_t __ret;
30725  __ret = (uint32x4_t)(__p0);
30726  return __ret;
30727}
30728#else
30729__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30730  uint32x4_t __ret;
30731  __ret = (uint32x4_t)(__p0);
30732  return __ret;
30733}
30734#endif
30735
30736#ifdef __LITTLE_ENDIAN__
30737__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30738  uint32x4_t __ret;
30739  __ret = (uint32x4_t)(__p0);
30740  return __ret;
30741}
30742#else
30743__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30744  uint32x4_t __ret;
30745  __ret = (uint32x4_t)(__p0);
30746  return __ret;
30747}
30748#endif
30749
30750#ifdef __LITTLE_ENDIAN__
30751__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30752  uint32x4_t __ret;
30753  __ret = (uint32x4_t)(__p0);
30754  return __ret;
30755}
30756#else
30757__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30758  uint32x4_t __ret;
30759  __ret = (uint32x4_t)(__p0);
30760  return __ret;
30761}
30762#endif
30763
30764#ifdef __LITTLE_ENDIAN__
30765__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30766  uint64x2_t __ret;
30767  __ret = (uint64x2_t)(__p0);
30768  return __ret;
30769}
30770#else
30771__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30772  uint64x2_t __ret;
30773  __ret = (uint64x2_t)(__p0);
30774  return __ret;
30775}
30776#endif
30777
30778#ifdef __LITTLE_ENDIAN__
30779__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30780  uint64x2_t __ret;
30781  __ret = (uint64x2_t)(__p0);
30782  return __ret;
30783}
30784#else
30785__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30786  uint64x2_t __ret;
30787  __ret = (uint64x2_t)(__p0);
30788  return __ret;
30789}
30790#endif
30791
30792#ifdef __LITTLE_ENDIAN__
30793__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30794  uint64x2_t __ret;
30795  __ret = (uint64x2_t)(__p0);
30796  return __ret;
30797}
30798#else
30799__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30800  uint64x2_t __ret;
30801  __ret = (uint64x2_t)(__p0);
30802  return __ret;
30803}
30804#endif
30805
30806#ifdef __LITTLE_ENDIAN__
30807__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30808  uint64x2_t __ret;
30809  __ret = (uint64x2_t)(__p0);
30810  return __ret;
30811}
30812#else
30813__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30814  uint64x2_t __ret;
30815  __ret = (uint64x2_t)(__p0);
30816  return __ret;
30817}
30818#endif
30819
30820#ifdef __LITTLE_ENDIAN__
30821__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30822  uint64x2_t __ret;
30823  __ret = (uint64x2_t)(__p0);
30824  return __ret;
30825}
30826#else
30827__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30828  uint64x2_t __ret;
30829  __ret = (uint64x2_t)(__p0);
30830  return __ret;
30831}
30832#endif
30833
30834#ifdef __LITTLE_ENDIAN__
30835__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30836  uint64x2_t __ret;
30837  __ret = (uint64x2_t)(__p0);
30838  return __ret;
30839}
30840#else
30841__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30842  uint64x2_t __ret;
30843  __ret = (uint64x2_t)(__p0);
30844  return __ret;
30845}
30846#endif
30847
30848#ifdef __LITTLE_ENDIAN__
30849__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30850  uint64x2_t __ret;
30851  __ret = (uint64x2_t)(__p0);
30852  return __ret;
30853}
30854#else
30855__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30856  uint64x2_t __ret;
30857  __ret = (uint64x2_t)(__p0);
30858  return __ret;
30859}
30860#endif
30861
30862#ifdef __LITTLE_ENDIAN__
30863__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30864  uint64x2_t __ret;
30865  __ret = (uint64x2_t)(__p0);
30866  return __ret;
30867}
30868#else
30869__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30870  uint64x2_t __ret;
30871  __ret = (uint64x2_t)(__p0);
30872  return __ret;
30873}
30874#endif
30875
30876#ifdef __LITTLE_ENDIAN__
30877__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30878  uint64x2_t __ret;
30879  __ret = (uint64x2_t)(__p0);
30880  return __ret;
30881}
30882#else
30883__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30884  uint64x2_t __ret;
30885  __ret = (uint64x2_t)(__p0);
30886  return __ret;
30887}
30888#endif
30889
30890#ifdef __LITTLE_ENDIAN__
30891__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30892  uint64x2_t __ret;
30893  __ret = (uint64x2_t)(__p0);
30894  return __ret;
30895}
30896#else
30897__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30898  uint64x2_t __ret;
30899  __ret = (uint64x2_t)(__p0);
30900  return __ret;
30901}
30902#endif
30903
30904#ifdef __LITTLE_ENDIAN__
30905__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30906  uint64x2_t __ret;
30907  __ret = (uint64x2_t)(__p0);
30908  return __ret;
30909}
30910#else
30911__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30912  uint64x2_t __ret;
30913  __ret = (uint64x2_t)(__p0);
30914  return __ret;
30915}
30916#endif
30917
30918#ifdef __LITTLE_ENDIAN__
30919__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
30920  uint16x8_t __ret;
30921  __ret = (uint16x8_t)(__p0);
30922  return __ret;
30923}
30924#else
30925__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
30926  uint16x8_t __ret;
30927  __ret = (uint16x8_t)(__p0);
30928  return __ret;
30929}
30930#endif
30931
30932#ifdef __LITTLE_ENDIAN__
30933__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
30934  uint16x8_t __ret;
30935  __ret = (uint16x8_t)(__p0);
30936  return __ret;
30937}
30938#else
30939__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
30940  uint16x8_t __ret;
30941  __ret = (uint16x8_t)(__p0);
30942  return __ret;
30943}
30944#endif
30945
30946#ifdef __LITTLE_ENDIAN__
30947__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
30948  uint16x8_t __ret;
30949  __ret = (uint16x8_t)(__p0);
30950  return __ret;
30951}
30952#else
30953__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
30954  uint16x8_t __ret;
30955  __ret = (uint16x8_t)(__p0);
30956  return __ret;
30957}
30958#endif
30959
30960#ifdef __LITTLE_ENDIAN__
30961__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
30962  uint16x8_t __ret;
30963  __ret = (uint16x8_t)(__p0);
30964  return __ret;
30965}
30966#else
30967__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
30968  uint16x8_t __ret;
30969  __ret = (uint16x8_t)(__p0);
30970  return __ret;
30971}
30972#endif
30973
30974#ifdef __LITTLE_ENDIAN__
30975__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
30976  uint16x8_t __ret;
30977  __ret = (uint16x8_t)(__p0);
30978  return __ret;
30979}
30980#else
30981__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
30982  uint16x8_t __ret;
30983  __ret = (uint16x8_t)(__p0);
30984  return __ret;
30985}
30986#endif
30987
30988#ifdef __LITTLE_ENDIAN__
30989__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
30990  uint16x8_t __ret;
30991  __ret = (uint16x8_t)(__p0);
30992  return __ret;
30993}
30994#else
30995__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
30996  uint16x8_t __ret;
30997  __ret = (uint16x8_t)(__p0);
30998  return __ret;
30999}
31000#endif
31001
31002#ifdef __LITTLE_ENDIAN__
31003__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31004  uint16x8_t __ret;
31005  __ret = (uint16x8_t)(__p0);
31006  return __ret;
31007}
31008#else
31009__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31010  uint16x8_t __ret;
31011  __ret = (uint16x8_t)(__p0);
31012  return __ret;
31013}
31014#endif
31015
31016#ifdef __LITTLE_ENDIAN__
31017__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31018  uint16x8_t __ret;
31019  __ret = (uint16x8_t)(__p0);
31020  return __ret;
31021}
31022#else
31023__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31024  uint16x8_t __ret;
31025  __ret = (uint16x8_t)(__p0);
31026  return __ret;
31027}
31028#endif
31029
31030#ifdef __LITTLE_ENDIAN__
31031__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31032  uint16x8_t __ret;
31033  __ret = (uint16x8_t)(__p0);
31034  return __ret;
31035}
31036#else
31037__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31038  uint16x8_t __ret;
31039  __ret = (uint16x8_t)(__p0);
31040  return __ret;
31041}
31042#endif
31043
31044#ifdef __LITTLE_ENDIAN__
31045__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31046  uint16x8_t __ret;
31047  __ret = (uint16x8_t)(__p0);
31048  return __ret;
31049}
31050#else
31051__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31052  uint16x8_t __ret;
31053  __ret = (uint16x8_t)(__p0);
31054  return __ret;
31055}
31056#endif
31057
31058#ifdef __LITTLE_ENDIAN__
31059__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31060  uint16x8_t __ret;
31061  __ret = (uint16x8_t)(__p0);
31062  return __ret;
31063}
31064#else
31065__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31066  uint16x8_t __ret;
31067  __ret = (uint16x8_t)(__p0);
31068  return __ret;
31069}
31070#endif
31071
31072#ifdef __LITTLE_ENDIAN__
31073__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31074  int8x16_t __ret;
31075  __ret = (int8x16_t)(__p0);
31076  return __ret;
31077}
31078#else
31079__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31080  int8x16_t __ret;
31081  __ret = (int8x16_t)(__p0);
31082  return __ret;
31083}
31084#endif
31085
31086#ifdef __LITTLE_ENDIAN__
31087__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31088  int8x16_t __ret;
31089  __ret = (int8x16_t)(__p0);
31090  return __ret;
31091}
31092#else
31093__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31094  int8x16_t __ret;
31095  __ret = (int8x16_t)(__p0);
31096  return __ret;
31097}
31098#endif
31099
31100#ifdef __LITTLE_ENDIAN__
31101__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31102  int8x16_t __ret;
31103  __ret = (int8x16_t)(__p0);
31104  return __ret;
31105}
31106#else
31107__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31108  int8x16_t __ret;
31109  __ret = (int8x16_t)(__p0);
31110  return __ret;
31111}
31112#endif
31113
31114#ifdef __LITTLE_ENDIAN__
31115__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31116  int8x16_t __ret;
31117  __ret = (int8x16_t)(__p0);
31118  return __ret;
31119}
31120#else
31121__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31122  int8x16_t __ret;
31123  __ret = (int8x16_t)(__p0);
31124  return __ret;
31125}
31126#endif
31127
31128#ifdef __LITTLE_ENDIAN__
31129__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31130  int8x16_t __ret;
31131  __ret = (int8x16_t)(__p0);
31132  return __ret;
31133}
31134#else
31135__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31136  int8x16_t __ret;
31137  __ret = (int8x16_t)(__p0);
31138  return __ret;
31139}
31140#endif
31141
31142#ifdef __LITTLE_ENDIAN__
31143__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31144  int8x16_t __ret;
31145  __ret = (int8x16_t)(__p0);
31146  return __ret;
31147}
31148#else
31149__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31150  int8x16_t __ret;
31151  __ret = (int8x16_t)(__p0);
31152  return __ret;
31153}
31154#endif
31155
31156#ifdef __LITTLE_ENDIAN__
31157__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31158  int8x16_t __ret;
31159  __ret = (int8x16_t)(__p0);
31160  return __ret;
31161}
31162#else
31163__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31164  int8x16_t __ret;
31165  __ret = (int8x16_t)(__p0);
31166  return __ret;
31167}
31168#endif
31169
31170#ifdef __LITTLE_ENDIAN__
31171__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31172  int8x16_t __ret;
31173  __ret = (int8x16_t)(__p0);
31174  return __ret;
31175}
31176#else
31177__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31178  int8x16_t __ret;
31179  __ret = (int8x16_t)(__p0);
31180  return __ret;
31181}
31182#endif
31183
31184#ifdef __LITTLE_ENDIAN__
31185__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31186  int8x16_t __ret;
31187  __ret = (int8x16_t)(__p0);
31188  return __ret;
31189}
31190#else
31191__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31192  int8x16_t __ret;
31193  __ret = (int8x16_t)(__p0);
31194  return __ret;
31195}
31196#endif
31197
31198#ifdef __LITTLE_ENDIAN__
31199__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31200  int8x16_t __ret;
31201  __ret = (int8x16_t)(__p0);
31202  return __ret;
31203}
31204#else
31205__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31206  int8x16_t __ret;
31207  __ret = (int8x16_t)(__p0);
31208  return __ret;
31209}
31210#endif
31211
31212#ifdef __LITTLE_ENDIAN__
31213__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31214  int8x16_t __ret;
31215  __ret = (int8x16_t)(__p0);
31216  return __ret;
31217}
31218#else
31219__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31220  int8x16_t __ret;
31221  __ret = (int8x16_t)(__p0);
31222  return __ret;
31223}
31224#endif
31225
31226#ifdef __LITTLE_ENDIAN__
31227__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31228  float32x4_t __ret;
31229  __ret = (float32x4_t)(__p0);
31230  return __ret;
31231}
31232#else
31233__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31234  float32x4_t __ret;
31235  __ret = (float32x4_t)(__p0);
31236  return __ret;
31237}
31238#endif
31239
31240#ifdef __LITTLE_ENDIAN__
31241__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31242  float32x4_t __ret;
31243  __ret = (float32x4_t)(__p0);
31244  return __ret;
31245}
31246#else
31247__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31248  float32x4_t __ret;
31249  __ret = (float32x4_t)(__p0);
31250  return __ret;
31251}
31252#endif
31253
31254#ifdef __LITTLE_ENDIAN__
31255__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31256  float32x4_t __ret;
31257  __ret = (float32x4_t)(__p0);
31258  return __ret;
31259}
31260#else
31261__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31262  float32x4_t __ret;
31263  __ret = (float32x4_t)(__p0);
31264  return __ret;
31265}
31266#endif
31267
31268#ifdef __LITTLE_ENDIAN__
31269__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31270  float32x4_t __ret;
31271  __ret = (float32x4_t)(__p0);
31272  return __ret;
31273}
31274#else
31275__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31276  float32x4_t __ret;
31277  __ret = (float32x4_t)(__p0);
31278  return __ret;
31279}
31280#endif
31281
31282#ifdef __LITTLE_ENDIAN__
31283__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31284  float32x4_t __ret;
31285  __ret = (float32x4_t)(__p0);
31286  return __ret;
31287}
31288#else
31289__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31290  float32x4_t __ret;
31291  __ret = (float32x4_t)(__p0);
31292  return __ret;
31293}
31294#endif
31295
31296#ifdef __LITTLE_ENDIAN__
31297__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31298  float32x4_t __ret;
31299  __ret = (float32x4_t)(__p0);
31300  return __ret;
31301}
31302#else
31303__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31304  float32x4_t __ret;
31305  __ret = (float32x4_t)(__p0);
31306  return __ret;
31307}
31308#endif
31309
31310#ifdef __LITTLE_ENDIAN__
31311__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31312  float32x4_t __ret;
31313  __ret = (float32x4_t)(__p0);
31314  return __ret;
31315}
31316#else
31317__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31318  float32x4_t __ret;
31319  __ret = (float32x4_t)(__p0);
31320  return __ret;
31321}
31322#endif
31323
31324#ifdef __LITTLE_ENDIAN__
31325__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31326  float32x4_t __ret;
31327  __ret = (float32x4_t)(__p0);
31328  return __ret;
31329}
31330#else
31331__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31332  float32x4_t __ret;
31333  __ret = (float32x4_t)(__p0);
31334  return __ret;
31335}
31336#endif
31337
31338#ifdef __LITTLE_ENDIAN__
31339__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31340  float32x4_t __ret;
31341  __ret = (float32x4_t)(__p0);
31342  return __ret;
31343}
31344#else
31345__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31346  float32x4_t __ret;
31347  __ret = (float32x4_t)(__p0);
31348  return __ret;
31349}
31350#endif
31351
31352#ifdef __LITTLE_ENDIAN__
31353__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31354  float32x4_t __ret;
31355  __ret = (float32x4_t)(__p0);
31356  return __ret;
31357}
31358#else
31359__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31360  float32x4_t __ret;
31361  __ret = (float32x4_t)(__p0);
31362  return __ret;
31363}
31364#endif
31365
31366#ifdef __LITTLE_ENDIAN__
31367__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31368  float32x4_t __ret;
31369  __ret = (float32x4_t)(__p0);
31370  return __ret;
31371}
31372#else
31373__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31374  float32x4_t __ret;
31375  __ret = (float32x4_t)(__p0);
31376  return __ret;
31377}
31378#endif
31379
31380#ifdef __LITTLE_ENDIAN__
31381__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31382  float16x8_t __ret;
31383  __ret = (float16x8_t)(__p0);
31384  return __ret;
31385}
31386#else
31387__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31388  float16x8_t __ret;
31389  __ret = (float16x8_t)(__p0);
31390  return __ret;
31391}
31392#endif
31393
31394#ifdef __LITTLE_ENDIAN__
31395__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31396  float16x8_t __ret;
31397  __ret = (float16x8_t)(__p0);
31398  return __ret;
31399}
31400#else
31401__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31402  float16x8_t __ret;
31403  __ret = (float16x8_t)(__p0);
31404  return __ret;
31405}
31406#endif
31407
31408#ifdef __LITTLE_ENDIAN__
31409__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31410  float16x8_t __ret;
31411  __ret = (float16x8_t)(__p0);
31412  return __ret;
31413}
31414#else
31415__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31416  float16x8_t __ret;
31417  __ret = (float16x8_t)(__p0);
31418  return __ret;
31419}
31420#endif
31421
31422#ifdef __LITTLE_ENDIAN__
31423__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31424  float16x8_t __ret;
31425  __ret = (float16x8_t)(__p0);
31426  return __ret;
31427}
31428#else
31429__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31430  float16x8_t __ret;
31431  __ret = (float16x8_t)(__p0);
31432  return __ret;
31433}
31434#endif
31435
31436#ifdef __LITTLE_ENDIAN__
31437__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31438  float16x8_t __ret;
31439  __ret = (float16x8_t)(__p0);
31440  return __ret;
31441}
31442#else
31443__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31444  float16x8_t __ret;
31445  __ret = (float16x8_t)(__p0);
31446  return __ret;
31447}
31448#endif
31449
31450#ifdef __LITTLE_ENDIAN__
31451__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31452  float16x8_t __ret;
31453  __ret = (float16x8_t)(__p0);
31454  return __ret;
31455}
31456#else
31457__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31458  float16x8_t __ret;
31459  __ret = (float16x8_t)(__p0);
31460  return __ret;
31461}
31462#endif
31463
31464#ifdef __LITTLE_ENDIAN__
31465__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31466  float16x8_t __ret;
31467  __ret = (float16x8_t)(__p0);
31468  return __ret;
31469}
31470#else
31471__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31472  float16x8_t __ret;
31473  __ret = (float16x8_t)(__p0);
31474  return __ret;
31475}
31476#endif
31477
31478#ifdef __LITTLE_ENDIAN__
31479__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31480  float16x8_t __ret;
31481  __ret = (float16x8_t)(__p0);
31482  return __ret;
31483}
31484#else
31485__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31486  float16x8_t __ret;
31487  __ret = (float16x8_t)(__p0);
31488  return __ret;
31489}
31490#endif
31491
31492#ifdef __LITTLE_ENDIAN__
31493__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31494  float16x8_t __ret;
31495  __ret = (float16x8_t)(__p0);
31496  return __ret;
31497}
31498#else
31499__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31500  float16x8_t __ret;
31501  __ret = (float16x8_t)(__p0);
31502  return __ret;
31503}
31504#endif
31505
31506#ifdef __LITTLE_ENDIAN__
31507__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31508  float16x8_t __ret;
31509  __ret = (float16x8_t)(__p0);
31510  return __ret;
31511}
31512#else
31513__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31514  float16x8_t __ret;
31515  __ret = (float16x8_t)(__p0);
31516  return __ret;
31517}
31518#endif
31519
31520#ifdef __LITTLE_ENDIAN__
31521__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31522  float16x8_t __ret;
31523  __ret = (float16x8_t)(__p0);
31524  return __ret;
31525}
31526#else
31527__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31528  float16x8_t __ret;
31529  __ret = (float16x8_t)(__p0);
31530  return __ret;
31531}
31532#endif
31533
31534#ifdef __LITTLE_ENDIAN__
31535__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31536  int32x4_t __ret;
31537  __ret = (int32x4_t)(__p0);
31538  return __ret;
31539}
31540#else
31541__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31542  int32x4_t __ret;
31543  __ret = (int32x4_t)(__p0);
31544  return __ret;
31545}
31546#endif
31547
31548#ifdef __LITTLE_ENDIAN__
31549__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31550  int32x4_t __ret;
31551  __ret = (int32x4_t)(__p0);
31552  return __ret;
31553}
31554#else
31555__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31556  int32x4_t __ret;
31557  __ret = (int32x4_t)(__p0);
31558  return __ret;
31559}
31560#endif
31561
31562#ifdef __LITTLE_ENDIAN__
31563__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31564  int32x4_t __ret;
31565  __ret = (int32x4_t)(__p0);
31566  return __ret;
31567}
31568#else
31569__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31570  int32x4_t __ret;
31571  __ret = (int32x4_t)(__p0);
31572  return __ret;
31573}
31574#endif
31575
31576#ifdef __LITTLE_ENDIAN__
31577__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31578  int32x4_t __ret;
31579  __ret = (int32x4_t)(__p0);
31580  return __ret;
31581}
31582#else
31583__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31584  int32x4_t __ret;
31585  __ret = (int32x4_t)(__p0);
31586  return __ret;
31587}
31588#endif
31589
31590#ifdef __LITTLE_ENDIAN__
31591__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31592  int32x4_t __ret;
31593  __ret = (int32x4_t)(__p0);
31594  return __ret;
31595}
31596#else
31597__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31598  int32x4_t __ret;
31599  __ret = (int32x4_t)(__p0);
31600  return __ret;
31601}
31602#endif
31603
31604#ifdef __LITTLE_ENDIAN__
31605__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31606  int32x4_t __ret;
31607  __ret = (int32x4_t)(__p0);
31608  return __ret;
31609}
31610#else
31611__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31612  int32x4_t __ret;
31613  __ret = (int32x4_t)(__p0);
31614  return __ret;
31615}
31616#endif
31617
31618#ifdef __LITTLE_ENDIAN__
31619__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31620  int32x4_t __ret;
31621  __ret = (int32x4_t)(__p0);
31622  return __ret;
31623}
31624#else
31625__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31626  int32x4_t __ret;
31627  __ret = (int32x4_t)(__p0);
31628  return __ret;
31629}
31630#endif
31631
31632#ifdef __LITTLE_ENDIAN__
31633__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31634  int32x4_t __ret;
31635  __ret = (int32x4_t)(__p0);
31636  return __ret;
31637}
31638#else
31639__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31640  int32x4_t __ret;
31641  __ret = (int32x4_t)(__p0);
31642  return __ret;
31643}
31644#endif
31645
31646#ifdef __LITTLE_ENDIAN__
31647__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31648  int32x4_t __ret;
31649  __ret = (int32x4_t)(__p0);
31650  return __ret;
31651}
31652#else
31653__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31654  int32x4_t __ret;
31655  __ret = (int32x4_t)(__p0);
31656  return __ret;
31657}
31658#endif
31659
31660#ifdef __LITTLE_ENDIAN__
31661__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31662  int32x4_t __ret;
31663  __ret = (int32x4_t)(__p0);
31664  return __ret;
31665}
31666#else
31667__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31668  int32x4_t __ret;
31669  __ret = (int32x4_t)(__p0);
31670  return __ret;
31671}
31672#endif
31673
31674#ifdef __LITTLE_ENDIAN__
31675__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31676  int32x4_t __ret;
31677  __ret = (int32x4_t)(__p0);
31678  return __ret;
31679}
31680#else
31681__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31682  int32x4_t __ret;
31683  __ret = (int32x4_t)(__p0);
31684  return __ret;
31685}
31686#endif
31687
31688#ifdef __LITTLE_ENDIAN__
31689__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31690  int64x2_t __ret;
31691  __ret = (int64x2_t)(__p0);
31692  return __ret;
31693}
31694#else
31695__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31696  int64x2_t __ret;
31697  __ret = (int64x2_t)(__p0);
31698  return __ret;
31699}
31700#endif
31701
31702#ifdef __LITTLE_ENDIAN__
31703__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31704  int64x2_t __ret;
31705  __ret = (int64x2_t)(__p0);
31706  return __ret;
31707}
31708#else
31709__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31710  int64x2_t __ret;
31711  __ret = (int64x2_t)(__p0);
31712  return __ret;
31713}
31714#endif
31715
31716#ifdef __LITTLE_ENDIAN__
31717__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31718  int64x2_t __ret;
31719  __ret = (int64x2_t)(__p0);
31720  return __ret;
31721}
31722#else
31723__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31724  int64x2_t __ret;
31725  __ret = (int64x2_t)(__p0);
31726  return __ret;
31727}
31728#endif
31729
31730#ifdef __LITTLE_ENDIAN__
31731__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31732  int64x2_t __ret;
31733  __ret = (int64x2_t)(__p0);
31734  return __ret;
31735}
31736#else
31737__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31738  int64x2_t __ret;
31739  __ret = (int64x2_t)(__p0);
31740  return __ret;
31741}
31742#endif
31743
31744#ifdef __LITTLE_ENDIAN__
31745__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31746  int64x2_t __ret;
31747  __ret = (int64x2_t)(__p0);
31748  return __ret;
31749}
31750#else
31751__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31752  int64x2_t __ret;
31753  __ret = (int64x2_t)(__p0);
31754  return __ret;
31755}
31756#endif
31757
31758#ifdef __LITTLE_ENDIAN__
31759__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31760  int64x2_t __ret;
31761  __ret = (int64x2_t)(__p0);
31762  return __ret;
31763}
31764#else
31765__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31766  int64x2_t __ret;
31767  __ret = (int64x2_t)(__p0);
31768  return __ret;
31769}
31770#endif
31771
31772#ifdef __LITTLE_ENDIAN__
31773__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31774  int64x2_t __ret;
31775  __ret = (int64x2_t)(__p0);
31776  return __ret;
31777}
31778#else
31779__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31780  int64x2_t __ret;
31781  __ret = (int64x2_t)(__p0);
31782  return __ret;
31783}
31784#endif
31785
31786#ifdef __LITTLE_ENDIAN__
31787__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31788  int64x2_t __ret;
31789  __ret = (int64x2_t)(__p0);
31790  return __ret;
31791}
31792#else
31793__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31794  int64x2_t __ret;
31795  __ret = (int64x2_t)(__p0);
31796  return __ret;
31797}
31798#endif
31799
31800#ifdef __LITTLE_ENDIAN__
31801__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31802  int64x2_t __ret;
31803  __ret = (int64x2_t)(__p0);
31804  return __ret;
31805}
31806#else
31807__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31808  int64x2_t __ret;
31809  __ret = (int64x2_t)(__p0);
31810  return __ret;
31811}
31812#endif
31813
31814#ifdef __LITTLE_ENDIAN__
31815__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31816  int64x2_t __ret;
31817  __ret = (int64x2_t)(__p0);
31818  return __ret;
31819}
31820#else
31821__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31822  int64x2_t __ret;
31823  __ret = (int64x2_t)(__p0);
31824  return __ret;
31825}
31826#endif
31827
31828#ifdef __LITTLE_ENDIAN__
31829__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31830  int64x2_t __ret;
31831  __ret = (int64x2_t)(__p0);
31832  return __ret;
31833}
31834#else
31835__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31836  int64x2_t __ret;
31837  __ret = (int64x2_t)(__p0);
31838  return __ret;
31839}
31840#endif
31841
31842#ifdef __LITTLE_ENDIAN__
31843__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31844  int16x8_t __ret;
31845  __ret = (int16x8_t)(__p0);
31846  return __ret;
31847}
31848#else
31849__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31850  int16x8_t __ret;
31851  __ret = (int16x8_t)(__p0);
31852  return __ret;
31853}
31854#endif
31855
31856#ifdef __LITTLE_ENDIAN__
31857__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31858  int16x8_t __ret;
31859  __ret = (int16x8_t)(__p0);
31860  return __ret;
31861}
31862#else
31863__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31864  int16x8_t __ret;
31865  __ret = (int16x8_t)(__p0);
31866  return __ret;
31867}
31868#endif
31869
31870#ifdef __LITTLE_ENDIAN__
31871__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31872  int16x8_t __ret;
31873  __ret = (int16x8_t)(__p0);
31874  return __ret;
31875}
31876#else
31877__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31878  int16x8_t __ret;
31879  __ret = (int16x8_t)(__p0);
31880  return __ret;
31881}
31882#endif
31883
31884#ifdef __LITTLE_ENDIAN__
31885__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31886  int16x8_t __ret;
31887  __ret = (int16x8_t)(__p0);
31888  return __ret;
31889}
31890#else
31891__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31892  int16x8_t __ret;
31893  __ret = (int16x8_t)(__p0);
31894  return __ret;
31895}
31896#endif
31897
31898#ifdef __LITTLE_ENDIAN__
31899__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31900  int16x8_t __ret;
31901  __ret = (int16x8_t)(__p0);
31902  return __ret;
31903}
31904#else
31905__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31906  int16x8_t __ret;
31907  __ret = (int16x8_t)(__p0);
31908  return __ret;
31909}
31910#endif
31911
31912#ifdef __LITTLE_ENDIAN__
31913__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
31914  int16x8_t __ret;
31915  __ret = (int16x8_t)(__p0);
31916  return __ret;
31917}
31918#else
31919__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
31920  int16x8_t __ret;
31921  __ret = (int16x8_t)(__p0);
31922  return __ret;
31923}
31924#endif
31925
31926#ifdef __LITTLE_ENDIAN__
31927__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
31928  int16x8_t __ret;
31929  __ret = (int16x8_t)(__p0);
31930  return __ret;
31931}
31932#else
31933__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
31934  int16x8_t __ret;
31935  __ret = (int16x8_t)(__p0);
31936  return __ret;
31937}
31938#endif
31939
31940#ifdef __LITTLE_ENDIAN__
31941__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
31942  int16x8_t __ret;
31943  __ret = (int16x8_t)(__p0);
31944  return __ret;
31945}
31946#else
31947__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
31948  int16x8_t __ret;
31949  __ret = (int16x8_t)(__p0);
31950  return __ret;
31951}
31952#endif
31953
31954#ifdef __LITTLE_ENDIAN__
31955__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
31956  int16x8_t __ret;
31957  __ret = (int16x8_t)(__p0);
31958  return __ret;
31959}
31960#else
31961__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
31962  int16x8_t __ret;
31963  __ret = (int16x8_t)(__p0);
31964  return __ret;
31965}
31966#endif
31967
31968#ifdef __LITTLE_ENDIAN__
31969__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
31970  int16x8_t __ret;
31971  __ret = (int16x8_t)(__p0);
31972  return __ret;
31973}
31974#else
31975__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
31976  int16x8_t __ret;
31977  __ret = (int16x8_t)(__p0);
31978  return __ret;
31979}
31980#endif
31981
31982#ifdef __LITTLE_ENDIAN__
31983__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
31984  int16x8_t __ret;
31985  __ret = (int16x8_t)(__p0);
31986  return __ret;
31987}
31988#else
31989__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
31990  int16x8_t __ret;
31991  __ret = (int16x8_t)(__p0);
31992  return __ret;
31993}
31994#endif
31995
31996#ifdef __LITTLE_ENDIAN__
31997__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
31998  uint8x8_t __ret;
31999  __ret = (uint8x8_t)(__p0);
32000  return __ret;
32001}
32002#else
32003__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32004  uint8x8_t __ret;
32005  __ret = (uint8x8_t)(__p0);
32006  return __ret;
32007}
32008#endif
32009
32010#ifdef __LITTLE_ENDIAN__
32011__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32012  uint8x8_t __ret;
32013  __ret = (uint8x8_t)(__p0);
32014  return __ret;
32015}
32016#else
32017__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32018  uint8x8_t __ret;
32019  __ret = (uint8x8_t)(__p0);
32020  return __ret;
32021}
32022#endif
32023
32024#ifdef __LITTLE_ENDIAN__
32025__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32026  uint8x8_t __ret;
32027  __ret = (uint8x8_t)(__p0);
32028  return __ret;
32029}
32030#else
32031__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32032  uint8x8_t __ret;
32033  __ret = (uint8x8_t)(__p0);
32034  return __ret;
32035}
32036#endif
32037
32038#ifdef __LITTLE_ENDIAN__
32039__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32040  uint8x8_t __ret;
32041  __ret = (uint8x8_t)(__p0);
32042  return __ret;
32043}
32044#else
32045__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32046  uint8x8_t __ret;
32047  __ret = (uint8x8_t)(__p0);
32048  return __ret;
32049}
32050#endif
32051
32052#ifdef __LITTLE_ENDIAN__
32053__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32054  uint8x8_t __ret;
32055  __ret = (uint8x8_t)(__p0);
32056  return __ret;
32057}
32058#else
32059__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32060  uint8x8_t __ret;
32061  __ret = (uint8x8_t)(__p0);
32062  return __ret;
32063}
32064#endif
32065
32066#ifdef __LITTLE_ENDIAN__
32067__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32068  uint8x8_t __ret;
32069  __ret = (uint8x8_t)(__p0);
32070  return __ret;
32071}
32072#else
32073__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32074  uint8x8_t __ret;
32075  __ret = (uint8x8_t)(__p0);
32076  return __ret;
32077}
32078#endif
32079
32080#ifdef __LITTLE_ENDIAN__
32081__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32082  uint8x8_t __ret;
32083  __ret = (uint8x8_t)(__p0);
32084  return __ret;
32085}
32086#else
32087__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32088  uint8x8_t __ret;
32089  __ret = (uint8x8_t)(__p0);
32090  return __ret;
32091}
32092#endif
32093
32094#ifdef __LITTLE_ENDIAN__
32095__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32096  uint8x8_t __ret;
32097  __ret = (uint8x8_t)(__p0);
32098  return __ret;
32099}
32100#else
32101__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32102  uint8x8_t __ret;
32103  __ret = (uint8x8_t)(__p0);
32104  return __ret;
32105}
32106#endif
32107
32108#ifdef __LITTLE_ENDIAN__
32109__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32110  uint8x8_t __ret;
32111  __ret = (uint8x8_t)(__p0);
32112  return __ret;
32113}
32114#else
32115__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32116  uint8x8_t __ret;
32117  __ret = (uint8x8_t)(__p0);
32118  return __ret;
32119}
32120#endif
32121
32122#ifdef __LITTLE_ENDIAN__
32123__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32124  uint8x8_t __ret;
32125  __ret = (uint8x8_t)(__p0);
32126  return __ret;
32127}
32128#else
32129__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32130  uint8x8_t __ret;
32131  __ret = (uint8x8_t)(__p0);
32132  return __ret;
32133}
32134#endif
32135
32136#ifdef __LITTLE_ENDIAN__
32137__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32138  uint8x8_t __ret;
32139  __ret = (uint8x8_t)(__p0);
32140  return __ret;
32141}
32142#else
32143__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32144  uint8x8_t __ret;
32145  __ret = (uint8x8_t)(__p0);
32146  return __ret;
32147}
32148#endif
32149
32150#ifdef __LITTLE_ENDIAN__
32151__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32152  uint32x2_t __ret;
32153  __ret = (uint32x2_t)(__p0);
32154  return __ret;
32155}
32156#else
32157__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32158  uint32x2_t __ret;
32159  __ret = (uint32x2_t)(__p0);
32160  return __ret;
32161}
32162#endif
32163
32164#ifdef __LITTLE_ENDIAN__
32165__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32166  uint32x2_t __ret;
32167  __ret = (uint32x2_t)(__p0);
32168  return __ret;
32169}
32170#else
32171__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32172  uint32x2_t __ret;
32173  __ret = (uint32x2_t)(__p0);
32174  return __ret;
32175}
32176#endif
32177
32178#ifdef __LITTLE_ENDIAN__
32179__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32180  uint32x2_t __ret;
32181  __ret = (uint32x2_t)(__p0);
32182  return __ret;
32183}
32184#else
32185__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32186  uint32x2_t __ret;
32187  __ret = (uint32x2_t)(__p0);
32188  return __ret;
32189}
32190#endif
32191
32192#ifdef __LITTLE_ENDIAN__
32193__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32194  uint32x2_t __ret;
32195  __ret = (uint32x2_t)(__p0);
32196  return __ret;
32197}
32198#else
32199__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32200  uint32x2_t __ret;
32201  __ret = (uint32x2_t)(__p0);
32202  return __ret;
32203}
32204#endif
32205
32206#ifdef __LITTLE_ENDIAN__
32207__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32208  uint32x2_t __ret;
32209  __ret = (uint32x2_t)(__p0);
32210  return __ret;
32211}
32212#else
32213__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32214  uint32x2_t __ret;
32215  __ret = (uint32x2_t)(__p0);
32216  return __ret;
32217}
32218#endif
32219
32220#ifdef __LITTLE_ENDIAN__
32221__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32222  uint32x2_t __ret;
32223  __ret = (uint32x2_t)(__p0);
32224  return __ret;
32225}
32226#else
32227__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32228  uint32x2_t __ret;
32229  __ret = (uint32x2_t)(__p0);
32230  return __ret;
32231}
32232#endif
32233
32234#ifdef __LITTLE_ENDIAN__
32235__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32236  uint32x2_t __ret;
32237  __ret = (uint32x2_t)(__p0);
32238  return __ret;
32239}
32240#else
32241__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32242  uint32x2_t __ret;
32243  __ret = (uint32x2_t)(__p0);
32244  return __ret;
32245}
32246#endif
32247
32248#ifdef __LITTLE_ENDIAN__
32249__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32250  uint32x2_t __ret;
32251  __ret = (uint32x2_t)(__p0);
32252  return __ret;
32253}
32254#else
32255__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32256  uint32x2_t __ret;
32257  __ret = (uint32x2_t)(__p0);
32258  return __ret;
32259}
32260#endif
32261
32262#ifdef __LITTLE_ENDIAN__
32263__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32264  uint32x2_t __ret;
32265  __ret = (uint32x2_t)(__p0);
32266  return __ret;
32267}
32268#else
32269__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32270  uint32x2_t __ret;
32271  __ret = (uint32x2_t)(__p0);
32272  return __ret;
32273}
32274#endif
32275
32276#ifdef __LITTLE_ENDIAN__
32277__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32278  uint32x2_t __ret;
32279  __ret = (uint32x2_t)(__p0);
32280  return __ret;
32281}
32282#else
32283__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32284  uint32x2_t __ret;
32285  __ret = (uint32x2_t)(__p0);
32286  return __ret;
32287}
32288#endif
32289
32290#ifdef __LITTLE_ENDIAN__
32291__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32292  uint32x2_t __ret;
32293  __ret = (uint32x2_t)(__p0);
32294  return __ret;
32295}
32296#else
32297__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32298  uint32x2_t __ret;
32299  __ret = (uint32x2_t)(__p0);
32300  return __ret;
32301}
32302#endif
32303
32304#ifdef __LITTLE_ENDIAN__
32305__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32306  uint64x1_t __ret;
32307  __ret = (uint64x1_t)(__p0);
32308  return __ret;
32309}
32310#else
32311__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32312  uint64x1_t __ret;
32313  __ret = (uint64x1_t)(__p0);
32314  return __ret;
32315}
32316#endif
32317
32318#ifdef __LITTLE_ENDIAN__
32319__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32320  uint64x1_t __ret;
32321  __ret = (uint64x1_t)(__p0);
32322  return __ret;
32323}
32324#else
32325__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32326  uint64x1_t __ret;
32327  __ret = (uint64x1_t)(__p0);
32328  return __ret;
32329}
32330#endif
32331
32332#ifdef __LITTLE_ENDIAN__
32333__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32334  uint64x1_t __ret;
32335  __ret = (uint64x1_t)(__p0);
32336  return __ret;
32337}
32338#else
32339__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32340  uint64x1_t __ret;
32341  __ret = (uint64x1_t)(__p0);
32342  return __ret;
32343}
32344#endif
32345
32346#ifdef __LITTLE_ENDIAN__
32347__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32348  uint64x1_t __ret;
32349  __ret = (uint64x1_t)(__p0);
32350  return __ret;
32351}
32352#else
32353__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32354  uint64x1_t __ret;
32355  __ret = (uint64x1_t)(__p0);
32356  return __ret;
32357}
32358#endif
32359
32360#ifdef __LITTLE_ENDIAN__
32361__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32362  uint64x1_t __ret;
32363  __ret = (uint64x1_t)(__p0);
32364  return __ret;
32365}
32366#else
32367__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32368  uint64x1_t __ret;
32369  __ret = (uint64x1_t)(__p0);
32370  return __ret;
32371}
32372#endif
32373
32374#ifdef __LITTLE_ENDIAN__
32375__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32376  uint64x1_t __ret;
32377  __ret = (uint64x1_t)(__p0);
32378  return __ret;
32379}
32380#else
32381__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32382  uint64x1_t __ret;
32383  __ret = (uint64x1_t)(__p0);
32384  return __ret;
32385}
32386#endif
32387
32388#ifdef __LITTLE_ENDIAN__
32389__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32390  uint64x1_t __ret;
32391  __ret = (uint64x1_t)(__p0);
32392  return __ret;
32393}
32394#else
32395__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32396  uint64x1_t __ret;
32397  __ret = (uint64x1_t)(__p0);
32398  return __ret;
32399}
32400#endif
32401
32402#ifdef __LITTLE_ENDIAN__
32403__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32404  uint64x1_t __ret;
32405  __ret = (uint64x1_t)(__p0);
32406  return __ret;
32407}
32408#else
32409__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32410  uint64x1_t __ret;
32411  __ret = (uint64x1_t)(__p0);
32412  return __ret;
32413}
32414#endif
32415
32416#ifdef __LITTLE_ENDIAN__
32417__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32418  uint64x1_t __ret;
32419  __ret = (uint64x1_t)(__p0);
32420  return __ret;
32421}
32422#else
32423__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32424  uint64x1_t __ret;
32425  __ret = (uint64x1_t)(__p0);
32426  return __ret;
32427}
32428#endif
32429
32430#ifdef __LITTLE_ENDIAN__
32431__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32432  uint64x1_t __ret;
32433  __ret = (uint64x1_t)(__p0);
32434  return __ret;
32435}
32436#else
32437__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32438  uint64x1_t __ret;
32439  __ret = (uint64x1_t)(__p0);
32440  return __ret;
32441}
32442#endif
32443
32444#ifdef __LITTLE_ENDIAN__
32445__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32446  uint64x1_t __ret;
32447  __ret = (uint64x1_t)(__p0);
32448  return __ret;
32449}
32450#else
32451__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32452  uint64x1_t __ret;
32453  __ret = (uint64x1_t)(__p0);
32454  return __ret;
32455}
32456#endif
32457
32458#ifdef __LITTLE_ENDIAN__
32459__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32460  uint16x4_t __ret;
32461  __ret = (uint16x4_t)(__p0);
32462  return __ret;
32463}
32464#else
32465__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32466  uint16x4_t __ret;
32467  __ret = (uint16x4_t)(__p0);
32468  return __ret;
32469}
32470#endif
32471
32472#ifdef __LITTLE_ENDIAN__
32473__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32474  uint16x4_t __ret;
32475  __ret = (uint16x4_t)(__p0);
32476  return __ret;
32477}
32478#else
32479__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32480  uint16x4_t __ret;
32481  __ret = (uint16x4_t)(__p0);
32482  return __ret;
32483}
32484#endif
32485
32486#ifdef __LITTLE_ENDIAN__
32487__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32488  uint16x4_t __ret;
32489  __ret = (uint16x4_t)(__p0);
32490  return __ret;
32491}
32492#else
32493__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32494  uint16x4_t __ret;
32495  __ret = (uint16x4_t)(__p0);
32496  return __ret;
32497}
32498#endif
32499
32500#ifdef __LITTLE_ENDIAN__
32501__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32502  uint16x4_t __ret;
32503  __ret = (uint16x4_t)(__p0);
32504  return __ret;
32505}
32506#else
32507__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32508  uint16x4_t __ret;
32509  __ret = (uint16x4_t)(__p0);
32510  return __ret;
32511}
32512#endif
32513
32514#ifdef __LITTLE_ENDIAN__
32515__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32516  uint16x4_t __ret;
32517  __ret = (uint16x4_t)(__p0);
32518  return __ret;
32519}
32520#else
32521__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32522  uint16x4_t __ret;
32523  __ret = (uint16x4_t)(__p0);
32524  return __ret;
32525}
32526#endif
32527
32528#ifdef __LITTLE_ENDIAN__
32529__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32530  uint16x4_t __ret;
32531  __ret = (uint16x4_t)(__p0);
32532  return __ret;
32533}
32534#else
32535__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32536  uint16x4_t __ret;
32537  __ret = (uint16x4_t)(__p0);
32538  return __ret;
32539}
32540#endif
32541
32542#ifdef __LITTLE_ENDIAN__
32543__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32544  uint16x4_t __ret;
32545  __ret = (uint16x4_t)(__p0);
32546  return __ret;
32547}
32548#else
32549__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32550  uint16x4_t __ret;
32551  __ret = (uint16x4_t)(__p0);
32552  return __ret;
32553}
32554#endif
32555
32556#ifdef __LITTLE_ENDIAN__
32557__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32558  uint16x4_t __ret;
32559  __ret = (uint16x4_t)(__p0);
32560  return __ret;
32561}
32562#else
32563__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32564  uint16x4_t __ret;
32565  __ret = (uint16x4_t)(__p0);
32566  return __ret;
32567}
32568#endif
32569
32570#ifdef __LITTLE_ENDIAN__
32571__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32572  uint16x4_t __ret;
32573  __ret = (uint16x4_t)(__p0);
32574  return __ret;
32575}
32576#else
32577__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32578  uint16x4_t __ret;
32579  __ret = (uint16x4_t)(__p0);
32580  return __ret;
32581}
32582#endif
32583
32584#ifdef __LITTLE_ENDIAN__
32585__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32586  uint16x4_t __ret;
32587  __ret = (uint16x4_t)(__p0);
32588  return __ret;
32589}
32590#else
32591__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32592  uint16x4_t __ret;
32593  __ret = (uint16x4_t)(__p0);
32594  return __ret;
32595}
32596#endif
32597
32598#ifdef __LITTLE_ENDIAN__
32599__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32600  uint16x4_t __ret;
32601  __ret = (uint16x4_t)(__p0);
32602  return __ret;
32603}
32604#else
32605__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32606  uint16x4_t __ret;
32607  __ret = (uint16x4_t)(__p0);
32608  return __ret;
32609}
32610#endif
32611
32612#ifdef __LITTLE_ENDIAN__
32613__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32614  int8x8_t __ret;
32615  __ret = (int8x8_t)(__p0);
32616  return __ret;
32617}
32618#else
32619__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32620  int8x8_t __ret;
32621  __ret = (int8x8_t)(__p0);
32622  return __ret;
32623}
32624#endif
32625
32626#ifdef __LITTLE_ENDIAN__
32627__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32628  int8x8_t __ret;
32629  __ret = (int8x8_t)(__p0);
32630  return __ret;
32631}
32632#else
32633__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32634  int8x8_t __ret;
32635  __ret = (int8x8_t)(__p0);
32636  return __ret;
32637}
32638#endif
32639
32640#ifdef __LITTLE_ENDIAN__
32641__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32642  int8x8_t __ret;
32643  __ret = (int8x8_t)(__p0);
32644  return __ret;
32645}
32646#else
32647__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32648  int8x8_t __ret;
32649  __ret = (int8x8_t)(__p0);
32650  return __ret;
32651}
32652#endif
32653
32654#ifdef __LITTLE_ENDIAN__
32655__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32656  int8x8_t __ret;
32657  __ret = (int8x8_t)(__p0);
32658  return __ret;
32659}
32660#else
32661__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32662  int8x8_t __ret;
32663  __ret = (int8x8_t)(__p0);
32664  return __ret;
32665}
32666#endif
32667
32668#ifdef __LITTLE_ENDIAN__
32669__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32670  int8x8_t __ret;
32671  __ret = (int8x8_t)(__p0);
32672  return __ret;
32673}
32674#else
32675__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32676  int8x8_t __ret;
32677  __ret = (int8x8_t)(__p0);
32678  return __ret;
32679}
32680#endif
32681
32682#ifdef __LITTLE_ENDIAN__
32683__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32684  int8x8_t __ret;
32685  __ret = (int8x8_t)(__p0);
32686  return __ret;
32687}
32688#else
32689__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32690  int8x8_t __ret;
32691  __ret = (int8x8_t)(__p0);
32692  return __ret;
32693}
32694#endif
32695
32696#ifdef __LITTLE_ENDIAN__
32697__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32698  int8x8_t __ret;
32699  __ret = (int8x8_t)(__p0);
32700  return __ret;
32701}
32702#else
32703__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32704  int8x8_t __ret;
32705  __ret = (int8x8_t)(__p0);
32706  return __ret;
32707}
32708#endif
32709
32710#ifdef __LITTLE_ENDIAN__
32711__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32712  int8x8_t __ret;
32713  __ret = (int8x8_t)(__p0);
32714  return __ret;
32715}
32716#else
32717__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32718  int8x8_t __ret;
32719  __ret = (int8x8_t)(__p0);
32720  return __ret;
32721}
32722#endif
32723
32724#ifdef __LITTLE_ENDIAN__
32725__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32726  int8x8_t __ret;
32727  __ret = (int8x8_t)(__p0);
32728  return __ret;
32729}
32730#else
32731__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32732  int8x8_t __ret;
32733  __ret = (int8x8_t)(__p0);
32734  return __ret;
32735}
32736#endif
32737
32738#ifdef __LITTLE_ENDIAN__
32739__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32740  int8x8_t __ret;
32741  __ret = (int8x8_t)(__p0);
32742  return __ret;
32743}
32744#else
32745__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32746  int8x8_t __ret;
32747  __ret = (int8x8_t)(__p0);
32748  return __ret;
32749}
32750#endif
32751
32752#ifdef __LITTLE_ENDIAN__
32753__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32754  int8x8_t __ret;
32755  __ret = (int8x8_t)(__p0);
32756  return __ret;
32757}
32758#else
32759__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32760  int8x8_t __ret;
32761  __ret = (int8x8_t)(__p0);
32762  return __ret;
32763}
32764#endif
32765
32766#ifdef __LITTLE_ENDIAN__
32767__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32768  float32x2_t __ret;
32769  __ret = (float32x2_t)(__p0);
32770  return __ret;
32771}
32772#else
32773__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32774  float32x2_t __ret;
32775  __ret = (float32x2_t)(__p0);
32776  return __ret;
32777}
32778#endif
32779
32780#ifdef __LITTLE_ENDIAN__
32781__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32782  float32x2_t __ret;
32783  __ret = (float32x2_t)(__p0);
32784  return __ret;
32785}
32786#else
32787__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32788  float32x2_t __ret;
32789  __ret = (float32x2_t)(__p0);
32790  return __ret;
32791}
32792#endif
32793
32794#ifdef __LITTLE_ENDIAN__
32795__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32796  float32x2_t __ret;
32797  __ret = (float32x2_t)(__p0);
32798  return __ret;
32799}
32800#else
32801__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32802  float32x2_t __ret;
32803  __ret = (float32x2_t)(__p0);
32804  return __ret;
32805}
32806#endif
32807
32808#ifdef __LITTLE_ENDIAN__
32809__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32810  float32x2_t __ret;
32811  __ret = (float32x2_t)(__p0);
32812  return __ret;
32813}
32814#else
32815__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32816  float32x2_t __ret;
32817  __ret = (float32x2_t)(__p0);
32818  return __ret;
32819}
32820#endif
32821
32822#ifdef __LITTLE_ENDIAN__
32823__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32824  float32x2_t __ret;
32825  __ret = (float32x2_t)(__p0);
32826  return __ret;
32827}
32828#else
32829__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32830  float32x2_t __ret;
32831  __ret = (float32x2_t)(__p0);
32832  return __ret;
32833}
32834#endif
32835
32836#ifdef __LITTLE_ENDIAN__
32837__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32838  float32x2_t __ret;
32839  __ret = (float32x2_t)(__p0);
32840  return __ret;
32841}
32842#else
32843__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32844  float32x2_t __ret;
32845  __ret = (float32x2_t)(__p0);
32846  return __ret;
32847}
32848#endif
32849
32850#ifdef __LITTLE_ENDIAN__
32851__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32852  float32x2_t __ret;
32853  __ret = (float32x2_t)(__p0);
32854  return __ret;
32855}
32856#else
32857__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32858  float32x2_t __ret;
32859  __ret = (float32x2_t)(__p0);
32860  return __ret;
32861}
32862#endif
32863
32864#ifdef __LITTLE_ENDIAN__
32865__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32866  float32x2_t __ret;
32867  __ret = (float32x2_t)(__p0);
32868  return __ret;
32869}
32870#else
32871__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32872  float32x2_t __ret;
32873  __ret = (float32x2_t)(__p0);
32874  return __ret;
32875}
32876#endif
32877
32878#ifdef __LITTLE_ENDIAN__
32879__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32880  float32x2_t __ret;
32881  __ret = (float32x2_t)(__p0);
32882  return __ret;
32883}
32884#else
32885__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32886  float32x2_t __ret;
32887  __ret = (float32x2_t)(__p0);
32888  return __ret;
32889}
32890#endif
32891
32892#ifdef __LITTLE_ENDIAN__
32893__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32894  float32x2_t __ret;
32895  __ret = (float32x2_t)(__p0);
32896  return __ret;
32897}
32898#else
32899__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32900  float32x2_t __ret;
32901  __ret = (float32x2_t)(__p0);
32902  return __ret;
32903}
32904#endif
32905
32906#ifdef __LITTLE_ENDIAN__
32907__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32908  float32x2_t __ret;
32909  __ret = (float32x2_t)(__p0);
32910  return __ret;
32911}
32912#else
32913__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32914  float32x2_t __ret;
32915  __ret = (float32x2_t)(__p0);
32916  return __ret;
32917}
32918#endif
32919
32920#ifdef __LITTLE_ENDIAN__
32921__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
32922  float16x4_t __ret;
32923  __ret = (float16x4_t)(__p0);
32924  return __ret;
32925}
32926#else
32927__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
32928  float16x4_t __ret;
32929  __ret = (float16x4_t)(__p0);
32930  return __ret;
32931}
32932#endif
32933
32934#ifdef __LITTLE_ENDIAN__
32935__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
32936  float16x4_t __ret;
32937  __ret = (float16x4_t)(__p0);
32938  return __ret;
32939}
32940#else
32941__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
32942  float16x4_t __ret;
32943  __ret = (float16x4_t)(__p0);
32944  return __ret;
32945}
32946#endif
32947
32948#ifdef __LITTLE_ENDIAN__
32949__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
32950  float16x4_t __ret;
32951  __ret = (float16x4_t)(__p0);
32952  return __ret;
32953}
32954#else
32955__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
32956  float16x4_t __ret;
32957  __ret = (float16x4_t)(__p0);
32958  return __ret;
32959}
32960#endif
32961
32962#ifdef __LITTLE_ENDIAN__
32963__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
32964  float16x4_t __ret;
32965  __ret = (float16x4_t)(__p0);
32966  return __ret;
32967}
32968#else
32969__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
32970  float16x4_t __ret;
32971  __ret = (float16x4_t)(__p0);
32972  return __ret;
32973}
32974#endif
32975
32976#ifdef __LITTLE_ENDIAN__
32977__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
32978  float16x4_t __ret;
32979  __ret = (float16x4_t)(__p0);
32980  return __ret;
32981}
32982#else
32983__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
32984  float16x4_t __ret;
32985  __ret = (float16x4_t)(__p0);
32986  return __ret;
32987}
32988#endif
32989
32990#ifdef __LITTLE_ENDIAN__
32991__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
32992  float16x4_t __ret;
32993  __ret = (float16x4_t)(__p0);
32994  return __ret;
32995}
32996#else
32997__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
32998  float16x4_t __ret;
32999  __ret = (float16x4_t)(__p0);
33000  return __ret;
33001}
33002#endif
33003
33004#ifdef __LITTLE_ENDIAN__
33005__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33006  float16x4_t __ret;
33007  __ret = (float16x4_t)(__p0);
33008  return __ret;
33009}
33010#else
33011__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33012  float16x4_t __ret;
33013  __ret = (float16x4_t)(__p0);
33014  return __ret;
33015}
33016#endif
33017
33018#ifdef __LITTLE_ENDIAN__
33019__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33020  float16x4_t __ret;
33021  __ret = (float16x4_t)(__p0);
33022  return __ret;
33023}
33024#else
33025__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33026  float16x4_t __ret;
33027  __ret = (float16x4_t)(__p0);
33028  return __ret;
33029}
33030#endif
33031
33032#ifdef __LITTLE_ENDIAN__
33033__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33034  float16x4_t __ret;
33035  __ret = (float16x4_t)(__p0);
33036  return __ret;
33037}
33038#else
33039__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33040  float16x4_t __ret;
33041  __ret = (float16x4_t)(__p0);
33042  return __ret;
33043}
33044#endif
33045
33046#ifdef __LITTLE_ENDIAN__
33047__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33048  float16x4_t __ret;
33049  __ret = (float16x4_t)(__p0);
33050  return __ret;
33051}
33052#else
33053__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33054  float16x4_t __ret;
33055  __ret = (float16x4_t)(__p0);
33056  return __ret;
33057}
33058#endif
33059
33060#ifdef __LITTLE_ENDIAN__
33061__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33062  float16x4_t __ret;
33063  __ret = (float16x4_t)(__p0);
33064  return __ret;
33065}
33066#else
33067__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33068  float16x4_t __ret;
33069  __ret = (float16x4_t)(__p0);
33070  return __ret;
33071}
33072#endif
33073
33074#ifdef __LITTLE_ENDIAN__
33075__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33076  int32x2_t __ret;
33077  __ret = (int32x2_t)(__p0);
33078  return __ret;
33079}
33080#else
33081__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33082  int32x2_t __ret;
33083  __ret = (int32x2_t)(__p0);
33084  return __ret;
33085}
33086#endif
33087
33088#ifdef __LITTLE_ENDIAN__
33089__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33090  int32x2_t __ret;
33091  __ret = (int32x2_t)(__p0);
33092  return __ret;
33093}
33094#else
33095__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33096  int32x2_t __ret;
33097  __ret = (int32x2_t)(__p0);
33098  return __ret;
33099}
33100#endif
33101
33102#ifdef __LITTLE_ENDIAN__
33103__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33104  int32x2_t __ret;
33105  __ret = (int32x2_t)(__p0);
33106  return __ret;
33107}
33108#else
33109__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33110  int32x2_t __ret;
33111  __ret = (int32x2_t)(__p0);
33112  return __ret;
33113}
33114#endif
33115
33116#ifdef __LITTLE_ENDIAN__
33117__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33118  int32x2_t __ret;
33119  __ret = (int32x2_t)(__p0);
33120  return __ret;
33121}
33122#else
33123__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33124  int32x2_t __ret;
33125  __ret = (int32x2_t)(__p0);
33126  return __ret;
33127}
33128#endif
33129
33130#ifdef __LITTLE_ENDIAN__
33131__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33132  int32x2_t __ret;
33133  __ret = (int32x2_t)(__p0);
33134  return __ret;
33135}
33136#else
33137__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33138  int32x2_t __ret;
33139  __ret = (int32x2_t)(__p0);
33140  return __ret;
33141}
33142#endif
33143
33144#ifdef __LITTLE_ENDIAN__
33145__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33146  int32x2_t __ret;
33147  __ret = (int32x2_t)(__p0);
33148  return __ret;
33149}
33150#else
33151__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33152  int32x2_t __ret;
33153  __ret = (int32x2_t)(__p0);
33154  return __ret;
33155}
33156#endif
33157
33158#ifdef __LITTLE_ENDIAN__
33159__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33160  int32x2_t __ret;
33161  __ret = (int32x2_t)(__p0);
33162  return __ret;
33163}
33164#else
33165__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33166  int32x2_t __ret;
33167  __ret = (int32x2_t)(__p0);
33168  return __ret;
33169}
33170#endif
33171
33172#ifdef __LITTLE_ENDIAN__
33173__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33174  int32x2_t __ret;
33175  __ret = (int32x2_t)(__p0);
33176  return __ret;
33177}
33178#else
33179__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33180  int32x2_t __ret;
33181  __ret = (int32x2_t)(__p0);
33182  return __ret;
33183}
33184#endif
33185
33186#ifdef __LITTLE_ENDIAN__
33187__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33188  int32x2_t __ret;
33189  __ret = (int32x2_t)(__p0);
33190  return __ret;
33191}
33192#else
33193__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33194  int32x2_t __ret;
33195  __ret = (int32x2_t)(__p0);
33196  return __ret;
33197}
33198#endif
33199
33200#ifdef __LITTLE_ENDIAN__
33201__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33202  int32x2_t __ret;
33203  __ret = (int32x2_t)(__p0);
33204  return __ret;
33205}
33206#else
33207__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33208  int32x2_t __ret;
33209  __ret = (int32x2_t)(__p0);
33210  return __ret;
33211}
33212#endif
33213
33214#ifdef __LITTLE_ENDIAN__
33215__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33216  int32x2_t __ret;
33217  __ret = (int32x2_t)(__p0);
33218  return __ret;
33219}
33220#else
33221__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33222  int32x2_t __ret;
33223  __ret = (int32x2_t)(__p0);
33224  return __ret;
33225}
33226#endif
33227
33228#ifdef __LITTLE_ENDIAN__
33229__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33230  int64x1_t __ret;
33231  __ret = (int64x1_t)(__p0);
33232  return __ret;
33233}
33234#else
33235__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33236  int64x1_t __ret;
33237  __ret = (int64x1_t)(__p0);
33238  return __ret;
33239}
33240#endif
33241
33242#ifdef __LITTLE_ENDIAN__
33243__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33244  int64x1_t __ret;
33245  __ret = (int64x1_t)(__p0);
33246  return __ret;
33247}
33248#else
33249__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33250  int64x1_t __ret;
33251  __ret = (int64x1_t)(__p0);
33252  return __ret;
33253}
33254#endif
33255
33256#ifdef __LITTLE_ENDIAN__
33257__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33258  int64x1_t __ret;
33259  __ret = (int64x1_t)(__p0);
33260  return __ret;
33261}
33262#else
33263__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33264  int64x1_t __ret;
33265  __ret = (int64x1_t)(__p0);
33266  return __ret;
33267}
33268#endif
33269
33270#ifdef __LITTLE_ENDIAN__
33271__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33272  int64x1_t __ret;
33273  __ret = (int64x1_t)(__p0);
33274  return __ret;
33275}
33276#else
33277__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33278  int64x1_t __ret;
33279  __ret = (int64x1_t)(__p0);
33280  return __ret;
33281}
33282#endif
33283
33284#ifdef __LITTLE_ENDIAN__
33285__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33286  int64x1_t __ret;
33287  __ret = (int64x1_t)(__p0);
33288  return __ret;
33289}
33290#else
33291__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33292  int64x1_t __ret;
33293  __ret = (int64x1_t)(__p0);
33294  return __ret;
33295}
33296#endif
33297
33298#ifdef __LITTLE_ENDIAN__
33299__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33300  int64x1_t __ret;
33301  __ret = (int64x1_t)(__p0);
33302  return __ret;
33303}
33304#else
33305__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33306  int64x1_t __ret;
33307  __ret = (int64x1_t)(__p0);
33308  return __ret;
33309}
33310#endif
33311
33312#ifdef __LITTLE_ENDIAN__
33313__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33314  int64x1_t __ret;
33315  __ret = (int64x1_t)(__p0);
33316  return __ret;
33317}
33318#else
33319__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33320  int64x1_t __ret;
33321  __ret = (int64x1_t)(__p0);
33322  return __ret;
33323}
33324#endif
33325
33326#ifdef __LITTLE_ENDIAN__
33327__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33328  int64x1_t __ret;
33329  __ret = (int64x1_t)(__p0);
33330  return __ret;
33331}
33332#else
33333__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33334  int64x1_t __ret;
33335  __ret = (int64x1_t)(__p0);
33336  return __ret;
33337}
33338#endif
33339
33340#ifdef __LITTLE_ENDIAN__
33341__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33342  int64x1_t __ret;
33343  __ret = (int64x1_t)(__p0);
33344  return __ret;
33345}
33346#else
33347__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33348  int64x1_t __ret;
33349  __ret = (int64x1_t)(__p0);
33350  return __ret;
33351}
33352#endif
33353
33354#ifdef __LITTLE_ENDIAN__
33355__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33356  int64x1_t __ret;
33357  __ret = (int64x1_t)(__p0);
33358  return __ret;
33359}
33360#else
33361__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33362  int64x1_t __ret;
33363  __ret = (int64x1_t)(__p0);
33364  return __ret;
33365}
33366#endif
33367
33368#ifdef __LITTLE_ENDIAN__
33369__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33370  int64x1_t __ret;
33371  __ret = (int64x1_t)(__p0);
33372  return __ret;
33373}
33374#else
33375__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33376  int64x1_t __ret;
33377  __ret = (int64x1_t)(__p0);
33378  return __ret;
33379}
33380#endif
33381
33382#ifdef __LITTLE_ENDIAN__
33383__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33384  int16x4_t __ret;
33385  __ret = (int16x4_t)(__p0);
33386  return __ret;
33387}
33388#else
33389__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33390  int16x4_t __ret;
33391  __ret = (int16x4_t)(__p0);
33392  return __ret;
33393}
33394#endif
33395
33396#ifdef __LITTLE_ENDIAN__
33397__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33398  int16x4_t __ret;
33399  __ret = (int16x4_t)(__p0);
33400  return __ret;
33401}
33402#else
33403__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33404  int16x4_t __ret;
33405  __ret = (int16x4_t)(__p0);
33406  return __ret;
33407}
33408#endif
33409
33410#ifdef __LITTLE_ENDIAN__
33411__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33412  int16x4_t __ret;
33413  __ret = (int16x4_t)(__p0);
33414  return __ret;
33415}
33416#else
33417__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33418  int16x4_t __ret;
33419  __ret = (int16x4_t)(__p0);
33420  return __ret;
33421}
33422#endif
33423
33424#ifdef __LITTLE_ENDIAN__
33425__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33426  int16x4_t __ret;
33427  __ret = (int16x4_t)(__p0);
33428  return __ret;
33429}
33430#else
33431__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33432  int16x4_t __ret;
33433  __ret = (int16x4_t)(__p0);
33434  return __ret;
33435}
33436#endif
33437
33438#ifdef __LITTLE_ENDIAN__
33439__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33440  int16x4_t __ret;
33441  __ret = (int16x4_t)(__p0);
33442  return __ret;
33443}
33444#else
33445__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33446  int16x4_t __ret;
33447  __ret = (int16x4_t)(__p0);
33448  return __ret;
33449}
33450#endif
33451
33452#ifdef __LITTLE_ENDIAN__
33453__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33454  int16x4_t __ret;
33455  __ret = (int16x4_t)(__p0);
33456  return __ret;
33457}
33458#else
33459__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33460  int16x4_t __ret;
33461  __ret = (int16x4_t)(__p0);
33462  return __ret;
33463}
33464#endif
33465
33466#ifdef __LITTLE_ENDIAN__
33467__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33468  int16x4_t __ret;
33469  __ret = (int16x4_t)(__p0);
33470  return __ret;
33471}
33472#else
33473__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33474  int16x4_t __ret;
33475  __ret = (int16x4_t)(__p0);
33476  return __ret;
33477}
33478#endif
33479
33480#ifdef __LITTLE_ENDIAN__
33481__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33482  int16x4_t __ret;
33483  __ret = (int16x4_t)(__p0);
33484  return __ret;
33485}
33486#else
33487__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33488  int16x4_t __ret;
33489  __ret = (int16x4_t)(__p0);
33490  return __ret;
33491}
33492#endif
33493
33494#ifdef __LITTLE_ENDIAN__
33495__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33496  int16x4_t __ret;
33497  __ret = (int16x4_t)(__p0);
33498  return __ret;
33499}
33500#else
33501__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33502  int16x4_t __ret;
33503  __ret = (int16x4_t)(__p0);
33504  return __ret;
33505}
33506#endif
33507
33508#ifdef __LITTLE_ENDIAN__
33509__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33510  int16x4_t __ret;
33511  __ret = (int16x4_t)(__p0);
33512  return __ret;
33513}
33514#else
33515__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33516  int16x4_t __ret;
33517  __ret = (int16x4_t)(__p0);
33518  return __ret;
33519}
33520#endif
33521
33522#ifdef __LITTLE_ENDIAN__
33523__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33524  int16x4_t __ret;
33525  __ret = (int16x4_t)(__p0);
33526  return __ret;
33527}
33528#else
33529__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33530  int16x4_t __ret;
33531  __ret = (int16x4_t)(__p0);
33532  return __ret;
33533}
33534#endif
33535
33536#endif
33537#if (__ARM_FP & 2)
33538#ifdef __LITTLE_ENDIAN__
33539__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
33540  float16x4_t __ret;
33541  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
33542  return __ret;
33543}
33544#else
33545__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
33546  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33547  float16x4_t __ret;
33548  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 8);
33549  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33550  return __ret;
33551}
33552__ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
33553  float16x4_t __ret;
33554  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
33555  return __ret;
33556}
33557#endif
33558
33559#ifdef __LITTLE_ENDIAN__
33560__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
33561  float32x4_t __ret;
33562  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
33563  return __ret;
33564}
33565#else
33566__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
33567  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33568  float32x4_t __ret;
33569  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 41);
33570  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33571  return __ret;
33572}
33573__ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
33574  float32x4_t __ret;
33575  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
33576  return __ret;
33577}
33578#endif
33579
33580#endif
33581#if __ARM_ARCH >= 8
33582#ifdef __LITTLE_ENDIAN__
33583__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33584  int32x4_t __ret;
33585  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
33586  return __ret;
33587}
33588#else
33589__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33590  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33591  int32x4_t __ret;
33592  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
33593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33594  return __ret;
33595}
33596#endif
33597
33598#ifdef __LITTLE_ENDIAN__
33599__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33600  int32x2_t __ret;
33601  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
33602  return __ret;
33603}
33604#else
33605__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33606  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33607  int32x2_t __ret;
33608  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
33609  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33610  return __ret;
33611}
33612#endif
33613
33614#ifdef __LITTLE_ENDIAN__
33615__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33616  uint32x4_t __ret;
33617  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
33618  return __ret;
33619}
33620#else
33621__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33622  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33623  uint32x4_t __ret;
33624  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
33625  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33626  return __ret;
33627}
33628#endif
33629
33630#ifdef __LITTLE_ENDIAN__
33631__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33632  uint32x2_t __ret;
33633  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
33634  return __ret;
33635}
33636#else
33637__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33638  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33639  uint32x2_t __ret;
33640  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
33641  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33642  return __ret;
33643}
33644#endif
33645
33646#ifdef __LITTLE_ENDIAN__
33647__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33648  int32x4_t __ret;
33649  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
33650  return __ret;
33651}
33652#else
33653__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33654  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33655  int32x4_t __ret;
33656  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
33657  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33658  return __ret;
33659}
33660#endif
33661
33662#ifdef __LITTLE_ENDIAN__
33663__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33664  int32x2_t __ret;
33665  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
33666  return __ret;
33667}
33668#else
33669__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33670  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33671  int32x2_t __ret;
33672  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
33673  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33674  return __ret;
33675}
33676#endif
33677
33678#ifdef __LITTLE_ENDIAN__
33679__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33680  uint32x4_t __ret;
33681  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
33682  return __ret;
33683}
33684#else
33685__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33686  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33687  uint32x4_t __ret;
33688  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
33689  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33690  return __ret;
33691}
33692#endif
33693
33694#ifdef __LITTLE_ENDIAN__
33695__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33696  uint32x2_t __ret;
33697  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
33698  return __ret;
33699}
33700#else
33701__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33702  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33703  uint32x2_t __ret;
33704  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
33705  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33706  return __ret;
33707}
33708#endif
33709
33710#ifdef __LITTLE_ENDIAN__
33711__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33712  int32x4_t __ret;
33713  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
33714  return __ret;
33715}
33716#else
33717__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33718  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33719  int32x4_t __ret;
33720  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
33721  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33722  return __ret;
33723}
33724#endif
33725
33726#ifdef __LITTLE_ENDIAN__
33727__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33728  int32x2_t __ret;
33729  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
33730  return __ret;
33731}
33732#else
33733__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33734  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33735  int32x2_t __ret;
33736  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
33737  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33738  return __ret;
33739}
33740#endif
33741
33742#ifdef __LITTLE_ENDIAN__
33743__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33744  uint32x4_t __ret;
33745  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
33746  return __ret;
33747}
33748#else
33749__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33750  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33751  uint32x4_t __ret;
33752  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
33753  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33754  return __ret;
33755}
33756#endif
33757
33758#ifdef __LITTLE_ENDIAN__
33759__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33760  uint32x2_t __ret;
33761  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
33762  return __ret;
33763}
33764#else
33765__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33766  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33767  uint32x2_t __ret;
33768  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
33769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33770  return __ret;
33771}
33772#endif
33773
33774#ifdef __LITTLE_ENDIAN__
33775__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33776  int32x4_t __ret;
33777  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
33778  return __ret;
33779}
33780#else
33781__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33782  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33783  int32x4_t __ret;
33784  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
33785  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33786  return __ret;
33787}
33788#endif
33789
33790#ifdef __LITTLE_ENDIAN__
33791__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33792  int32x2_t __ret;
33793  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
33794  return __ret;
33795}
33796#else
33797__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33798  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33799  int32x2_t __ret;
33800  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
33801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33802  return __ret;
33803}
33804#endif
33805
33806#ifdef __LITTLE_ENDIAN__
33807__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33808  uint32x4_t __ret;
33809  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
33810  return __ret;
33811}
33812#else
33813__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33814  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33815  uint32x4_t __ret;
33816  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
33817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33818  return __ret;
33819}
33820#endif
33821
33822#ifdef __LITTLE_ENDIAN__
33823__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33824  uint32x2_t __ret;
33825  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
33826  return __ret;
33827}
33828#else
33829__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33830  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33831  uint32x2_t __ret;
33832  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
33833  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33834  return __ret;
33835}
33836#endif
33837
33838#endif
33839#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
33840#ifdef __LITTLE_ENDIAN__
33841__ai float32x4_t vrndq_f32(float32x4_t __p0) {
33842  float32x4_t __ret;
33843  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
33844  return __ret;
33845}
33846#else
33847__ai float32x4_t vrndq_f32(float32x4_t __p0) {
33848  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33849  float32x4_t __ret;
33850  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
33851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33852  return __ret;
33853}
33854#endif
33855
33856#ifdef __LITTLE_ENDIAN__
33857__ai float32x2_t vrnd_f32(float32x2_t __p0) {
33858  float32x2_t __ret;
33859  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
33860  return __ret;
33861}
33862#else
33863__ai float32x2_t vrnd_f32(float32x2_t __p0) {
33864  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33865  float32x2_t __ret;
33866  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
33867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33868  return __ret;
33869}
33870#endif
33871
33872#ifdef __LITTLE_ENDIAN__
33873__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33874  float32x4_t __ret;
33875  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
33876  return __ret;
33877}
33878#else
33879__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33880  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33881  float32x4_t __ret;
33882  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
33883  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33884  return __ret;
33885}
33886#endif
33887
33888#ifdef __LITTLE_ENDIAN__
33889__ai float32x2_t vrnda_f32(float32x2_t __p0) {
33890  float32x2_t __ret;
33891  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
33892  return __ret;
33893}
33894#else
33895__ai float32x2_t vrnda_f32(float32x2_t __p0) {
33896  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33897  float32x2_t __ret;
33898  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
33899  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33900  return __ret;
33901}
33902#endif
33903
33904#ifdef __LITTLE_ENDIAN__
33905__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33906  float32x4_t __ret;
33907  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
33908  return __ret;
33909}
33910#else
33911__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33912  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33913  float32x4_t __ret;
33914  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
33915  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33916  return __ret;
33917}
33918#endif
33919
33920#ifdef __LITTLE_ENDIAN__
33921__ai float32x2_t vrndm_f32(float32x2_t __p0) {
33922  float32x2_t __ret;
33923  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
33924  return __ret;
33925}
33926#else
33927__ai float32x2_t vrndm_f32(float32x2_t __p0) {
33928  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33929  float32x2_t __ret;
33930  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
33931  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33932  return __ret;
33933}
33934#endif
33935
33936#ifdef __LITTLE_ENDIAN__
33937__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33938  float32x4_t __ret;
33939  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
33940  return __ret;
33941}
33942#else
33943__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33944  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33945  float32x4_t __ret;
33946  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
33947  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33948  return __ret;
33949}
33950#endif
33951
33952#ifdef __LITTLE_ENDIAN__
33953__ai float32x2_t vrndn_f32(float32x2_t __p0) {
33954  float32x2_t __ret;
33955  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
33956  return __ret;
33957}
33958#else
33959__ai float32x2_t vrndn_f32(float32x2_t __p0) {
33960  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33961  float32x2_t __ret;
33962  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
33963  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33964  return __ret;
33965}
33966#endif
33967
33968#ifdef __LITTLE_ENDIAN__
33969__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
33970  float32x4_t __ret;
33971  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
33972  return __ret;
33973}
33974#else
33975__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
33976  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33977  float32x4_t __ret;
33978  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
33979  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33980  return __ret;
33981}
33982#endif
33983
33984#ifdef __LITTLE_ENDIAN__
33985__ai float32x2_t vrndp_f32(float32x2_t __p0) {
33986  float32x2_t __ret;
33987  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
33988  return __ret;
33989}
33990#else
33991__ai float32x2_t vrndp_f32(float32x2_t __p0) {
33992  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33993  float32x2_t __ret;
33994  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
33995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33996  return __ret;
33997}
33998#endif
33999
34000#ifdef __LITTLE_ENDIAN__
34001__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34002  float32x4_t __ret;
34003  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
34004  return __ret;
34005}
34006#else
34007__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34008  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34009  float32x4_t __ret;
34010  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
34011  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34012  return __ret;
34013}
34014#endif
34015
34016#ifdef __LITTLE_ENDIAN__
34017__ai float32x2_t vrndx_f32(float32x2_t __p0) {
34018  float32x2_t __ret;
34019  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
34020  return __ret;
34021}
34022#else
34023__ai float32x2_t vrndx_f32(float32x2_t __p0) {
34024  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34025  float32x2_t __ret;
34026  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
34027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34028  return __ret;
34029}
34030#endif
34031
34032#endif
34033#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
34034#ifdef __LITTLE_ENDIAN__
34035__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34036  float32x4_t __ret;
34037  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34038  return __ret;
34039}
34040#else
34041__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34042  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34043  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34044  float32x4_t __ret;
34045  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34046  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34047  return __ret;
34048}
34049#endif
34050
34051#ifdef __LITTLE_ENDIAN__
34052__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34053  float32x2_t __ret;
34054  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34055  return __ret;
34056}
34057#else
34058__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34059  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34060  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34061  float32x2_t __ret;
34062  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34063  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34064  return __ret;
34065}
34066#endif
34067
34068#ifdef __LITTLE_ENDIAN__
34069__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34070  float32x4_t __ret;
34071  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34072  return __ret;
34073}
34074#else
34075__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34076  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34077  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34078  float32x4_t __ret;
34079  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34080  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34081  return __ret;
34082}
34083#endif
34084
34085#ifdef __LITTLE_ENDIAN__
34086__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34087  float32x2_t __ret;
34088  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34089  return __ret;
34090}
34091#else
34092__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34093  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34094  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34095  float32x2_t __ret;
34096  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34097  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34098  return __ret;
34099}
34100#endif
34101
34102#endif
34103#if __ARM_ARCH >= 8 && defined(__aarch64__)
34104#ifdef __LITTLE_ENDIAN__
34105__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34106  int64x2_t __ret;
34107  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
34108  return __ret;
34109}
34110#else
34111__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34112  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34113  int64x2_t __ret;
34114  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
34115  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34116  return __ret;
34117}
34118#endif
34119
34120#ifdef __LITTLE_ENDIAN__
34121__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34122  int64x1_t __ret;
34123  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34124  return __ret;
34125}
34126#else
34127__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34128  int64x1_t __ret;
34129  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34130  return __ret;
34131}
34132#endif
34133
34134#ifdef __LITTLE_ENDIAN__
34135__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34136  uint64x2_t __ret;
34137  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
34138  return __ret;
34139}
34140#else
34141__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34142  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34143  uint64x2_t __ret;
34144  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
34145  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34146  return __ret;
34147}
34148#endif
34149
34150#ifdef __LITTLE_ENDIAN__
34151__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34152  uint64x1_t __ret;
34153  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34154  return __ret;
34155}
34156#else
34157__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34158  uint64x1_t __ret;
34159  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34160  return __ret;
34161}
34162#endif
34163
34164#ifdef __LITTLE_ENDIAN__
34165__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34166  int64x2_t __ret;
34167  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
34168  return __ret;
34169}
34170#else
34171__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34172  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34173  int64x2_t __ret;
34174  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
34175  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34176  return __ret;
34177}
34178#endif
34179
34180#ifdef __LITTLE_ENDIAN__
34181__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34182  int64x1_t __ret;
34183  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34184  return __ret;
34185}
34186#else
34187__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34188  int64x1_t __ret;
34189  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34190  return __ret;
34191}
34192#endif
34193
34194#ifdef __LITTLE_ENDIAN__
34195__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34196  uint64x2_t __ret;
34197  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
34198  return __ret;
34199}
34200#else
34201__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34202  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34203  uint64x2_t __ret;
34204  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
34205  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34206  return __ret;
34207}
34208#endif
34209
34210#ifdef __LITTLE_ENDIAN__
34211__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34212  uint64x1_t __ret;
34213  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34214  return __ret;
34215}
34216#else
34217__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34218  uint64x1_t __ret;
34219  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34220  return __ret;
34221}
34222#endif
34223
34224#ifdef __LITTLE_ENDIAN__
34225__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34226  int64x2_t __ret;
34227  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
34228  return __ret;
34229}
34230#else
34231__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34232  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34233  int64x2_t __ret;
34234  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
34235  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34236  return __ret;
34237}
34238#endif
34239
34240#ifdef __LITTLE_ENDIAN__
34241__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34242  int64x1_t __ret;
34243  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34244  return __ret;
34245}
34246#else
34247__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34248  int64x1_t __ret;
34249  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34250  return __ret;
34251}
34252#endif
34253
34254#ifdef __LITTLE_ENDIAN__
34255__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34256  uint64x2_t __ret;
34257  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
34258  return __ret;
34259}
34260#else
34261__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34262  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34263  uint64x2_t __ret;
34264  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
34265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34266  return __ret;
34267}
34268#endif
34269
34270#ifdef __LITTLE_ENDIAN__
34271__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34272  uint64x1_t __ret;
34273  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34274  return __ret;
34275}
34276#else
34277__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34278  uint64x1_t __ret;
34279  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34280  return __ret;
34281}
34282#endif
34283
34284#ifdef __LITTLE_ENDIAN__
34285__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34286  int64x2_t __ret;
34287  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
34288  return __ret;
34289}
34290#else
34291__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34292  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34293  int64x2_t __ret;
34294  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
34295  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34296  return __ret;
34297}
34298#endif
34299
34300#ifdef __LITTLE_ENDIAN__
34301__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34302  int64x1_t __ret;
34303  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34304  return __ret;
34305}
34306#else
34307__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34308  int64x1_t __ret;
34309  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34310  return __ret;
34311}
34312#endif
34313
34314#ifdef __LITTLE_ENDIAN__
34315__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34316  uint64x2_t __ret;
34317  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
34318  return __ret;
34319}
34320#else
34321__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34322  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34323  uint64x2_t __ret;
34324  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
34325  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34326  return __ret;
34327}
34328#endif
34329
34330#ifdef __LITTLE_ENDIAN__
34331__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34332  uint64x1_t __ret;
34333  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34334  return __ret;
34335}
34336#else
34337__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34338  uint64x1_t __ret;
34339  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34340  return __ret;
34341}
34342#endif
34343
34344#ifdef __LITTLE_ENDIAN__
34345__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34346  poly8x8_t __ret;
34347  __ret = (poly8x8_t)(__p0);
34348  return __ret;
34349}
34350#else
34351__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34352  poly8x8_t __ret;
34353  __ret = (poly8x8_t)(__p0);
34354  return __ret;
34355}
34356#endif
34357
34358#ifdef __LITTLE_ENDIAN__
34359__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34360  poly8x8_t __ret;
34361  __ret = (poly8x8_t)(__p0);
34362  return __ret;
34363}
34364#else
34365__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34366  poly8x8_t __ret;
34367  __ret = (poly8x8_t)(__p0);
34368  return __ret;
34369}
34370#endif
34371
34372#ifdef __LITTLE_ENDIAN__
34373__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34374  poly8x8_t __ret;
34375  __ret = (poly8x8_t)(__p0);
34376  return __ret;
34377}
34378#else
34379__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34380  poly8x8_t __ret;
34381  __ret = (poly8x8_t)(__p0);
34382  return __ret;
34383}
34384#endif
34385
34386#ifdef __LITTLE_ENDIAN__
34387__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34388  poly8x8_t __ret;
34389  __ret = (poly8x8_t)(__p0);
34390  return __ret;
34391}
34392#else
34393__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34394  poly8x8_t __ret;
34395  __ret = (poly8x8_t)(__p0);
34396  return __ret;
34397}
34398#endif
34399
34400#ifdef __LITTLE_ENDIAN__
34401__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34402  poly8x8_t __ret;
34403  __ret = (poly8x8_t)(__p0);
34404  return __ret;
34405}
34406#else
34407__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34408  poly8x8_t __ret;
34409  __ret = (poly8x8_t)(__p0);
34410  return __ret;
34411}
34412#endif
34413
34414#ifdef __LITTLE_ENDIAN__
34415__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34416  poly8x8_t __ret;
34417  __ret = (poly8x8_t)(__p0);
34418  return __ret;
34419}
34420#else
34421__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34422  poly8x8_t __ret;
34423  __ret = (poly8x8_t)(__p0);
34424  return __ret;
34425}
34426#endif
34427
34428#ifdef __LITTLE_ENDIAN__
34429__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34430  poly8x8_t __ret;
34431  __ret = (poly8x8_t)(__p0);
34432  return __ret;
34433}
34434#else
34435__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34436  poly8x8_t __ret;
34437  __ret = (poly8x8_t)(__p0);
34438  return __ret;
34439}
34440#endif
34441
34442#ifdef __LITTLE_ENDIAN__
34443__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34444  poly8x8_t __ret;
34445  __ret = (poly8x8_t)(__p0);
34446  return __ret;
34447}
34448#else
34449__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34450  poly8x8_t __ret;
34451  __ret = (poly8x8_t)(__p0);
34452  return __ret;
34453}
34454#endif
34455
34456#ifdef __LITTLE_ENDIAN__
34457__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34458  poly8x8_t __ret;
34459  __ret = (poly8x8_t)(__p0);
34460  return __ret;
34461}
34462#else
34463__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34464  poly8x8_t __ret;
34465  __ret = (poly8x8_t)(__p0);
34466  return __ret;
34467}
34468#endif
34469
34470#ifdef __LITTLE_ENDIAN__
34471__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34472  poly8x8_t __ret;
34473  __ret = (poly8x8_t)(__p0);
34474  return __ret;
34475}
34476#else
34477__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34478  poly8x8_t __ret;
34479  __ret = (poly8x8_t)(__p0);
34480  return __ret;
34481}
34482#endif
34483
34484#ifdef __LITTLE_ENDIAN__
34485__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34486  poly8x8_t __ret;
34487  __ret = (poly8x8_t)(__p0);
34488  return __ret;
34489}
34490#else
34491__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34492  poly8x8_t __ret;
34493  __ret = (poly8x8_t)(__p0);
34494  return __ret;
34495}
34496#endif
34497
34498#ifdef __LITTLE_ENDIAN__
34499__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34500  poly8x8_t __ret;
34501  __ret = (poly8x8_t)(__p0);
34502  return __ret;
34503}
34504#else
34505__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34506  poly8x8_t __ret;
34507  __ret = (poly8x8_t)(__p0);
34508  return __ret;
34509}
34510#endif
34511
34512#ifdef __LITTLE_ENDIAN__
34513__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34514  poly8x8_t __ret;
34515  __ret = (poly8x8_t)(__p0);
34516  return __ret;
34517}
34518#else
34519__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34520  poly8x8_t __ret;
34521  __ret = (poly8x8_t)(__p0);
34522  return __ret;
34523}
34524#endif
34525
34526#ifdef __LITTLE_ENDIAN__
34527__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34528  poly64x1_t __ret;
34529  __ret = (poly64x1_t)(__p0);
34530  return __ret;
34531}
34532#else
34533__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34534  poly64x1_t __ret;
34535  __ret = (poly64x1_t)(__p0);
34536  return __ret;
34537}
34538#endif
34539
34540#ifdef __LITTLE_ENDIAN__
34541__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34542  poly64x1_t __ret;
34543  __ret = (poly64x1_t)(__p0);
34544  return __ret;
34545}
34546#else
34547__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34548  poly64x1_t __ret;
34549  __ret = (poly64x1_t)(__p0);
34550  return __ret;
34551}
34552#endif
34553
34554#ifdef __LITTLE_ENDIAN__
34555__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34556  poly64x1_t __ret;
34557  __ret = (poly64x1_t)(__p0);
34558  return __ret;
34559}
34560#else
34561__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34562  poly64x1_t __ret;
34563  __ret = (poly64x1_t)(__p0);
34564  return __ret;
34565}
34566#endif
34567
34568#ifdef __LITTLE_ENDIAN__
34569__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34570  poly64x1_t __ret;
34571  __ret = (poly64x1_t)(__p0);
34572  return __ret;
34573}
34574#else
34575__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34576  poly64x1_t __ret;
34577  __ret = (poly64x1_t)(__p0);
34578  return __ret;
34579}
34580#endif
34581
34582#ifdef __LITTLE_ENDIAN__
34583__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34584  poly64x1_t __ret;
34585  __ret = (poly64x1_t)(__p0);
34586  return __ret;
34587}
34588#else
34589__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34590  poly64x1_t __ret;
34591  __ret = (poly64x1_t)(__p0);
34592  return __ret;
34593}
34594#endif
34595
34596#ifdef __LITTLE_ENDIAN__
34597__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34598  poly64x1_t __ret;
34599  __ret = (poly64x1_t)(__p0);
34600  return __ret;
34601}
34602#else
34603__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34604  poly64x1_t __ret;
34605  __ret = (poly64x1_t)(__p0);
34606  return __ret;
34607}
34608#endif
34609
34610#ifdef __LITTLE_ENDIAN__
34611__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34612  poly64x1_t __ret;
34613  __ret = (poly64x1_t)(__p0);
34614  return __ret;
34615}
34616#else
34617__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34618  poly64x1_t __ret;
34619  __ret = (poly64x1_t)(__p0);
34620  return __ret;
34621}
34622#endif
34623
34624#ifdef __LITTLE_ENDIAN__
34625__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34626  poly64x1_t __ret;
34627  __ret = (poly64x1_t)(__p0);
34628  return __ret;
34629}
34630#else
34631__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34632  poly64x1_t __ret;
34633  __ret = (poly64x1_t)(__p0);
34634  return __ret;
34635}
34636#endif
34637
34638#ifdef __LITTLE_ENDIAN__
34639__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34640  poly64x1_t __ret;
34641  __ret = (poly64x1_t)(__p0);
34642  return __ret;
34643}
34644#else
34645__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34646  poly64x1_t __ret;
34647  __ret = (poly64x1_t)(__p0);
34648  return __ret;
34649}
34650#endif
34651
34652#ifdef __LITTLE_ENDIAN__
34653__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34654  poly64x1_t __ret;
34655  __ret = (poly64x1_t)(__p0);
34656  return __ret;
34657}
34658#else
34659__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34660  poly64x1_t __ret;
34661  __ret = (poly64x1_t)(__p0);
34662  return __ret;
34663}
34664#endif
34665
34666#ifdef __LITTLE_ENDIAN__
34667__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34668  poly64x1_t __ret;
34669  __ret = (poly64x1_t)(__p0);
34670  return __ret;
34671}
34672#else
34673__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34674  poly64x1_t __ret;
34675  __ret = (poly64x1_t)(__p0);
34676  return __ret;
34677}
34678#endif
34679
34680#ifdef __LITTLE_ENDIAN__
34681__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34682  poly64x1_t __ret;
34683  __ret = (poly64x1_t)(__p0);
34684  return __ret;
34685}
34686#else
34687__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34688  poly64x1_t __ret;
34689  __ret = (poly64x1_t)(__p0);
34690  return __ret;
34691}
34692#endif
34693
34694#ifdef __LITTLE_ENDIAN__
34695__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34696  poly64x1_t __ret;
34697  __ret = (poly64x1_t)(__p0);
34698  return __ret;
34699}
34700#else
34701__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34702  poly64x1_t __ret;
34703  __ret = (poly64x1_t)(__p0);
34704  return __ret;
34705}
34706#endif
34707
34708#ifdef __LITTLE_ENDIAN__
34709__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34710  poly16x4_t __ret;
34711  __ret = (poly16x4_t)(__p0);
34712  return __ret;
34713}
34714#else
34715__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34716  poly16x4_t __ret;
34717  __ret = (poly16x4_t)(__p0);
34718  return __ret;
34719}
34720#endif
34721
34722#ifdef __LITTLE_ENDIAN__
34723__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34724  poly16x4_t __ret;
34725  __ret = (poly16x4_t)(__p0);
34726  return __ret;
34727}
34728#else
34729__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34730  poly16x4_t __ret;
34731  __ret = (poly16x4_t)(__p0);
34732  return __ret;
34733}
34734#endif
34735
34736#ifdef __LITTLE_ENDIAN__
34737__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34738  poly16x4_t __ret;
34739  __ret = (poly16x4_t)(__p0);
34740  return __ret;
34741}
34742#else
34743__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34744  poly16x4_t __ret;
34745  __ret = (poly16x4_t)(__p0);
34746  return __ret;
34747}
34748#endif
34749
34750#ifdef __LITTLE_ENDIAN__
34751__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34752  poly16x4_t __ret;
34753  __ret = (poly16x4_t)(__p0);
34754  return __ret;
34755}
34756#else
34757__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34758  poly16x4_t __ret;
34759  __ret = (poly16x4_t)(__p0);
34760  return __ret;
34761}
34762#endif
34763
34764#ifdef __LITTLE_ENDIAN__
34765__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34766  poly16x4_t __ret;
34767  __ret = (poly16x4_t)(__p0);
34768  return __ret;
34769}
34770#else
34771__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34772  poly16x4_t __ret;
34773  __ret = (poly16x4_t)(__p0);
34774  return __ret;
34775}
34776#endif
34777
34778#ifdef __LITTLE_ENDIAN__
34779__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34780  poly16x4_t __ret;
34781  __ret = (poly16x4_t)(__p0);
34782  return __ret;
34783}
34784#else
34785__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34786  poly16x4_t __ret;
34787  __ret = (poly16x4_t)(__p0);
34788  return __ret;
34789}
34790#endif
34791
34792#ifdef __LITTLE_ENDIAN__
34793__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34794  poly16x4_t __ret;
34795  __ret = (poly16x4_t)(__p0);
34796  return __ret;
34797}
34798#else
34799__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34800  poly16x4_t __ret;
34801  __ret = (poly16x4_t)(__p0);
34802  return __ret;
34803}
34804#endif
34805
34806#ifdef __LITTLE_ENDIAN__
34807__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34808  poly16x4_t __ret;
34809  __ret = (poly16x4_t)(__p0);
34810  return __ret;
34811}
34812#else
34813__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34814  poly16x4_t __ret;
34815  __ret = (poly16x4_t)(__p0);
34816  return __ret;
34817}
34818#endif
34819
34820#ifdef __LITTLE_ENDIAN__
34821__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34822  poly16x4_t __ret;
34823  __ret = (poly16x4_t)(__p0);
34824  return __ret;
34825}
34826#else
34827__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34828  poly16x4_t __ret;
34829  __ret = (poly16x4_t)(__p0);
34830  return __ret;
34831}
34832#endif
34833
34834#ifdef __LITTLE_ENDIAN__
34835__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34836  poly16x4_t __ret;
34837  __ret = (poly16x4_t)(__p0);
34838  return __ret;
34839}
34840#else
34841__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34842  poly16x4_t __ret;
34843  __ret = (poly16x4_t)(__p0);
34844  return __ret;
34845}
34846#endif
34847
34848#ifdef __LITTLE_ENDIAN__
34849__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34850  poly16x4_t __ret;
34851  __ret = (poly16x4_t)(__p0);
34852  return __ret;
34853}
34854#else
34855__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34856  poly16x4_t __ret;
34857  __ret = (poly16x4_t)(__p0);
34858  return __ret;
34859}
34860#endif
34861
34862#ifdef __LITTLE_ENDIAN__
34863__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34864  poly16x4_t __ret;
34865  __ret = (poly16x4_t)(__p0);
34866  return __ret;
34867}
34868#else
34869__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34870  poly16x4_t __ret;
34871  __ret = (poly16x4_t)(__p0);
34872  return __ret;
34873}
34874#endif
34875
34876#ifdef __LITTLE_ENDIAN__
34877__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34878  poly16x4_t __ret;
34879  __ret = (poly16x4_t)(__p0);
34880  return __ret;
34881}
34882#else
34883__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34884  poly16x4_t __ret;
34885  __ret = (poly16x4_t)(__p0);
34886  return __ret;
34887}
34888#endif
34889
34890#ifdef __LITTLE_ENDIAN__
34891__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34892  poly8x16_t __ret;
34893  __ret = (poly8x16_t)(__p0);
34894  return __ret;
34895}
34896#else
34897__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34898  poly8x16_t __ret;
34899  __ret = (poly8x16_t)(__p0);
34900  return __ret;
34901}
34902#endif
34903
34904#ifdef __LITTLE_ENDIAN__
34905__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34906  poly8x16_t __ret;
34907  __ret = (poly8x16_t)(__p0);
34908  return __ret;
34909}
34910#else
34911__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34912  poly8x16_t __ret;
34913  __ret = (poly8x16_t)(__p0);
34914  return __ret;
34915}
34916#endif
34917
34918#ifdef __LITTLE_ENDIAN__
34919__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34920  poly8x16_t __ret;
34921  __ret = (poly8x16_t)(__p0);
34922  return __ret;
34923}
34924#else
34925__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34926  poly8x16_t __ret;
34927  __ret = (poly8x16_t)(__p0);
34928  return __ret;
34929}
34930#endif
34931
34932#ifdef __LITTLE_ENDIAN__
34933__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34934  poly8x16_t __ret;
34935  __ret = (poly8x16_t)(__p0);
34936  return __ret;
34937}
34938#else
34939__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34940  poly8x16_t __ret;
34941  __ret = (poly8x16_t)(__p0);
34942  return __ret;
34943}
34944#endif
34945
34946#ifdef __LITTLE_ENDIAN__
34947__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34948  poly8x16_t __ret;
34949  __ret = (poly8x16_t)(__p0);
34950  return __ret;
34951}
34952#else
34953__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34954  poly8x16_t __ret;
34955  __ret = (poly8x16_t)(__p0);
34956  return __ret;
34957}
34958#endif
34959
34960#ifdef __LITTLE_ENDIAN__
34961__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
34962  poly8x16_t __ret;
34963  __ret = (poly8x16_t)(__p0);
34964  return __ret;
34965}
34966#else
34967__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
34968  poly8x16_t __ret;
34969  __ret = (poly8x16_t)(__p0);
34970  return __ret;
34971}
34972#endif
34973
34974#ifdef __LITTLE_ENDIAN__
34975__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
34976  poly8x16_t __ret;
34977  __ret = (poly8x16_t)(__p0);
34978  return __ret;
34979}
34980#else
34981__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
34982  poly8x16_t __ret;
34983  __ret = (poly8x16_t)(__p0);
34984  return __ret;
34985}
34986#endif
34987
34988#ifdef __LITTLE_ENDIAN__
34989__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
34990  poly8x16_t __ret;
34991  __ret = (poly8x16_t)(__p0);
34992  return __ret;
34993}
34994#else
34995__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
34996  poly8x16_t __ret;
34997  __ret = (poly8x16_t)(__p0);
34998  return __ret;
34999}
35000#endif
35001
35002#ifdef __LITTLE_ENDIAN__
35003__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35004  poly8x16_t __ret;
35005  __ret = (poly8x16_t)(__p0);
35006  return __ret;
35007}
35008#else
35009__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35010  poly8x16_t __ret;
35011  __ret = (poly8x16_t)(__p0);
35012  return __ret;
35013}
35014#endif
35015
35016#ifdef __LITTLE_ENDIAN__
35017__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35018  poly8x16_t __ret;
35019  __ret = (poly8x16_t)(__p0);
35020  return __ret;
35021}
35022#else
35023__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35024  poly8x16_t __ret;
35025  __ret = (poly8x16_t)(__p0);
35026  return __ret;
35027}
35028#endif
35029
35030#ifdef __LITTLE_ENDIAN__
35031__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35032  poly8x16_t __ret;
35033  __ret = (poly8x16_t)(__p0);
35034  return __ret;
35035}
35036#else
35037__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35038  poly8x16_t __ret;
35039  __ret = (poly8x16_t)(__p0);
35040  return __ret;
35041}
35042#endif
35043
35044#ifdef __LITTLE_ENDIAN__
35045__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35046  poly8x16_t __ret;
35047  __ret = (poly8x16_t)(__p0);
35048  return __ret;
35049}
35050#else
35051__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35052  poly8x16_t __ret;
35053  __ret = (poly8x16_t)(__p0);
35054  return __ret;
35055}
35056#endif
35057
35058#ifdef __LITTLE_ENDIAN__
35059__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35060  poly8x16_t __ret;
35061  __ret = (poly8x16_t)(__p0);
35062  return __ret;
35063}
35064#else
35065__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35066  poly8x16_t __ret;
35067  __ret = (poly8x16_t)(__p0);
35068  return __ret;
35069}
35070#endif
35071
35072#ifdef __LITTLE_ENDIAN__
35073__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35074  poly8x16_t __ret;
35075  __ret = (poly8x16_t)(__p0);
35076  return __ret;
35077}
35078#else
35079__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35080  poly8x16_t __ret;
35081  __ret = (poly8x16_t)(__p0);
35082  return __ret;
35083}
35084#endif
35085
35086#ifdef __LITTLE_ENDIAN__
35087__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35088  poly128_t __ret;
35089  __ret = (poly128_t)(__p0);
35090  return __ret;
35091}
35092#else
35093__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35094  poly128_t __ret;
35095  __ret = (poly128_t)(__p0);
35096  return __ret;
35097}
35098#endif
35099
35100#ifdef __LITTLE_ENDIAN__
35101__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35102  poly128_t __ret;
35103  __ret = (poly128_t)(__p0);
35104  return __ret;
35105}
35106#else
35107__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35108  poly128_t __ret;
35109  __ret = (poly128_t)(__p0);
35110  return __ret;
35111}
35112#endif
35113
35114#ifdef __LITTLE_ENDIAN__
35115__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35116  poly128_t __ret;
35117  __ret = (poly128_t)(__p0);
35118  return __ret;
35119}
35120#else
35121__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35122  poly128_t __ret;
35123  __ret = (poly128_t)(__p0);
35124  return __ret;
35125}
35126#endif
35127
35128#ifdef __LITTLE_ENDIAN__
35129__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35130  poly128_t __ret;
35131  __ret = (poly128_t)(__p0);
35132  return __ret;
35133}
35134#else
35135__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35136  poly128_t __ret;
35137  __ret = (poly128_t)(__p0);
35138  return __ret;
35139}
35140#endif
35141
35142#ifdef __LITTLE_ENDIAN__
35143__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35144  poly128_t __ret;
35145  __ret = (poly128_t)(__p0);
35146  return __ret;
35147}
35148#else
35149__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35150  poly128_t __ret;
35151  __ret = (poly128_t)(__p0);
35152  return __ret;
35153}
35154#endif
35155
35156#ifdef __LITTLE_ENDIAN__
35157__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35158  poly128_t __ret;
35159  __ret = (poly128_t)(__p0);
35160  return __ret;
35161}
35162#else
35163__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35164  poly128_t __ret;
35165  __ret = (poly128_t)(__p0);
35166  return __ret;
35167}
35168#endif
35169
35170#ifdef __LITTLE_ENDIAN__
35171__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35172  poly128_t __ret;
35173  __ret = (poly128_t)(__p0);
35174  return __ret;
35175}
35176#else
35177__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35178  poly128_t __ret;
35179  __ret = (poly128_t)(__p0);
35180  return __ret;
35181}
35182#endif
35183
35184#ifdef __LITTLE_ENDIAN__
35185__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35186  poly128_t __ret;
35187  __ret = (poly128_t)(__p0);
35188  return __ret;
35189}
35190#else
35191__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35192  poly128_t __ret;
35193  __ret = (poly128_t)(__p0);
35194  return __ret;
35195}
35196#endif
35197
35198#ifdef __LITTLE_ENDIAN__
35199__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35200  poly128_t __ret;
35201  __ret = (poly128_t)(__p0);
35202  return __ret;
35203}
35204#else
35205__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35206  poly128_t __ret;
35207  __ret = (poly128_t)(__p0);
35208  return __ret;
35209}
35210#endif
35211
35212#ifdef __LITTLE_ENDIAN__
35213__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35214  poly128_t __ret;
35215  __ret = (poly128_t)(__p0);
35216  return __ret;
35217}
35218#else
35219__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35220  poly128_t __ret;
35221  __ret = (poly128_t)(__p0);
35222  return __ret;
35223}
35224#endif
35225
35226#ifdef __LITTLE_ENDIAN__
35227__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35228  poly128_t __ret;
35229  __ret = (poly128_t)(__p0);
35230  return __ret;
35231}
35232#else
35233__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35234  poly128_t __ret;
35235  __ret = (poly128_t)(__p0);
35236  return __ret;
35237}
35238#endif
35239
35240#ifdef __LITTLE_ENDIAN__
35241__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35242  poly128_t __ret;
35243  __ret = (poly128_t)(__p0);
35244  return __ret;
35245}
35246#else
35247__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35248  poly128_t __ret;
35249  __ret = (poly128_t)(__p0);
35250  return __ret;
35251}
35252#endif
35253
35254#ifdef __LITTLE_ENDIAN__
35255__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35256  poly128_t __ret;
35257  __ret = (poly128_t)(__p0);
35258  return __ret;
35259}
35260#else
35261__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35262  poly128_t __ret;
35263  __ret = (poly128_t)(__p0);
35264  return __ret;
35265}
35266#endif
35267
35268#ifdef __LITTLE_ENDIAN__
35269__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35270  poly128_t __ret;
35271  __ret = (poly128_t)(__p0);
35272  return __ret;
35273}
35274#else
35275__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35276  poly128_t __ret;
35277  __ret = (poly128_t)(__p0);
35278  return __ret;
35279}
35280#endif
35281
35282#ifdef __LITTLE_ENDIAN__
35283__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35284  poly64x2_t __ret;
35285  __ret = (poly64x2_t)(__p0);
35286  return __ret;
35287}
35288#else
35289__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35290  poly64x2_t __ret;
35291  __ret = (poly64x2_t)(__p0);
35292  return __ret;
35293}
35294#endif
35295
35296#ifdef __LITTLE_ENDIAN__
35297__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35298  poly64x2_t __ret;
35299  __ret = (poly64x2_t)(__p0);
35300  return __ret;
35301}
35302#else
35303__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35304  poly64x2_t __ret;
35305  __ret = (poly64x2_t)(__p0);
35306  return __ret;
35307}
35308#endif
35309
35310#ifdef __LITTLE_ENDIAN__
35311__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35312  poly64x2_t __ret;
35313  __ret = (poly64x2_t)(__p0);
35314  return __ret;
35315}
35316#else
35317__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35318  poly64x2_t __ret;
35319  __ret = (poly64x2_t)(__p0);
35320  return __ret;
35321}
35322#endif
35323
35324#ifdef __LITTLE_ENDIAN__
35325__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35326  poly64x2_t __ret;
35327  __ret = (poly64x2_t)(__p0);
35328  return __ret;
35329}
35330#else
35331__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35332  poly64x2_t __ret;
35333  __ret = (poly64x2_t)(__p0);
35334  return __ret;
35335}
35336#endif
35337
35338#ifdef __LITTLE_ENDIAN__
35339__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35340  poly64x2_t __ret;
35341  __ret = (poly64x2_t)(__p0);
35342  return __ret;
35343}
35344#else
35345__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35346  poly64x2_t __ret;
35347  __ret = (poly64x2_t)(__p0);
35348  return __ret;
35349}
35350#endif
35351
35352#ifdef __LITTLE_ENDIAN__
35353__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35354  poly64x2_t __ret;
35355  __ret = (poly64x2_t)(__p0);
35356  return __ret;
35357}
35358#else
35359__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35360  poly64x2_t __ret;
35361  __ret = (poly64x2_t)(__p0);
35362  return __ret;
35363}
35364#endif
35365
35366#ifdef __LITTLE_ENDIAN__
35367__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35368  poly64x2_t __ret;
35369  __ret = (poly64x2_t)(__p0);
35370  return __ret;
35371}
35372#else
35373__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35374  poly64x2_t __ret;
35375  __ret = (poly64x2_t)(__p0);
35376  return __ret;
35377}
35378#endif
35379
35380#ifdef __LITTLE_ENDIAN__
35381__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35382  poly64x2_t __ret;
35383  __ret = (poly64x2_t)(__p0);
35384  return __ret;
35385}
35386#else
35387__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35388  poly64x2_t __ret;
35389  __ret = (poly64x2_t)(__p0);
35390  return __ret;
35391}
35392#endif
35393
35394#ifdef __LITTLE_ENDIAN__
35395__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35396  poly64x2_t __ret;
35397  __ret = (poly64x2_t)(__p0);
35398  return __ret;
35399}
35400#else
35401__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35402  poly64x2_t __ret;
35403  __ret = (poly64x2_t)(__p0);
35404  return __ret;
35405}
35406#endif
35407
35408#ifdef __LITTLE_ENDIAN__
35409__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35410  poly64x2_t __ret;
35411  __ret = (poly64x2_t)(__p0);
35412  return __ret;
35413}
35414#else
35415__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35416  poly64x2_t __ret;
35417  __ret = (poly64x2_t)(__p0);
35418  return __ret;
35419}
35420#endif
35421
35422#ifdef __LITTLE_ENDIAN__
35423__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35424  poly64x2_t __ret;
35425  __ret = (poly64x2_t)(__p0);
35426  return __ret;
35427}
35428#else
35429__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35430  poly64x2_t __ret;
35431  __ret = (poly64x2_t)(__p0);
35432  return __ret;
35433}
35434#endif
35435
35436#ifdef __LITTLE_ENDIAN__
35437__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35438  poly64x2_t __ret;
35439  __ret = (poly64x2_t)(__p0);
35440  return __ret;
35441}
35442#else
35443__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35444  poly64x2_t __ret;
35445  __ret = (poly64x2_t)(__p0);
35446  return __ret;
35447}
35448#endif
35449
35450#ifdef __LITTLE_ENDIAN__
35451__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35452  poly64x2_t __ret;
35453  __ret = (poly64x2_t)(__p0);
35454  return __ret;
35455}
35456#else
35457__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35458  poly64x2_t __ret;
35459  __ret = (poly64x2_t)(__p0);
35460  return __ret;
35461}
35462#endif
35463
35464#ifdef __LITTLE_ENDIAN__
35465__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35466  poly64x2_t __ret;
35467  __ret = (poly64x2_t)(__p0);
35468  return __ret;
35469}
35470#else
35471__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35472  poly64x2_t __ret;
35473  __ret = (poly64x2_t)(__p0);
35474  return __ret;
35475}
35476#endif
35477
35478#ifdef __LITTLE_ENDIAN__
35479__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35480  poly16x8_t __ret;
35481  __ret = (poly16x8_t)(__p0);
35482  return __ret;
35483}
35484#else
35485__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35486  poly16x8_t __ret;
35487  __ret = (poly16x8_t)(__p0);
35488  return __ret;
35489}
35490#endif
35491
35492#ifdef __LITTLE_ENDIAN__
35493__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35494  poly16x8_t __ret;
35495  __ret = (poly16x8_t)(__p0);
35496  return __ret;
35497}
35498#else
35499__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35500  poly16x8_t __ret;
35501  __ret = (poly16x8_t)(__p0);
35502  return __ret;
35503}
35504#endif
35505
35506#ifdef __LITTLE_ENDIAN__
35507__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35508  poly16x8_t __ret;
35509  __ret = (poly16x8_t)(__p0);
35510  return __ret;
35511}
35512#else
35513__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35514  poly16x8_t __ret;
35515  __ret = (poly16x8_t)(__p0);
35516  return __ret;
35517}
35518#endif
35519
35520#ifdef __LITTLE_ENDIAN__
35521__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35522  poly16x8_t __ret;
35523  __ret = (poly16x8_t)(__p0);
35524  return __ret;
35525}
35526#else
35527__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35528  poly16x8_t __ret;
35529  __ret = (poly16x8_t)(__p0);
35530  return __ret;
35531}
35532#endif
35533
35534#ifdef __LITTLE_ENDIAN__
35535__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35536  poly16x8_t __ret;
35537  __ret = (poly16x8_t)(__p0);
35538  return __ret;
35539}
35540#else
35541__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35542  poly16x8_t __ret;
35543  __ret = (poly16x8_t)(__p0);
35544  return __ret;
35545}
35546#endif
35547
35548#ifdef __LITTLE_ENDIAN__
35549__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35550  poly16x8_t __ret;
35551  __ret = (poly16x8_t)(__p0);
35552  return __ret;
35553}
35554#else
35555__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35556  poly16x8_t __ret;
35557  __ret = (poly16x8_t)(__p0);
35558  return __ret;
35559}
35560#endif
35561
35562#ifdef __LITTLE_ENDIAN__
35563__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35564  poly16x8_t __ret;
35565  __ret = (poly16x8_t)(__p0);
35566  return __ret;
35567}
35568#else
35569__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35570  poly16x8_t __ret;
35571  __ret = (poly16x8_t)(__p0);
35572  return __ret;
35573}
35574#endif
35575
35576#ifdef __LITTLE_ENDIAN__
35577__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35578  poly16x8_t __ret;
35579  __ret = (poly16x8_t)(__p0);
35580  return __ret;
35581}
35582#else
35583__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35584  poly16x8_t __ret;
35585  __ret = (poly16x8_t)(__p0);
35586  return __ret;
35587}
35588#endif
35589
35590#ifdef __LITTLE_ENDIAN__
35591__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35592  poly16x8_t __ret;
35593  __ret = (poly16x8_t)(__p0);
35594  return __ret;
35595}
35596#else
35597__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35598  poly16x8_t __ret;
35599  __ret = (poly16x8_t)(__p0);
35600  return __ret;
35601}
35602#endif
35603
35604#ifdef __LITTLE_ENDIAN__
35605__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35606  poly16x8_t __ret;
35607  __ret = (poly16x8_t)(__p0);
35608  return __ret;
35609}
35610#else
35611__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35612  poly16x8_t __ret;
35613  __ret = (poly16x8_t)(__p0);
35614  return __ret;
35615}
35616#endif
35617
35618#ifdef __LITTLE_ENDIAN__
35619__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35620  poly16x8_t __ret;
35621  __ret = (poly16x8_t)(__p0);
35622  return __ret;
35623}
35624#else
35625__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35626  poly16x8_t __ret;
35627  __ret = (poly16x8_t)(__p0);
35628  return __ret;
35629}
35630#endif
35631
35632#ifdef __LITTLE_ENDIAN__
35633__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35634  poly16x8_t __ret;
35635  __ret = (poly16x8_t)(__p0);
35636  return __ret;
35637}
35638#else
35639__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35640  poly16x8_t __ret;
35641  __ret = (poly16x8_t)(__p0);
35642  return __ret;
35643}
35644#endif
35645
35646#ifdef __LITTLE_ENDIAN__
35647__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35648  poly16x8_t __ret;
35649  __ret = (poly16x8_t)(__p0);
35650  return __ret;
35651}
35652#else
35653__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35654  poly16x8_t __ret;
35655  __ret = (poly16x8_t)(__p0);
35656  return __ret;
35657}
35658#endif
35659
35660#ifdef __LITTLE_ENDIAN__
35661__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35662  poly16x8_t __ret;
35663  __ret = (poly16x8_t)(__p0);
35664  return __ret;
35665}
35666#else
35667__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35668  poly16x8_t __ret;
35669  __ret = (poly16x8_t)(__p0);
35670  return __ret;
35671}
35672#endif
35673
35674#ifdef __LITTLE_ENDIAN__
35675__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35676  uint8x16_t __ret;
35677  __ret = (uint8x16_t)(__p0);
35678  return __ret;
35679}
35680#else
35681__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35682  uint8x16_t __ret;
35683  __ret = (uint8x16_t)(__p0);
35684  return __ret;
35685}
35686#endif
35687
35688#ifdef __LITTLE_ENDIAN__
35689__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35690  uint8x16_t __ret;
35691  __ret = (uint8x16_t)(__p0);
35692  return __ret;
35693}
35694#else
35695__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35696  uint8x16_t __ret;
35697  __ret = (uint8x16_t)(__p0);
35698  return __ret;
35699}
35700#endif
35701
35702#ifdef __LITTLE_ENDIAN__
35703__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35704  uint8x16_t __ret;
35705  __ret = (uint8x16_t)(__p0);
35706  return __ret;
35707}
35708#else
35709__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35710  uint8x16_t __ret;
35711  __ret = (uint8x16_t)(__p0);
35712  return __ret;
35713}
35714#endif
35715
35716#ifdef __LITTLE_ENDIAN__
35717__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35718  uint8x16_t __ret;
35719  __ret = (uint8x16_t)(__p0);
35720  return __ret;
35721}
35722#else
35723__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35724  uint8x16_t __ret;
35725  __ret = (uint8x16_t)(__p0);
35726  return __ret;
35727}
35728#endif
35729
35730#ifdef __LITTLE_ENDIAN__
35731__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35732  uint8x16_t __ret;
35733  __ret = (uint8x16_t)(__p0);
35734  return __ret;
35735}
35736#else
35737__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35738  uint8x16_t __ret;
35739  __ret = (uint8x16_t)(__p0);
35740  return __ret;
35741}
35742#endif
35743
35744#ifdef __LITTLE_ENDIAN__
35745__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35746  uint8x16_t __ret;
35747  __ret = (uint8x16_t)(__p0);
35748  return __ret;
35749}
35750#else
35751__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35752  uint8x16_t __ret;
35753  __ret = (uint8x16_t)(__p0);
35754  return __ret;
35755}
35756#endif
35757
35758#ifdef __LITTLE_ENDIAN__
35759__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35760  uint8x16_t __ret;
35761  __ret = (uint8x16_t)(__p0);
35762  return __ret;
35763}
35764#else
35765__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35766  uint8x16_t __ret;
35767  __ret = (uint8x16_t)(__p0);
35768  return __ret;
35769}
35770#endif
35771
35772#ifdef __LITTLE_ENDIAN__
35773__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35774  uint8x16_t __ret;
35775  __ret = (uint8x16_t)(__p0);
35776  return __ret;
35777}
35778#else
35779__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35780  uint8x16_t __ret;
35781  __ret = (uint8x16_t)(__p0);
35782  return __ret;
35783}
35784#endif
35785
35786#ifdef __LITTLE_ENDIAN__
35787__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35788  uint8x16_t __ret;
35789  __ret = (uint8x16_t)(__p0);
35790  return __ret;
35791}
35792#else
35793__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35794  uint8x16_t __ret;
35795  __ret = (uint8x16_t)(__p0);
35796  return __ret;
35797}
35798#endif
35799
35800#ifdef __LITTLE_ENDIAN__
35801__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35802  uint8x16_t __ret;
35803  __ret = (uint8x16_t)(__p0);
35804  return __ret;
35805}
35806#else
35807__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35808  uint8x16_t __ret;
35809  __ret = (uint8x16_t)(__p0);
35810  return __ret;
35811}
35812#endif
35813
35814#ifdef __LITTLE_ENDIAN__
35815__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35816  uint8x16_t __ret;
35817  __ret = (uint8x16_t)(__p0);
35818  return __ret;
35819}
35820#else
35821__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35822  uint8x16_t __ret;
35823  __ret = (uint8x16_t)(__p0);
35824  return __ret;
35825}
35826#endif
35827
35828#ifdef __LITTLE_ENDIAN__
35829__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35830  uint8x16_t __ret;
35831  __ret = (uint8x16_t)(__p0);
35832  return __ret;
35833}
35834#else
35835__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35836  uint8x16_t __ret;
35837  __ret = (uint8x16_t)(__p0);
35838  return __ret;
35839}
35840#endif
35841
35842#ifdef __LITTLE_ENDIAN__
35843__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35844  uint8x16_t __ret;
35845  __ret = (uint8x16_t)(__p0);
35846  return __ret;
35847}
35848#else
35849__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35850  uint8x16_t __ret;
35851  __ret = (uint8x16_t)(__p0);
35852  return __ret;
35853}
35854#endif
35855
35856#ifdef __LITTLE_ENDIAN__
35857__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35858  uint8x16_t __ret;
35859  __ret = (uint8x16_t)(__p0);
35860  return __ret;
35861}
35862#else
35863__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35864  uint8x16_t __ret;
35865  __ret = (uint8x16_t)(__p0);
35866  return __ret;
35867}
35868#endif
35869
35870#ifdef __LITTLE_ENDIAN__
35871__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35872  uint32x4_t __ret;
35873  __ret = (uint32x4_t)(__p0);
35874  return __ret;
35875}
35876#else
35877__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35878  uint32x4_t __ret;
35879  __ret = (uint32x4_t)(__p0);
35880  return __ret;
35881}
35882#endif
35883
35884#ifdef __LITTLE_ENDIAN__
35885__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35886  uint32x4_t __ret;
35887  __ret = (uint32x4_t)(__p0);
35888  return __ret;
35889}
35890#else
35891__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35892  uint32x4_t __ret;
35893  __ret = (uint32x4_t)(__p0);
35894  return __ret;
35895}
35896#endif
35897
35898#ifdef __LITTLE_ENDIAN__
35899__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35900  uint32x4_t __ret;
35901  __ret = (uint32x4_t)(__p0);
35902  return __ret;
35903}
35904#else
35905__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35906  uint32x4_t __ret;
35907  __ret = (uint32x4_t)(__p0);
35908  return __ret;
35909}
35910#endif
35911
35912#ifdef __LITTLE_ENDIAN__
35913__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35914  uint32x4_t __ret;
35915  __ret = (uint32x4_t)(__p0);
35916  return __ret;
35917}
35918#else
35919__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35920  uint32x4_t __ret;
35921  __ret = (uint32x4_t)(__p0);
35922  return __ret;
35923}
35924#endif
35925
35926#ifdef __LITTLE_ENDIAN__
35927__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35928  uint32x4_t __ret;
35929  __ret = (uint32x4_t)(__p0);
35930  return __ret;
35931}
35932#else
35933__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35934  uint32x4_t __ret;
35935  __ret = (uint32x4_t)(__p0);
35936  return __ret;
35937}
35938#endif
35939
35940#ifdef __LITTLE_ENDIAN__
35941__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35942  uint32x4_t __ret;
35943  __ret = (uint32x4_t)(__p0);
35944  return __ret;
35945}
35946#else
35947__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35948  uint32x4_t __ret;
35949  __ret = (uint32x4_t)(__p0);
35950  return __ret;
35951}
35952#endif
35953
35954#ifdef __LITTLE_ENDIAN__
35955__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
35956  uint32x4_t __ret;
35957  __ret = (uint32x4_t)(__p0);
35958  return __ret;
35959}
35960#else
35961__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
35962  uint32x4_t __ret;
35963  __ret = (uint32x4_t)(__p0);
35964  return __ret;
35965}
35966#endif
35967
35968#ifdef __LITTLE_ENDIAN__
35969__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
35970  uint32x4_t __ret;
35971  __ret = (uint32x4_t)(__p0);
35972  return __ret;
35973}
35974#else
35975__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
35976  uint32x4_t __ret;
35977  __ret = (uint32x4_t)(__p0);
35978  return __ret;
35979}
35980#endif
35981
35982#ifdef __LITTLE_ENDIAN__
35983__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
35984  uint32x4_t __ret;
35985  __ret = (uint32x4_t)(__p0);
35986  return __ret;
35987}
35988#else
35989__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
35990  uint32x4_t __ret;
35991  __ret = (uint32x4_t)(__p0);
35992  return __ret;
35993}
35994#endif
35995
35996#ifdef __LITTLE_ENDIAN__
35997__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
35998  uint32x4_t __ret;
35999  __ret = (uint32x4_t)(__p0);
36000  return __ret;
36001}
36002#else
36003__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36004  uint32x4_t __ret;
36005  __ret = (uint32x4_t)(__p0);
36006  return __ret;
36007}
36008#endif
36009
36010#ifdef __LITTLE_ENDIAN__
36011__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36012  uint32x4_t __ret;
36013  __ret = (uint32x4_t)(__p0);
36014  return __ret;
36015}
36016#else
36017__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36018  uint32x4_t __ret;
36019  __ret = (uint32x4_t)(__p0);
36020  return __ret;
36021}
36022#endif
36023
36024#ifdef __LITTLE_ENDIAN__
36025__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36026  uint32x4_t __ret;
36027  __ret = (uint32x4_t)(__p0);
36028  return __ret;
36029}
36030#else
36031__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36032  uint32x4_t __ret;
36033  __ret = (uint32x4_t)(__p0);
36034  return __ret;
36035}
36036#endif
36037
36038#ifdef __LITTLE_ENDIAN__
36039__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36040  uint32x4_t __ret;
36041  __ret = (uint32x4_t)(__p0);
36042  return __ret;
36043}
36044#else
36045__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36046  uint32x4_t __ret;
36047  __ret = (uint32x4_t)(__p0);
36048  return __ret;
36049}
36050#endif
36051
36052#ifdef __LITTLE_ENDIAN__
36053__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36054  uint32x4_t __ret;
36055  __ret = (uint32x4_t)(__p0);
36056  return __ret;
36057}
36058#else
36059__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36060  uint32x4_t __ret;
36061  __ret = (uint32x4_t)(__p0);
36062  return __ret;
36063}
36064#endif
36065
36066#ifdef __LITTLE_ENDIAN__
36067__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36068  uint64x2_t __ret;
36069  __ret = (uint64x2_t)(__p0);
36070  return __ret;
36071}
36072#else
36073__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36074  uint64x2_t __ret;
36075  __ret = (uint64x2_t)(__p0);
36076  return __ret;
36077}
36078#endif
36079
36080#ifdef __LITTLE_ENDIAN__
36081__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36082  uint64x2_t __ret;
36083  __ret = (uint64x2_t)(__p0);
36084  return __ret;
36085}
36086#else
36087__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36088  uint64x2_t __ret;
36089  __ret = (uint64x2_t)(__p0);
36090  return __ret;
36091}
36092#endif
36093
36094#ifdef __LITTLE_ENDIAN__
36095__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36096  uint64x2_t __ret;
36097  __ret = (uint64x2_t)(__p0);
36098  return __ret;
36099}
36100#else
36101__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36102  uint64x2_t __ret;
36103  __ret = (uint64x2_t)(__p0);
36104  return __ret;
36105}
36106#endif
36107
36108#ifdef __LITTLE_ENDIAN__
36109__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36110  uint64x2_t __ret;
36111  __ret = (uint64x2_t)(__p0);
36112  return __ret;
36113}
36114#else
36115__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36116  uint64x2_t __ret;
36117  __ret = (uint64x2_t)(__p0);
36118  return __ret;
36119}
36120#endif
36121
36122#ifdef __LITTLE_ENDIAN__
36123__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36124  uint64x2_t __ret;
36125  __ret = (uint64x2_t)(__p0);
36126  return __ret;
36127}
36128#else
36129__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36130  uint64x2_t __ret;
36131  __ret = (uint64x2_t)(__p0);
36132  return __ret;
36133}
36134#endif
36135
36136#ifdef __LITTLE_ENDIAN__
36137__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36138  uint64x2_t __ret;
36139  __ret = (uint64x2_t)(__p0);
36140  return __ret;
36141}
36142#else
36143__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36144  uint64x2_t __ret;
36145  __ret = (uint64x2_t)(__p0);
36146  return __ret;
36147}
36148#endif
36149
36150#ifdef __LITTLE_ENDIAN__
36151__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36152  uint64x2_t __ret;
36153  __ret = (uint64x2_t)(__p0);
36154  return __ret;
36155}
36156#else
36157__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36158  uint64x2_t __ret;
36159  __ret = (uint64x2_t)(__p0);
36160  return __ret;
36161}
36162#endif
36163
36164#ifdef __LITTLE_ENDIAN__
36165__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36166  uint64x2_t __ret;
36167  __ret = (uint64x2_t)(__p0);
36168  return __ret;
36169}
36170#else
36171__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36172  uint64x2_t __ret;
36173  __ret = (uint64x2_t)(__p0);
36174  return __ret;
36175}
36176#endif
36177
36178#ifdef __LITTLE_ENDIAN__
36179__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36180  uint64x2_t __ret;
36181  __ret = (uint64x2_t)(__p0);
36182  return __ret;
36183}
36184#else
36185__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36186  uint64x2_t __ret;
36187  __ret = (uint64x2_t)(__p0);
36188  return __ret;
36189}
36190#endif
36191
36192#ifdef __LITTLE_ENDIAN__
36193__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36194  uint64x2_t __ret;
36195  __ret = (uint64x2_t)(__p0);
36196  return __ret;
36197}
36198#else
36199__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36200  uint64x2_t __ret;
36201  __ret = (uint64x2_t)(__p0);
36202  return __ret;
36203}
36204#endif
36205
36206#ifdef __LITTLE_ENDIAN__
36207__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36208  uint64x2_t __ret;
36209  __ret = (uint64x2_t)(__p0);
36210  return __ret;
36211}
36212#else
36213__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36214  uint64x2_t __ret;
36215  __ret = (uint64x2_t)(__p0);
36216  return __ret;
36217}
36218#endif
36219
36220#ifdef __LITTLE_ENDIAN__
36221__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36222  uint64x2_t __ret;
36223  __ret = (uint64x2_t)(__p0);
36224  return __ret;
36225}
36226#else
36227__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36228  uint64x2_t __ret;
36229  __ret = (uint64x2_t)(__p0);
36230  return __ret;
36231}
36232#endif
36233
36234#ifdef __LITTLE_ENDIAN__
36235__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36236  uint64x2_t __ret;
36237  __ret = (uint64x2_t)(__p0);
36238  return __ret;
36239}
36240#else
36241__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36242  uint64x2_t __ret;
36243  __ret = (uint64x2_t)(__p0);
36244  return __ret;
36245}
36246#endif
36247
36248#ifdef __LITTLE_ENDIAN__
36249__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36250  uint64x2_t __ret;
36251  __ret = (uint64x2_t)(__p0);
36252  return __ret;
36253}
36254#else
36255__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36256  uint64x2_t __ret;
36257  __ret = (uint64x2_t)(__p0);
36258  return __ret;
36259}
36260#endif
36261
36262#ifdef __LITTLE_ENDIAN__
36263__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36264  uint16x8_t __ret;
36265  __ret = (uint16x8_t)(__p0);
36266  return __ret;
36267}
36268#else
36269__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36270  uint16x8_t __ret;
36271  __ret = (uint16x8_t)(__p0);
36272  return __ret;
36273}
36274#endif
36275
36276#ifdef __LITTLE_ENDIAN__
36277__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36278  uint16x8_t __ret;
36279  __ret = (uint16x8_t)(__p0);
36280  return __ret;
36281}
36282#else
36283__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36284  uint16x8_t __ret;
36285  __ret = (uint16x8_t)(__p0);
36286  return __ret;
36287}
36288#endif
36289
36290#ifdef __LITTLE_ENDIAN__
36291__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36292  uint16x8_t __ret;
36293  __ret = (uint16x8_t)(__p0);
36294  return __ret;
36295}
36296#else
36297__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36298  uint16x8_t __ret;
36299  __ret = (uint16x8_t)(__p0);
36300  return __ret;
36301}
36302#endif
36303
36304#ifdef __LITTLE_ENDIAN__
36305__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36306  uint16x8_t __ret;
36307  __ret = (uint16x8_t)(__p0);
36308  return __ret;
36309}
36310#else
36311__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36312  uint16x8_t __ret;
36313  __ret = (uint16x8_t)(__p0);
36314  return __ret;
36315}
36316#endif
36317
36318#ifdef __LITTLE_ENDIAN__
36319__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36320  uint16x8_t __ret;
36321  __ret = (uint16x8_t)(__p0);
36322  return __ret;
36323}
36324#else
36325__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36326  uint16x8_t __ret;
36327  __ret = (uint16x8_t)(__p0);
36328  return __ret;
36329}
36330#endif
36331
36332#ifdef __LITTLE_ENDIAN__
36333__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36334  uint16x8_t __ret;
36335  __ret = (uint16x8_t)(__p0);
36336  return __ret;
36337}
36338#else
36339__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36340  uint16x8_t __ret;
36341  __ret = (uint16x8_t)(__p0);
36342  return __ret;
36343}
36344#endif
36345
36346#ifdef __LITTLE_ENDIAN__
36347__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36348  uint16x8_t __ret;
36349  __ret = (uint16x8_t)(__p0);
36350  return __ret;
36351}
36352#else
36353__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36354  uint16x8_t __ret;
36355  __ret = (uint16x8_t)(__p0);
36356  return __ret;
36357}
36358#endif
36359
36360#ifdef __LITTLE_ENDIAN__
36361__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36362  uint16x8_t __ret;
36363  __ret = (uint16x8_t)(__p0);
36364  return __ret;
36365}
36366#else
36367__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36368  uint16x8_t __ret;
36369  __ret = (uint16x8_t)(__p0);
36370  return __ret;
36371}
36372#endif
36373
36374#ifdef __LITTLE_ENDIAN__
36375__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36376  uint16x8_t __ret;
36377  __ret = (uint16x8_t)(__p0);
36378  return __ret;
36379}
36380#else
36381__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36382  uint16x8_t __ret;
36383  __ret = (uint16x8_t)(__p0);
36384  return __ret;
36385}
36386#endif
36387
36388#ifdef __LITTLE_ENDIAN__
36389__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36390  uint16x8_t __ret;
36391  __ret = (uint16x8_t)(__p0);
36392  return __ret;
36393}
36394#else
36395__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36396  uint16x8_t __ret;
36397  __ret = (uint16x8_t)(__p0);
36398  return __ret;
36399}
36400#endif
36401
36402#ifdef __LITTLE_ENDIAN__
36403__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36404  uint16x8_t __ret;
36405  __ret = (uint16x8_t)(__p0);
36406  return __ret;
36407}
36408#else
36409__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36410  uint16x8_t __ret;
36411  __ret = (uint16x8_t)(__p0);
36412  return __ret;
36413}
36414#endif
36415
36416#ifdef __LITTLE_ENDIAN__
36417__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36418  uint16x8_t __ret;
36419  __ret = (uint16x8_t)(__p0);
36420  return __ret;
36421}
36422#else
36423__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36424  uint16x8_t __ret;
36425  __ret = (uint16x8_t)(__p0);
36426  return __ret;
36427}
36428#endif
36429
36430#ifdef __LITTLE_ENDIAN__
36431__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36432  uint16x8_t __ret;
36433  __ret = (uint16x8_t)(__p0);
36434  return __ret;
36435}
36436#else
36437__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36438  uint16x8_t __ret;
36439  __ret = (uint16x8_t)(__p0);
36440  return __ret;
36441}
36442#endif
36443
36444#ifdef __LITTLE_ENDIAN__
36445__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36446  uint16x8_t __ret;
36447  __ret = (uint16x8_t)(__p0);
36448  return __ret;
36449}
36450#else
36451__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36452  uint16x8_t __ret;
36453  __ret = (uint16x8_t)(__p0);
36454  return __ret;
36455}
36456#endif
36457
36458#ifdef __LITTLE_ENDIAN__
36459__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36460  int8x16_t __ret;
36461  __ret = (int8x16_t)(__p0);
36462  return __ret;
36463}
36464#else
36465__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36466  int8x16_t __ret;
36467  __ret = (int8x16_t)(__p0);
36468  return __ret;
36469}
36470#endif
36471
36472#ifdef __LITTLE_ENDIAN__
36473__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36474  int8x16_t __ret;
36475  __ret = (int8x16_t)(__p0);
36476  return __ret;
36477}
36478#else
36479__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36480  int8x16_t __ret;
36481  __ret = (int8x16_t)(__p0);
36482  return __ret;
36483}
36484#endif
36485
36486#ifdef __LITTLE_ENDIAN__
36487__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36488  int8x16_t __ret;
36489  __ret = (int8x16_t)(__p0);
36490  return __ret;
36491}
36492#else
36493__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36494  int8x16_t __ret;
36495  __ret = (int8x16_t)(__p0);
36496  return __ret;
36497}
36498#endif
36499
36500#ifdef __LITTLE_ENDIAN__
36501__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36502  int8x16_t __ret;
36503  __ret = (int8x16_t)(__p0);
36504  return __ret;
36505}
36506#else
36507__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36508  int8x16_t __ret;
36509  __ret = (int8x16_t)(__p0);
36510  return __ret;
36511}
36512#endif
36513
36514#ifdef __LITTLE_ENDIAN__
36515__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36516  int8x16_t __ret;
36517  __ret = (int8x16_t)(__p0);
36518  return __ret;
36519}
36520#else
36521__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36522  int8x16_t __ret;
36523  __ret = (int8x16_t)(__p0);
36524  return __ret;
36525}
36526#endif
36527
36528#ifdef __LITTLE_ENDIAN__
36529__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36530  int8x16_t __ret;
36531  __ret = (int8x16_t)(__p0);
36532  return __ret;
36533}
36534#else
36535__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36536  int8x16_t __ret;
36537  __ret = (int8x16_t)(__p0);
36538  return __ret;
36539}
36540#endif
36541
36542#ifdef __LITTLE_ENDIAN__
36543__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36544  int8x16_t __ret;
36545  __ret = (int8x16_t)(__p0);
36546  return __ret;
36547}
36548#else
36549__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36550  int8x16_t __ret;
36551  __ret = (int8x16_t)(__p0);
36552  return __ret;
36553}
36554#endif
36555
36556#ifdef __LITTLE_ENDIAN__
36557__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36558  int8x16_t __ret;
36559  __ret = (int8x16_t)(__p0);
36560  return __ret;
36561}
36562#else
36563__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36564  int8x16_t __ret;
36565  __ret = (int8x16_t)(__p0);
36566  return __ret;
36567}
36568#endif
36569
36570#ifdef __LITTLE_ENDIAN__
36571__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36572  int8x16_t __ret;
36573  __ret = (int8x16_t)(__p0);
36574  return __ret;
36575}
36576#else
36577__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36578  int8x16_t __ret;
36579  __ret = (int8x16_t)(__p0);
36580  return __ret;
36581}
36582#endif
36583
36584#ifdef __LITTLE_ENDIAN__
36585__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36586  int8x16_t __ret;
36587  __ret = (int8x16_t)(__p0);
36588  return __ret;
36589}
36590#else
36591__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36592  int8x16_t __ret;
36593  __ret = (int8x16_t)(__p0);
36594  return __ret;
36595}
36596#endif
36597
36598#ifdef __LITTLE_ENDIAN__
36599__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36600  int8x16_t __ret;
36601  __ret = (int8x16_t)(__p0);
36602  return __ret;
36603}
36604#else
36605__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36606  int8x16_t __ret;
36607  __ret = (int8x16_t)(__p0);
36608  return __ret;
36609}
36610#endif
36611
36612#ifdef __LITTLE_ENDIAN__
36613__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36614  int8x16_t __ret;
36615  __ret = (int8x16_t)(__p0);
36616  return __ret;
36617}
36618#else
36619__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36620  int8x16_t __ret;
36621  __ret = (int8x16_t)(__p0);
36622  return __ret;
36623}
36624#endif
36625
36626#ifdef __LITTLE_ENDIAN__
36627__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36628  int8x16_t __ret;
36629  __ret = (int8x16_t)(__p0);
36630  return __ret;
36631}
36632#else
36633__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36634  int8x16_t __ret;
36635  __ret = (int8x16_t)(__p0);
36636  return __ret;
36637}
36638#endif
36639
36640#ifdef __LITTLE_ENDIAN__
36641__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36642  int8x16_t __ret;
36643  __ret = (int8x16_t)(__p0);
36644  return __ret;
36645}
36646#else
36647__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36648  int8x16_t __ret;
36649  __ret = (int8x16_t)(__p0);
36650  return __ret;
36651}
36652#endif
36653
36654#ifdef __LITTLE_ENDIAN__
36655__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36656  float64x2_t __ret;
36657  __ret = (float64x2_t)(__p0);
36658  return __ret;
36659}
36660#else
36661__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36662  float64x2_t __ret;
36663  __ret = (float64x2_t)(__p0);
36664  return __ret;
36665}
36666#endif
36667
36668#ifdef __LITTLE_ENDIAN__
36669__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36670  float64x2_t __ret;
36671  __ret = (float64x2_t)(__p0);
36672  return __ret;
36673}
36674#else
36675__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36676  float64x2_t __ret;
36677  __ret = (float64x2_t)(__p0);
36678  return __ret;
36679}
36680#endif
36681
36682#ifdef __LITTLE_ENDIAN__
36683__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36684  float64x2_t __ret;
36685  __ret = (float64x2_t)(__p0);
36686  return __ret;
36687}
36688#else
36689__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36690  float64x2_t __ret;
36691  __ret = (float64x2_t)(__p0);
36692  return __ret;
36693}
36694#endif
36695
36696#ifdef __LITTLE_ENDIAN__
36697__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36698  float64x2_t __ret;
36699  __ret = (float64x2_t)(__p0);
36700  return __ret;
36701}
36702#else
36703__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36704  float64x2_t __ret;
36705  __ret = (float64x2_t)(__p0);
36706  return __ret;
36707}
36708#endif
36709
36710#ifdef __LITTLE_ENDIAN__
36711__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36712  float64x2_t __ret;
36713  __ret = (float64x2_t)(__p0);
36714  return __ret;
36715}
36716#else
36717__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36718  float64x2_t __ret;
36719  __ret = (float64x2_t)(__p0);
36720  return __ret;
36721}
36722#endif
36723
36724#ifdef __LITTLE_ENDIAN__
36725__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36726  float64x2_t __ret;
36727  __ret = (float64x2_t)(__p0);
36728  return __ret;
36729}
36730#else
36731__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36732  float64x2_t __ret;
36733  __ret = (float64x2_t)(__p0);
36734  return __ret;
36735}
36736#endif
36737
36738#ifdef __LITTLE_ENDIAN__
36739__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36740  float64x2_t __ret;
36741  __ret = (float64x2_t)(__p0);
36742  return __ret;
36743}
36744#else
36745__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36746  float64x2_t __ret;
36747  __ret = (float64x2_t)(__p0);
36748  return __ret;
36749}
36750#endif
36751
36752#ifdef __LITTLE_ENDIAN__
36753__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36754  float64x2_t __ret;
36755  __ret = (float64x2_t)(__p0);
36756  return __ret;
36757}
36758#else
36759__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36760  float64x2_t __ret;
36761  __ret = (float64x2_t)(__p0);
36762  return __ret;
36763}
36764#endif
36765
36766#ifdef __LITTLE_ENDIAN__
36767__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36768  float64x2_t __ret;
36769  __ret = (float64x2_t)(__p0);
36770  return __ret;
36771}
36772#else
36773__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36774  float64x2_t __ret;
36775  __ret = (float64x2_t)(__p0);
36776  return __ret;
36777}
36778#endif
36779
36780#ifdef __LITTLE_ENDIAN__
36781__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36782  float64x2_t __ret;
36783  __ret = (float64x2_t)(__p0);
36784  return __ret;
36785}
36786#else
36787__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36788  float64x2_t __ret;
36789  __ret = (float64x2_t)(__p0);
36790  return __ret;
36791}
36792#endif
36793
36794#ifdef __LITTLE_ENDIAN__
36795__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36796  float64x2_t __ret;
36797  __ret = (float64x2_t)(__p0);
36798  return __ret;
36799}
36800#else
36801__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36802  float64x2_t __ret;
36803  __ret = (float64x2_t)(__p0);
36804  return __ret;
36805}
36806#endif
36807
36808#ifdef __LITTLE_ENDIAN__
36809__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36810  float64x2_t __ret;
36811  __ret = (float64x2_t)(__p0);
36812  return __ret;
36813}
36814#else
36815__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36816  float64x2_t __ret;
36817  __ret = (float64x2_t)(__p0);
36818  return __ret;
36819}
36820#endif
36821
36822#ifdef __LITTLE_ENDIAN__
36823__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36824  float64x2_t __ret;
36825  __ret = (float64x2_t)(__p0);
36826  return __ret;
36827}
36828#else
36829__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36830  float64x2_t __ret;
36831  __ret = (float64x2_t)(__p0);
36832  return __ret;
36833}
36834#endif
36835
36836#ifdef __LITTLE_ENDIAN__
36837__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36838  float64x2_t __ret;
36839  __ret = (float64x2_t)(__p0);
36840  return __ret;
36841}
36842#else
36843__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36844  float64x2_t __ret;
36845  __ret = (float64x2_t)(__p0);
36846  return __ret;
36847}
36848#endif
36849
36850#ifdef __LITTLE_ENDIAN__
36851__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36852  float32x4_t __ret;
36853  __ret = (float32x4_t)(__p0);
36854  return __ret;
36855}
36856#else
36857__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36858  float32x4_t __ret;
36859  __ret = (float32x4_t)(__p0);
36860  return __ret;
36861}
36862#endif
36863
36864#ifdef __LITTLE_ENDIAN__
36865__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36866  float32x4_t __ret;
36867  __ret = (float32x4_t)(__p0);
36868  return __ret;
36869}
36870#else
36871__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36872  float32x4_t __ret;
36873  __ret = (float32x4_t)(__p0);
36874  return __ret;
36875}
36876#endif
36877
36878#ifdef __LITTLE_ENDIAN__
36879__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36880  float32x4_t __ret;
36881  __ret = (float32x4_t)(__p0);
36882  return __ret;
36883}
36884#else
36885__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36886  float32x4_t __ret;
36887  __ret = (float32x4_t)(__p0);
36888  return __ret;
36889}
36890#endif
36891
36892#ifdef __LITTLE_ENDIAN__
36893__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36894  float32x4_t __ret;
36895  __ret = (float32x4_t)(__p0);
36896  return __ret;
36897}
36898#else
36899__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36900  float32x4_t __ret;
36901  __ret = (float32x4_t)(__p0);
36902  return __ret;
36903}
36904#endif
36905
36906#ifdef __LITTLE_ENDIAN__
36907__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36908  float32x4_t __ret;
36909  __ret = (float32x4_t)(__p0);
36910  return __ret;
36911}
36912#else
36913__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36914  float32x4_t __ret;
36915  __ret = (float32x4_t)(__p0);
36916  return __ret;
36917}
36918#endif
36919
36920#ifdef __LITTLE_ENDIAN__
36921__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36922  float32x4_t __ret;
36923  __ret = (float32x4_t)(__p0);
36924  return __ret;
36925}
36926#else
36927__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36928  float32x4_t __ret;
36929  __ret = (float32x4_t)(__p0);
36930  return __ret;
36931}
36932#endif
36933
36934#ifdef __LITTLE_ENDIAN__
36935__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36936  float32x4_t __ret;
36937  __ret = (float32x4_t)(__p0);
36938  return __ret;
36939}
36940#else
36941__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36942  float32x4_t __ret;
36943  __ret = (float32x4_t)(__p0);
36944  return __ret;
36945}
36946#endif
36947
36948#ifdef __LITTLE_ENDIAN__
36949__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36950  float32x4_t __ret;
36951  __ret = (float32x4_t)(__p0);
36952  return __ret;
36953}
36954#else
36955__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36956  float32x4_t __ret;
36957  __ret = (float32x4_t)(__p0);
36958  return __ret;
36959}
36960#endif
36961
36962#ifdef __LITTLE_ENDIAN__
36963__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
36964  float32x4_t __ret;
36965  __ret = (float32x4_t)(__p0);
36966  return __ret;
36967}
36968#else
36969__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
36970  float32x4_t __ret;
36971  __ret = (float32x4_t)(__p0);
36972  return __ret;
36973}
36974#endif
36975
36976#ifdef __LITTLE_ENDIAN__
36977__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
36978  float32x4_t __ret;
36979  __ret = (float32x4_t)(__p0);
36980  return __ret;
36981}
36982#else
36983__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
36984  float32x4_t __ret;
36985  __ret = (float32x4_t)(__p0);
36986  return __ret;
36987}
36988#endif
36989
36990#ifdef __LITTLE_ENDIAN__
36991__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
36992  float32x4_t __ret;
36993  __ret = (float32x4_t)(__p0);
36994  return __ret;
36995}
36996#else
36997__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
36998  float32x4_t __ret;
36999  __ret = (float32x4_t)(__p0);
37000  return __ret;
37001}
37002#endif
37003
37004#ifdef __LITTLE_ENDIAN__
37005__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37006  float32x4_t __ret;
37007  __ret = (float32x4_t)(__p0);
37008  return __ret;
37009}
37010#else
37011__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37012  float32x4_t __ret;
37013  __ret = (float32x4_t)(__p0);
37014  return __ret;
37015}
37016#endif
37017
37018#ifdef __LITTLE_ENDIAN__
37019__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37020  float32x4_t __ret;
37021  __ret = (float32x4_t)(__p0);
37022  return __ret;
37023}
37024#else
37025__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37026  float32x4_t __ret;
37027  __ret = (float32x4_t)(__p0);
37028  return __ret;
37029}
37030#endif
37031
37032#ifdef __LITTLE_ENDIAN__
37033__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37034  float32x4_t __ret;
37035  __ret = (float32x4_t)(__p0);
37036  return __ret;
37037}
37038#else
37039__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37040  float32x4_t __ret;
37041  __ret = (float32x4_t)(__p0);
37042  return __ret;
37043}
37044#endif
37045
37046#ifdef __LITTLE_ENDIAN__
37047__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37048  float16x8_t __ret;
37049  __ret = (float16x8_t)(__p0);
37050  return __ret;
37051}
37052#else
37053__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37054  float16x8_t __ret;
37055  __ret = (float16x8_t)(__p0);
37056  return __ret;
37057}
37058#endif
37059
37060#ifdef __LITTLE_ENDIAN__
37061__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37062  float16x8_t __ret;
37063  __ret = (float16x8_t)(__p0);
37064  return __ret;
37065}
37066#else
37067__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37068  float16x8_t __ret;
37069  __ret = (float16x8_t)(__p0);
37070  return __ret;
37071}
37072#endif
37073
37074#ifdef __LITTLE_ENDIAN__
37075__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37076  float16x8_t __ret;
37077  __ret = (float16x8_t)(__p0);
37078  return __ret;
37079}
37080#else
37081__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37082  float16x8_t __ret;
37083  __ret = (float16x8_t)(__p0);
37084  return __ret;
37085}
37086#endif
37087
37088#ifdef __LITTLE_ENDIAN__
37089__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37090  float16x8_t __ret;
37091  __ret = (float16x8_t)(__p0);
37092  return __ret;
37093}
37094#else
37095__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37096  float16x8_t __ret;
37097  __ret = (float16x8_t)(__p0);
37098  return __ret;
37099}
37100#endif
37101
37102#ifdef __LITTLE_ENDIAN__
37103__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37104  float16x8_t __ret;
37105  __ret = (float16x8_t)(__p0);
37106  return __ret;
37107}
37108#else
37109__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37110  float16x8_t __ret;
37111  __ret = (float16x8_t)(__p0);
37112  return __ret;
37113}
37114#endif
37115
37116#ifdef __LITTLE_ENDIAN__
37117__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37118  float16x8_t __ret;
37119  __ret = (float16x8_t)(__p0);
37120  return __ret;
37121}
37122#else
37123__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37124  float16x8_t __ret;
37125  __ret = (float16x8_t)(__p0);
37126  return __ret;
37127}
37128#endif
37129
37130#ifdef __LITTLE_ENDIAN__
37131__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37132  float16x8_t __ret;
37133  __ret = (float16x8_t)(__p0);
37134  return __ret;
37135}
37136#else
37137__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37138  float16x8_t __ret;
37139  __ret = (float16x8_t)(__p0);
37140  return __ret;
37141}
37142#endif
37143
37144#ifdef __LITTLE_ENDIAN__
37145__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37146  float16x8_t __ret;
37147  __ret = (float16x8_t)(__p0);
37148  return __ret;
37149}
37150#else
37151__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37152  float16x8_t __ret;
37153  __ret = (float16x8_t)(__p0);
37154  return __ret;
37155}
37156#endif
37157
37158#ifdef __LITTLE_ENDIAN__
37159__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37160  float16x8_t __ret;
37161  __ret = (float16x8_t)(__p0);
37162  return __ret;
37163}
37164#else
37165__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37166  float16x8_t __ret;
37167  __ret = (float16x8_t)(__p0);
37168  return __ret;
37169}
37170#endif
37171
37172#ifdef __LITTLE_ENDIAN__
37173__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37174  float16x8_t __ret;
37175  __ret = (float16x8_t)(__p0);
37176  return __ret;
37177}
37178#else
37179__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37180  float16x8_t __ret;
37181  __ret = (float16x8_t)(__p0);
37182  return __ret;
37183}
37184#endif
37185
37186#ifdef __LITTLE_ENDIAN__
37187__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37188  float16x8_t __ret;
37189  __ret = (float16x8_t)(__p0);
37190  return __ret;
37191}
37192#else
37193__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37194  float16x8_t __ret;
37195  __ret = (float16x8_t)(__p0);
37196  return __ret;
37197}
37198#endif
37199
37200#ifdef __LITTLE_ENDIAN__
37201__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37202  float16x8_t __ret;
37203  __ret = (float16x8_t)(__p0);
37204  return __ret;
37205}
37206#else
37207__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37208  float16x8_t __ret;
37209  __ret = (float16x8_t)(__p0);
37210  return __ret;
37211}
37212#endif
37213
37214#ifdef __LITTLE_ENDIAN__
37215__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37216  float16x8_t __ret;
37217  __ret = (float16x8_t)(__p0);
37218  return __ret;
37219}
37220#else
37221__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37222  float16x8_t __ret;
37223  __ret = (float16x8_t)(__p0);
37224  return __ret;
37225}
37226#endif
37227
37228#ifdef __LITTLE_ENDIAN__
37229__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37230  float16x8_t __ret;
37231  __ret = (float16x8_t)(__p0);
37232  return __ret;
37233}
37234#else
37235__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37236  float16x8_t __ret;
37237  __ret = (float16x8_t)(__p0);
37238  return __ret;
37239}
37240#endif
37241
37242#ifdef __LITTLE_ENDIAN__
37243__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37244  int32x4_t __ret;
37245  __ret = (int32x4_t)(__p0);
37246  return __ret;
37247}
37248#else
37249__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37250  int32x4_t __ret;
37251  __ret = (int32x4_t)(__p0);
37252  return __ret;
37253}
37254#endif
37255
37256#ifdef __LITTLE_ENDIAN__
37257__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37258  int32x4_t __ret;
37259  __ret = (int32x4_t)(__p0);
37260  return __ret;
37261}
37262#else
37263__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37264  int32x4_t __ret;
37265  __ret = (int32x4_t)(__p0);
37266  return __ret;
37267}
37268#endif
37269
37270#ifdef __LITTLE_ENDIAN__
37271__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37272  int32x4_t __ret;
37273  __ret = (int32x4_t)(__p0);
37274  return __ret;
37275}
37276#else
37277__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37278  int32x4_t __ret;
37279  __ret = (int32x4_t)(__p0);
37280  return __ret;
37281}
37282#endif
37283
37284#ifdef __LITTLE_ENDIAN__
37285__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37286  int32x4_t __ret;
37287  __ret = (int32x4_t)(__p0);
37288  return __ret;
37289}
37290#else
37291__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37292  int32x4_t __ret;
37293  __ret = (int32x4_t)(__p0);
37294  return __ret;
37295}
37296#endif
37297
37298#ifdef __LITTLE_ENDIAN__
37299__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37300  int32x4_t __ret;
37301  __ret = (int32x4_t)(__p0);
37302  return __ret;
37303}
37304#else
37305__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37306  int32x4_t __ret;
37307  __ret = (int32x4_t)(__p0);
37308  return __ret;
37309}
37310#endif
37311
37312#ifdef __LITTLE_ENDIAN__
37313__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37314  int32x4_t __ret;
37315  __ret = (int32x4_t)(__p0);
37316  return __ret;
37317}
37318#else
37319__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37320  int32x4_t __ret;
37321  __ret = (int32x4_t)(__p0);
37322  return __ret;
37323}
37324#endif
37325
37326#ifdef __LITTLE_ENDIAN__
37327__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37328  int32x4_t __ret;
37329  __ret = (int32x4_t)(__p0);
37330  return __ret;
37331}
37332#else
37333__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37334  int32x4_t __ret;
37335  __ret = (int32x4_t)(__p0);
37336  return __ret;
37337}
37338#endif
37339
37340#ifdef __LITTLE_ENDIAN__
37341__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37342  int32x4_t __ret;
37343  __ret = (int32x4_t)(__p0);
37344  return __ret;
37345}
37346#else
37347__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37348  int32x4_t __ret;
37349  __ret = (int32x4_t)(__p0);
37350  return __ret;
37351}
37352#endif
37353
37354#ifdef __LITTLE_ENDIAN__
37355__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37356  int32x4_t __ret;
37357  __ret = (int32x4_t)(__p0);
37358  return __ret;
37359}
37360#else
37361__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37362  int32x4_t __ret;
37363  __ret = (int32x4_t)(__p0);
37364  return __ret;
37365}
37366#endif
37367
37368#ifdef __LITTLE_ENDIAN__
37369__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37370  int32x4_t __ret;
37371  __ret = (int32x4_t)(__p0);
37372  return __ret;
37373}
37374#else
37375__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37376  int32x4_t __ret;
37377  __ret = (int32x4_t)(__p0);
37378  return __ret;
37379}
37380#endif
37381
37382#ifdef __LITTLE_ENDIAN__
37383__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37384  int32x4_t __ret;
37385  __ret = (int32x4_t)(__p0);
37386  return __ret;
37387}
37388#else
37389__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37390  int32x4_t __ret;
37391  __ret = (int32x4_t)(__p0);
37392  return __ret;
37393}
37394#endif
37395
37396#ifdef __LITTLE_ENDIAN__
37397__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37398  int32x4_t __ret;
37399  __ret = (int32x4_t)(__p0);
37400  return __ret;
37401}
37402#else
37403__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37404  int32x4_t __ret;
37405  __ret = (int32x4_t)(__p0);
37406  return __ret;
37407}
37408#endif
37409
37410#ifdef __LITTLE_ENDIAN__
37411__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37412  int32x4_t __ret;
37413  __ret = (int32x4_t)(__p0);
37414  return __ret;
37415}
37416#else
37417__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37418  int32x4_t __ret;
37419  __ret = (int32x4_t)(__p0);
37420  return __ret;
37421}
37422#endif
37423
37424#ifdef __LITTLE_ENDIAN__
37425__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37426  int32x4_t __ret;
37427  __ret = (int32x4_t)(__p0);
37428  return __ret;
37429}
37430#else
37431__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37432  int32x4_t __ret;
37433  __ret = (int32x4_t)(__p0);
37434  return __ret;
37435}
37436#endif
37437
37438#ifdef __LITTLE_ENDIAN__
37439__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37440  int64x2_t __ret;
37441  __ret = (int64x2_t)(__p0);
37442  return __ret;
37443}
37444#else
37445__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37446  int64x2_t __ret;
37447  __ret = (int64x2_t)(__p0);
37448  return __ret;
37449}
37450#endif
37451
37452#ifdef __LITTLE_ENDIAN__
37453__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37454  int64x2_t __ret;
37455  __ret = (int64x2_t)(__p0);
37456  return __ret;
37457}
37458#else
37459__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37460  int64x2_t __ret;
37461  __ret = (int64x2_t)(__p0);
37462  return __ret;
37463}
37464#endif
37465
37466#ifdef __LITTLE_ENDIAN__
37467__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37468  int64x2_t __ret;
37469  __ret = (int64x2_t)(__p0);
37470  return __ret;
37471}
37472#else
37473__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37474  int64x2_t __ret;
37475  __ret = (int64x2_t)(__p0);
37476  return __ret;
37477}
37478#endif
37479
37480#ifdef __LITTLE_ENDIAN__
37481__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37482  int64x2_t __ret;
37483  __ret = (int64x2_t)(__p0);
37484  return __ret;
37485}
37486#else
37487__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37488  int64x2_t __ret;
37489  __ret = (int64x2_t)(__p0);
37490  return __ret;
37491}
37492#endif
37493
37494#ifdef __LITTLE_ENDIAN__
37495__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37496  int64x2_t __ret;
37497  __ret = (int64x2_t)(__p0);
37498  return __ret;
37499}
37500#else
37501__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37502  int64x2_t __ret;
37503  __ret = (int64x2_t)(__p0);
37504  return __ret;
37505}
37506#endif
37507
37508#ifdef __LITTLE_ENDIAN__
37509__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37510  int64x2_t __ret;
37511  __ret = (int64x2_t)(__p0);
37512  return __ret;
37513}
37514#else
37515__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37516  int64x2_t __ret;
37517  __ret = (int64x2_t)(__p0);
37518  return __ret;
37519}
37520#endif
37521
37522#ifdef __LITTLE_ENDIAN__
37523__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37524  int64x2_t __ret;
37525  __ret = (int64x2_t)(__p0);
37526  return __ret;
37527}
37528#else
37529__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37530  int64x2_t __ret;
37531  __ret = (int64x2_t)(__p0);
37532  return __ret;
37533}
37534#endif
37535
37536#ifdef __LITTLE_ENDIAN__
37537__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37538  int64x2_t __ret;
37539  __ret = (int64x2_t)(__p0);
37540  return __ret;
37541}
37542#else
37543__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37544  int64x2_t __ret;
37545  __ret = (int64x2_t)(__p0);
37546  return __ret;
37547}
37548#endif
37549
37550#ifdef __LITTLE_ENDIAN__
37551__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37552  int64x2_t __ret;
37553  __ret = (int64x2_t)(__p0);
37554  return __ret;
37555}
37556#else
37557__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37558  int64x2_t __ret;
37559  __ret = (int64x2_t)(__p0);
37560  return __ret;
37561}
37562#endif
37563
37564#ifdef __LITTLE_ENDIAN__
37565__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37566  int64x2_t __ret;
37567  __ret = (int64x2_t)(__p0);
37568  return __ret;
37569}
37570#else
37571__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37572  int64x2_t __ret;
37573  __ret = (int64x2_t)(__p0);
37574  return __ret;
37575}
37576#endif
37577
37578#ifdef __LITTLE_ENDIAN__
37579__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37580  int64x2_t __ret;
37581  __ret = (int64x2_t)(__p0);
37582  return __ret;
37583}
37584#else
37585__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37586  int64x2_t __ret;
37587  __ret = (int64x2_t)(__p0);
37588  return __ret;
37589}
37590#endif
37591
37592#ifdef __LITTLE_ENDIAN__
37593__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37594  int64x2_t __ret;
37595  __ret = (int64x2_t)(__p0);
37596  return __ret;
37597}
37598#else
37599__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37600  int64x2_t __ret;
37601  __ret = (int64x2_t)(__p0);
37602  return __ret;
37603}
37604#endif
37605
37606#ifdef __LITTLE_ENDIAN__
37607__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37608  int64x2_t __ret;
37609  __ret = (int64x2_t)(__p0);
37610  return __ret;
37611}
37612#else
37613__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37614  int64x2_t __ret;
37615  __ret = (int64x2_t)(__p0);
37616  return __ret;
37617}
37618#endif
37619
37620#ifdef __LITTLE_ENDIAN__
37621__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37622  int64x2_t __ret;
37623  __ret = (int64x2_t)(__p0);
37624  return __ret;
37625}
37626#else
37627__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37628  int64x2_t __ret;
37629  __ret = (int64x2_t)(__p0);
37630  return __ret;
37631}
37632#endif
37633
37634#ifdef __LITTLE_ENDIAN__
37635__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37636  int16x8_t __ret;
37637  __ret = (int16x8_t)(__p0);
37638  return __ret;
37639}
37640#else
37641__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37642  int16x8_t __ret;
37643  __ret = (int16x8_t)(__p0);
37644  return __ret;
37645}
37646#endif
37647
37648#ifdef __LITTLE_ENDIAN__
37649__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37650  int16x8_t __ret;
37651  __ret = (int16x8_t)(__p0);
37652  return __ret;
37653}
37654#else
37655__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37656  int16x8_t __ret;
37657  __ret = (int16x8_t)(__p0);
37658  return __ret;
37659}
37660#endif
37661
37662#ifdef __LITTLE_ENDIAN__
37663__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37664  int16x8_t __ret;
37665  __ret = (int16x8_t)(__p0);
37666  return __ret;
37667}
37668#else
37669__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37670  int16x8_t __ret;
37671  __ret = (int16x8_t)(__p0);
37672  return __ret;
37673}
37674#endif
37675
37676#ifdef __LITTLE_ENDIAN__
37677__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37678  int16x8_t __ret;
37679  __ret = (int16x8_t)(__p0);
37680  return __ret;
37681}
37682#else
37683__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37684  int16x8_t __ret;
37685  __ret = (int16x8_t)(__p0);
37686  return __ret;
37687}
37688#endif
37689
37690#ifdef __LITTLE_ENDIAN__
37691__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37692  int16x8_t __ret;
37693  __ret = (int16x8_t)(__p0);
37694  return __ret;
37695}
37696#else
37697__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37698  int16x8_t __ret;
37699  __ret = (int16x8_t)(__p0);
37700  return __ret;
37701}
37702#endif
37703
37704#ifdef __LITTLE_ENDIAN__
37705__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37706  int16x8_t __ret;
37707  __ret = (int16x8_t)(__p0);
37708  return __ret;
37709}
37710#else
37711__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37712  int16x8_t __ret;
37713  __ret = (int16x8_t)(__p0);
37714  return __ret;
37715}
37716#endif
37717
37718#ifdef __LITTLE_ENDIAN__
37719__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37720  int16x8_t __ret;
37721  __ret = (int16x8_t)(__p0);
37722  return __ret;
37723}
37724#else
37725__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37726  int16x8_t __ret;
37727  __ret = (int16x8_t)(__p0);
37728  return __ret;
37729}
37730#endif
37731
37732#ifdef __LITTLE_ENDIAN__
37733__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37734  int16x8_t __ret;
37735  __ret = (int16x8_t)(__p0);
37736  return __ret;
37737}
37738#else
37739__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37740  int16x8_t __ret;
37741  __ret = (int16x8_t)(__p0);
37742  return __ret;
37743}
37744#endif
37745
37746#ifdef __LITTLE_ENDIAN__
37747__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37748  int16x8_t __ret;
37749  __ret = (int16x8_t)(__p0);
37750  return __ret;
37751}
37752#else
37753__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37754  int16x8_t __ret;
37755  __ret = (int16x8_t)(__p0);
37756  return __ret;
37757}
37758#endif
37759
37760#ifdef __LITTLE_ENDIAN__
37761__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37762  int16x8_t __ret;
37763  __ret = (int16x8_t)(__p0);
37764  return __ret;
37765}
37766#else
37767__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37768  int16x8_t __ret;
37769  __ret = (int16x8_t)(__p0);
37770  return __ret;
37771}
37772#endif
37773
37774#ifdef __LITTLE_ENDIAN__
37775__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37776  int16x8_t __ret;
37777  __ret = (int16x8_t)(__p0);
37778  return __ret;
37779}
37780#else
37781__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37782  int16x8_t __ret;
37783  __ret = (int16x8_t)(__p0);
37784  return __ret;
37785}
37786#endif
37787
37788#ifdef __LITTLE_ENDIAN__
37789__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37790  int16x8_t __ret;
37791  __ret = (int16x8_t)(__p0);
37792  return __ret;
37793}
37794#else
37795__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37796  int16x8_t __ret;
37797  __ret = (int16x8_t)(__p0);
37798  return __ret;
37799}
37800#endif
37801
37802#ifdef __LITTLE_ENDIAN__
37803__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37804  int16x8_t __ret;
37805  __ret = (int16x8_t)(__p0);
37806  return __ret;
37807}
37808#else
37809__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37810  int16x8_t __ret;
37811  __ret = (int16x8_t)(__p0);
37812  return __ret;
37813}
37814#endif
37815
37816#ifdef __LITTLE_ENDIAN__
37817__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37818  int16x8_t __ret;
37819  __ret = (int16x8_t)(__p0);
37820  return __ret;
37821}
37822#else
37823__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37824  int16x8_t __ret;
37825  __ret = (int16x8_t)(__p0);
37826  return __ret;
37827}
37828#endif
37829
37830#ifdef __LITTLE_ENDIAN__
37831__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37832  uint8x8_t __ret;
37833  __ret = (uint8x8_t)(__p0);
37834  return __ret;
37835}
37836#else
37837__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37838  uint8x8_t __ret;
37839  __ret = (uint8x8_t)(__p0);
37840  return __ret;
37841}
37842#endif
37843
37844#ifdef __LITTLE_ENDIAN__
37845__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37846  uint8x8_t __ret;
37847  __ret = (uint8x8_t)(__p0);
37848  return __ret;
37849}
37850#else
37851__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37852  uint8x8_t __ret;
37853  __ret = (uint8x8_t)(__p0);
37854  return __ret;
37855}
37856#endif
37857
37858#ifdef __LITTLE_ENDIAN__
37859__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37860  uint8x8_t __ret;
37861  __ret = (uint8x8_t)(__p0);
37862  return __ret;
37863}
37864#else
37865__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37866  uint8x8_t __ret;
37867  __ret = (uint8x8_t)(__p0);
37868  return __ret;
37869}
37870#endif
37871
37872#ifdef __LITTLE_ENDIAN__
37873__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37874  uint8x8_t __ret;
37875  __ret = (uint8x8_t)(__p0);
37876  return __ret;
37877}
37878#else
37879__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37880  uint8x8_t __ret;
37881  __ret = (uint8x8_t)(__p0);
37882  return __ret;
37883}
37884#endif
37885
37886#ifdef __LITTLE_ENDIAN__
37887__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37888  uint8x8_t __ret;
37889  __ret = (uint8x8_t)(__p0);
37890  return __ret;
37891}
37892#else
37893__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37894  uint8x8_t __ret;
37895  __ret = (uint8x8_t)(__p0);
37896  return __ret;
37897}
37898#endif
37899
37900#ifdef __LITTLE_ENDIAN__
37901__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37902  uint8x8_t __ret;
37903  __ret = (uint8x8_t)(__p0);
37904  return __ret;
37905}
37906#else
37907__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37908  uint8x8_t __ret;
37909  __ret = (uint8x8_t)(__p0);
37910  return __ret;
37911}
37912#endif
37913
37914#ifdef __LITTLE_ENDIAN__
37915__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37916  uint8x8_t __ret;
37917  __ret = (uint8x8_t)(__p0);
37918  return __ret;
37919}
37920#else
37921__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37922  uint8x8_t __ret;
37923  __ret = (uint8x8_t)(__p0);
37924  return __ret;
37925}
37926#endif
37927
37928#ifdef __LITTLE_ENDIAN__
37929__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37930  uint8x8_t __ret;
37931  __ret = (uint8x8_t)(__p0);
37932  return __ret;
37933}
37934#else
37935__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37936  uint8x8_t __ret;
37937  __ret = (uint8x8_t)(__p0);
37938  return __ret;
37939}
37940#endif
37941
37942#ifdef __LITTLE_ENDIAN__
37943__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37944  uint8x8_t __ret;
37945  __ret = (uint8x8_t)(__p0);
37946  return __ret;
37947}
37948#else
37949__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37950  uint8x8_t __ret;
37951  __ret = (uint8x8_t)(__p0);
37952  return __ret;
37953}
37954#endif
37955
37956#ifdef __LITTLE_ENDIAN__
37957__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
37958  uint8x8_t __ret;
37959  __ret = (uint8x8_t)(__p0);
37960  return __ret;
37961}
37962#else
37963__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
37964  uint8x8_t __ret;
37965  __ret = (uint8x8_t)(__p0);
37966  return __ret;
37967}
37968#endif
37969
37970#ifdef __LITTLE_ENDIAN__
37971__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
37972  uint8x8_t __ret;
37973  __ret = (uint8x8_t)(__p0);
37974  return __ret;
37975}
37976#else
37977__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
37978  uint8x8_t __ret;
37979  __ret = (uint8x8_t)(__p0);
37980  return __ret;
37981}
37982#endif
37983
37984#ifdef __LITTLE_ENDIAN__
37985__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
37986  uint8x8_t __ret;
37987  __ret = (uint8x8_t)(__p0);
37988  return __ret;
37989}
37990#else
37991__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
37992  uint8x8_t __ret;
37993  __ret = (uint8x8_t)(__p0);
37994  return __ret;
37995}
37996#endif
37997
37998#ifdef __LITTLE_ENDIAN__
37999__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38000  uint8x8_t __ret;
38001  __ret = (uint8x8_t)(__p0);
38002  return __ret;
38003}
38004#else
38005__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38006  uint8x8_t __ret;
38007  __ret = (uint8x8_t)(__p0);
38008  return __ret;
38009}
38010#endif
38011
38012#ifdef __LITTLE_ENDIAN__
38013__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38014  uint32x2_t __ret;
38015  __ret = (uint32x2_t)(__p0);
38016  return __ret;
38017}
38018#else
38019__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38020  uint32x2_t __ret;
38021  __ret = (uint32x2_t)(__p0);
38022  return __ret;
38023}
38024#endif
38025
38026#ifdef __LITTLE_ENDIAN__
38027__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38028  uint32x2_t __ret;
38029  __ret = (uint32x2_t)(__p0);
38030  return __ret;
38031}
38032#else
38033__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38034  uint32x2_t __ret;
38035  __ret = (uint32x2_t)(__p0);
38036  return __ret;
38037}
38038#endif
38039
38040#ifdef __LITTLE_ENDIAN__
38041__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38042  uint32x2_t __ret;
38043  __ret = (uint32x2_t)(__p0);
38044  return __ret;
38045}
38046#else
38047__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38048  uint32x2_t __ret;
38049  __ret = (uint32x2_t)(__p0);
38050  return __ret;
38051}
38052#endif
38053
38054#ifdef __LITTLE_ENDIAN__
38055__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38056  uint32x2_t __ret;
38057  __ret = (uint32x2_t)(__p0);
38058  return __ret;
38059}
38060#else
38061__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38062  uint32x2_t __ret;
38063  __ret = (uint32x2_t)(__p0);
38064  return __ret;
38065}
38066#endif
38067
38068#ifdef __LITTLE_ENDIAN__
38069__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38070  uint32x2_t __ret;
38071  __ret = (uint32x2_t)(__p0);
38072  return __ret;
38073}
38074#else
38075__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38076  uint32x2_t __ret;
38077  __ret = (uint32x2_t)(__p0);
38078  return __ret;
38079}
38080#endif
38081
38082#ifdef __LITTLE_ENDIAN__
38083__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38084  uint32x2_t __ret;
38085  __ret = (uint32x2_t)(__p0);
38086  return __ret;
38087}
38088#else
38089__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38090  uint32x2_t __ret;
38091  __ret = (uint32x2_t)(__p0);
38092  return __ret;
38093}
38094#endif
38095
38096#ifdef __LITTLE_ENDIAN__
38097__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38098  uint32x2_t __ret;
38099  __ret = (uint32x2_t)(__p0);
38100  return __ret;
38101}
38102#else
38103__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38104  uint32x2_t __ret;
38105  __ret = (uint32x2_t)(__p0);
38106  return __ret;
38107}
38108#endif
38109
38110#ifdef __LITTLE_ENDIAN__
38111__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38112  uint32x2_t __ret;
38113  __ret = (uint32x2_t)(__p0);
38114  return __ret;
38115}
38116#else
38117__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38118  uint32x2_t __ret;
38119  __ret = (uint32x2_t)(__p0);
38120  return __ret;
38121}
38122#endif
38123
38124#ifdef __LITTLE_ENDIAN__
38125__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38126  uint32x2_t __ret;
38127  __ret = (uint32x2_t)(__p0);
38128  return __ret;
38129}
38130#else
38131__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38132  uint32x2_t __ret;
38133  __ret = (uint32x2_t)(__p0);
38134  return __ret;
38135}
38136#endif
38137
38138#ifdef __LITTLE_ENDIAN__
38139__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38140  uint32x2_t __ret;
38141  __ret = (uint32x2_t)(__p0);
38142  return __ret;
38143}
38144#else
38145__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38146  uint32x2_t __ret;
38147  __ret = (uint32x2_t)(__p0);
38148  return __ret;
38149}
38150#endif
38151
38152#ifdef __LITTLE_ENDIAN__
38153__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38154  uint32x2_t __ret;
38155  __ret = (uint32x2_t)(__p0);
38156  return __ret;
38157}
38158#else
38159__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38160  uint32x2_t __ret;
38161  __ret = (uint32x2_t)(__p0);
38162  return __ret;
38163}
38164#endif
38165
38166#ifdef __LITTLE_ENDIAN__
38167__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38168  uint32x2_t __ret;
38169  __ret = (uint32x2_t)(__p0);
38170  return __ret;
38171}
38172#else
38173__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38174  uint32x2_t __ret;
38175  __ret = (uint32x2_t)(__p0);
38176  return __ret;
38177}
38178#endif
38179
38180#ifdef __LITTLE_ENDIAN__
38181__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38182  uint32x2_t __ret;
38183  __ret = (uint32x2_t)(__p0);
38184  return __ret;
38185}
38186#else
38187__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38188  uint32x2_t __ret;
38189  __ret = (uint32x2_t)(__p0);
38190  return __ret;
38191}
38192#endif
38193
38194#ifdef __LITTLE_ENDIAN__
38195__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38196  uint64x1_t __ret;
38197  __ret = (uint64x1_t)(__p0);
38198  return __ret;
38199}
38200#else
38201__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38202  uint64x1_t __ret;
38203  __ret = (uint64x1_t)(__p0);
38204  return __ret;
38205}
38206#endif
38207
38208#ifdef __LITTLE_ENDIAN__
38209__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38210  uint64x1_t __ret;
38211  __ret = (uint64x1_t)(__p0);
38212  return __ret;
38213}
38214#else
38215__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38216  uint64x1_t __ret;
38217  __ret = (uint64x1_t)(__p0);
38218  return __ret;
38219}
38220#endif
38221
38222#ifdef __LITTLE_ENDIAN__
38223__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38224  uint64x1_t __ret;
38225  __ret = (uint64x1_t)(__p0);
38226  return __ret;
38227}
38228#else
38229__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38230  uint64x1_t __ret;
38231  __ret = (uint64x1_t)(__p0);
38232  return __ret;
38233}
38234#endif
38235
38236#ifdef __LITTLE_ENDIAN__
38237__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38238  uint64x1_t __ret;
38239  __ret = (uint64x1_t)(__p0);
38240  return __ret;
38241}
38242#else
38243__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38244  uint64x1_t __ret;
38245  __ret = (uint64x1_t)(__p0);
38246  return __ret;
38247}
38248#endif
38249
38250#ifdef __LITTLE_ENDIAN__
38251__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38252  uint64x1_t __ret;
38253  __ret = (uint64x1_t)(__p0);
38254  return __ret;
38255}
38256#else
38257__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38258  uint64x1_t __ret;
38259  __ret = (uint64x1_t)(__p0);
38260  return __ret;
38261}
38262#endif
38263
38264#ifdef __LITTLE_ENDIAN__
38265__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38266  uint64x1_t __ret;
38267  __ret = (uint64x1_t)(__p0);
38268  return __ret;
38269}
38270#else
38271__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38272  uint64x1_t __ret;
38273  __ret = (uint64x1_t)(__p0);
38274  return __ret;
38275}
38276#endif
38277
38278#ifdef __LITTLE_ENDIAN__
38279__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38280  uint64x1_t __ret;
38281  __ret = (uint64x1_t)(__p0);
38282  return __ret;
38283}
38284#else
38285__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38286  uint64x1_t __ret;
38287  __ret = (uint64x1_t)(__p0);
38288  return __ret;
38289}
38290#endif
38291
38292#ifdef __LITTLE_ENDIAN__
38293__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38294  uint64x1_t __ret;
38295  __ret = (uint64x1_t)(__p0);
38296  return __ret;
38297}
38298#else
38299__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38300  uint64x1_t __ret;
38301  __ret = (uint64x1_t)(__p0);
38302  return __ret;
38303}
38304#endif
38305
38306#ifdef __LITTLE_ENDIAN__
38307__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38308  uint64x1_t __ret;
38309  __ret = (uint64x1_t)(__p0);
38310  return __ret;
38311}
38312#else
38313__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38314  uint64x1_t __ret;
38315  __ret = (uint64x1_t)(__p0);
38316  return __ret;
38317}
38318#endif
38319
38320#ifdef __LITTLE_ENDIAN__
38321__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38322  uint64x1_t __ret;
38323  __ret = (uint64x1_t)(__p0);
38324  return __ret;
38325}
38326#else
38327__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38328  uint64x1_t __ret;
38329  __ret = (uint64x1_t)(__p0);
38330  return __ret;
38331}
38332#endif
38333
38334#ifdef __LITTLE_ENDIAN__
38335__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38336  uint64x1_t __ret;
38337  __ret = (uint64x1_t)(__p0);
38338  return __ret;
38339}
38340#else
38341__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38342  uint64x1_t __ret;
38343  __ret = (uint64x1_t)(__p0);
38344  return __ret;
38345}
38346#endif
38347
38348#ifdef __LITTLE_ENDIAN__
38349__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38350  uint64x1_t __ret;
38351  __ret = (uint64x1_t)(__p0);
38352  return __ret;
38353}
38354#else
38355__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38356  uint64x1_t __ret;
38357  __ret = (uint64x1_t)(__p0);
38358  return __ret;
38359}
38360#endif
38361
38362#ifdef __LITTLE_ENDIAN__
38363__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38364  uint64x1_t __ret;
38365  __ret = (uint64x1_t)(__p0);
38366  return __ret;
38367}
38368#else
38369__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38370  uint64x1_t __ret;
38371  __ret = (uint64x1_t)(__p0);
38372  return __ret;
38373}
38374#endif
38375
38376#ifdef __LITTLE_ENDIAN__
38377__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38378  uint16x4_t __ret;
38379  __ret = (uint16x4_t)(__p0);
38380  return __ret;
38381}
38382#else
38383__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38384  uint16x4_t __ret;
38385  __ret = (uint16x4_t)(__p0);
38386  return __ret;
38387}
38388#endif
38389
38390#ifdef __LITTLE_ENDIAN__
38391__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38392  uint16x4_t __ret;
38393  __ret = (uint16x4_t)(__p0);
38394  return __ret;
38395}
38396#else
38397__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38398  uint16x4_t __ret;
38399  __ret = (uint16x4_t)(__p0);
38400  return __ret;
38401}
38402#endif
38403
38404#ifdef __LITTLE_ENDIAN__
38405__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38406  uint16x4_t __ret;
38407  __ret = (uint16x4_t)(__p0);
38408  return __ret;
38409}
38410#else
38411__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38412  uint16x4_t __ret;
38413  __ret = (uint16x4_t)(__p0);
38414  return __ret;
38415}
38416#endif
38417
38418#ifdef __LITTLE_ENDIAN__
38419__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38420  uint16x4_t __ret;
38421  __ret = (uint16x4_t)(__p0);
38422  return __ret;
38423}
38424#else
38425__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38426  uint16x4_t __ret;
38427  __ret = (uint16x4_t)(__p0);
38428  return __ret;
38429}
38430#endif
38431
38432#ifdef __LITTLE_ENDIAN__
38433__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38434  uint16x4_t __ret;
38435  __ret = (uint16x4_t)(__p0);
38436  return __ret;
38437}
38438#else
38439__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38440  uint16x4_t __ret;
38441  __ret = (uint16x4_t)(__p0);
38442  return __ret;
38443}
38444#endif
38445
38446#ifdef __LITTLE_ENDIAN__
38447__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38448  uint16x4_t __ret;
38449  __ret = (uint16x4_t)(__p0);
38450  return __ret;
38451}
38452#else
38453__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38454  uint16x4_t __ret;
38455  __ret = (uint16x4_t)(__p0);
38456  return __ret;
38457}
38458#endif
38459
38460#ifdef __LITTLE_ENDIAN__
38461__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38462  uint16x4_t __ret;
38463  __ret = (uint16x4_t)(__p0);
38464  return __ret;
38465}
38466#else
38467__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38468  uint16x4_t __ret;
38469  __ret = (uint16x4_t)(__p0);
38470  return __ret;
38471}
38472#endif
38473
38474#ifdef __LITTLE_ENDIAN__
38475__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38476  uint16x4_t __ret;
38477  __ret = (uint16x4_t)(__p0);
38478  return __ret;
38479}
38480#else
38481__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38482  uint16x4_t __ret;
38483  __ret = (uint16x4_t)(__p0);
38484  return __ret;
38485}
38486#endif
38487
38488#ifdef __LITTLE_ENDIAN__
38489__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38490  uint16x4_t __ret;
38491  __ret = (uint16x4_t)(__p0);
38492  return __ret;
38493}
38494#else
38495__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38496  uint16x4_t __ret;
38497  __ret = (uint16x4_t)(__p0);
38498  return __ret;
38499}
38500#endif
38501
38502#ifdef __LITTLE_ENDIAN__
38503__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38504  uint16x4_t __ret;
38505  __ret = (uint16x4_t)(__p0);
38506  return __ret;
38507}
38508#else
38509__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38510  uint16x4_t __ret;
38511  __ret = (uint16x4_t)(__p0);
38512  return __ret;
38513}
38514#endif
38515
38516#ifdef __LITTLE_ENDIAN__
38517__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38518  uint16x4_t __ret;
38519  __ret = (uint16x4_t)(__p0);
38520  return __ret;
38521}
38522#else
38523__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38524  uint16x4_t __ret;
38525  __ret = (uint16x4_t)(__p0);
38526  return __ret;
38527}
38528#endif
38529
38530#ifdef __LITTLE_ENDIAN__
38531__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38532  uint16x4_t __ret;
38533  __ret = (uint16x4_t)(__p0);
38534  return __ret;
38535}
38536#else
38537__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38538  uint16x4_t __ret;
38539  __ret = (uint16x4_t)(__p0);
38540  return __ret;
38541}
38542#endif
38543
38544#ifdef __LITTLE_ENDIAN__
38545__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38546  uint16x4_t __ret;
38547  __ret = (uint16x4_t)(__p0);
38548  return __ret;
38549}
38550#else
38551__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38552  uint16x4_t __ret;
38553  __ret = (uint16x4_t)(__p0);
38554  return __ret;
38555}
38556#endif
38557
38558#ifdef __LITTLE_ENDIAN__
38559__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38560  int8x8_t __ret;
38561  __ret = (int8x8_t)(__p0);
38562  return __ret;
38563}
38564#else
38565__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38566  int8x8_t __ret;
38567  __ret = (int8x8_t)(__p0);
38568  return __ret;
38569}
38570#endif
38571
38572#ifdef __LITTLE_ENDIAN__
38573__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38574  int8x8_t __ret;
38575  __ret = (int8x8_t)(__p0);
38576  return __ret;
38577}
38578#else
38579__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38580  int8x8_t __ret;
38581  __ret = (int8x8_t)(__p0);
38582  return __ret;
38583}
38584#endif
38585
38586#ifdef __LITTLE_ENDIAN__
38587__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38588  int8x8_t __ret;
38589  __ret = (int8x8_t)(__p0);
38590  return __ret;
38591}
38592#else
38593__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38594  int8x8_t __ret;
38595  __ret = (int8x8_t)(__p0);
38596  return __ret;
38597}
38598#endif
38599
38600#ifdef __LITTLE_ENDIAN__
38601__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38602  int8x8_t __ret;
38603  __ret = (int8x8_t)(__p0);
38604  return __ret;
38605}
38606#else
38607__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38608  int8x8_t __ret;
38609  __ret = (int8x8_t)(__p0);
38610  return __ret;
38611}
38612#endif
38613
38614#ifdef __LITTLE_ENDIAN__
38615__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38616  int8x8_t __ret;
38617  __ret = (int8x8_t)(__p0);
38618  return __ret;
38619}
38620#else
38621__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38622  int8x8_t __ret;
38623  __ret = (int8x8_t)(__p0);
38624  return __ret;
38625}
38626#endif
38627
38628#ifdef __LITTLE_ENDIAN__
38629__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38630  int8x8_t __ret;
38631  __ret = (int8x8_t)(__p0);
38632  return __ret;
38633}
38634#else
38635__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38636  int8x8_t __ret;
38637  __ret = (int8x8_t)(__p0);
38638  return __ret;
38639}
38640#endif
38641
38642#ifdef __LITTLE_ENDIAN__
38643__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38644  int8x8_t __ret;
38645  __ret = (int8x8_t)(__p0);
38646  return __ret;
38647}
38648#else
38649__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38650  int8x8_t __ret;
38651  __ret = (int8x8_t)(__p0);
38652  return __ret;
38653}
38654#endif
38655
38656#ifdef __LITTLE_ENDIAN__
38657__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38658  int8x8_t __ret;
38659  __ret = (int8x8_t)(__p0);
38660  return __ret;
38661}
38662#else
38663__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38664  int8x8_t __ret;
38665  __ret = (int8x8_t)(__p0);
38666  return __ret;
38667}
38668#endif
38669
38670#ifdef __LITTLE_ENDIAN__
38671__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38672  int8x8_t __ret;
38673  __ret = (int8x8_t)(__p0);
38674  return __ret;
38675}
38676#else
38677__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38678  int8x8_t __ret;
38679  __ret = (int8x8_t)(__p0);
38680  return __ret;
38681}
38682#endif
38683
38684#ifdef __LITTLE_ENDIAN__
38685__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38686  int8x8_t __ret;
38687  __ret = (int8x8_t)(__p0);
38688  return __ret;
38689}
38690#else
38691__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38692  int8x8_t __ret;
38693  __ret = (int8x8_t)(__p0);
38694  return __ret;
38695}
38696#endif
38697
38698#ifdef __LITTLE_ENDIAN__
38699__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38700  int8x8_t __ret;
38701  __ret = (int8x8_t)(__p0);
38702  return __ret;
38703}
38704#else
38705__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38706  int8x8_t __ret;
38707  __ret = (int8x8_t)(__p0);
38708  return __ret;
38709}
38710#endif
38711
38712#ifdef __LITTLE_ENDIAN__
38713__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38714  int8x8_t __ret;
38715  __ret = (int8x8_t)(__p0);
38716  return __ret;
38717}
38718#else
38719__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38720  int8x8_t __ret;
38721  __ret = (int8x8_t)(__p0);
38722  return __ret;
38723}
38724#endif
38725
38726#ifdef __LITTLE_ENDIAN__
38727__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38728  int8x8_t __ret;
38729  __ret = (int8x8_t)(__p0);
38730  return __ret;
38731}
38732#else
38733__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38734  int8x8_t __ret;
38735  __ret = (int8x8_t)(__p0);
38736  return __ret;
38737}
38738#endif
38739
38740#ifdef __LITTLE_ENDIAN__
38741__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38742  float64x1_t __ret;
38743  __ret = (float64x1_t)(__p0);
38744  return __ret;
38745}
38746#else
38747__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38748  float64x1_t __ret;
38749  __ret = (float64x1_t)(__p0);
38750  return __ret;
38751}
38752#endif
38753
38754#ifdef __LITTLE_ENDIAN__
38755__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38756  float64x1_t __ret;
38757  __ret = (float64x1_t)(__p0);
38758  return __ret;
38759}
38760#else
38761__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38762  float64x1_t __ret;
38763  __ret = (float64x1_t)(__p0);
38764  return __ret;
38765}
38766#endif
38767
38768#ifdef __LITTLE_ENDIAN__
38769__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38770  float64x1_t __ret;
38771  __ret = (float64x1_t)(__p0);
38772  return __ret;
38773}
38774#else
38775__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38776  float64x1_t __ret;
38777  __ret = (float64x1_t)(__p0);
38778  return __ret;
38779}
38780#endif
38781
38782#ifdef __LITTLE_ENDIAN__
38783__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38784  float64x1_t __ret;
38785  __ret = (float64x1_t)(__p0);
38786  return __ret;
38787}
38788#else
38789__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38790  float64x1_t __ret;
38791  __ret = (float64x1_t)(__p0);
38792  return __ret;
38793}
38794#endif
38795
38796#ifdef __LITTLE_ENDIAN__
38797__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38798  float64x1_t __ret;
38799  __ret = (float64x1_t)(__p0);
38800  return __ret;
38801}
38802#else
38803__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38804  float64x1_t __ret;
38805  __ret = (float64x1_t)(__p0);
38806  return __ret;
38807}
38808#endif
38809
38810#ifdef __LITTLE_ENDIAN__
38811__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38812  float64x1_t __ret;
38813  __ret = (float64x1_t)(__p0);
38814  return __ret;
38815}
38816#else
38817__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38818  float64x1_t __ret;
38819  __ret = (float64x1_t)(__p0);
38820  return __ret;
38821}
38822#endif
38823
38824#ifdef __LITTLE_ENDIAN__
38825__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38826  float64x1_t __ret;
38827  __ret = (float64x1_t)(__p0);
38828  return __ret;
38829}
38830#else
38831__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38832  float64x1_t __ret;
38833  __ret = (float64x1_t)(__p0);
38834  return __ret;
38835}
38836#endif
38837
38838#ifdef __LITTLE_ENDIAN__
38839__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38840  float64x1_t __ret;
38841  __ret = (float64x1_t)(__p0);
38842  return __ret;
38843}
38844#else
38845__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38846  float64x1_t __ret;
38847  __ret = (float64x1_t)(__p0);
38848  return __ret;
38849}
38850#endif
38851
38852#ifdef __LITTLE_ENDIAN__
38853__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38854  float64x1_t __ret;
38855  __ret = (float64x1_t)(__p0);
38856  return __ret;
38857}
38858#else
38859__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38860  float64x1_t __ret;
38861  __ret = (float64x1_t)(__p0);
38862  return __ret;
38863}
38864#endif
38865
38866#ifdef __LITTLE_ENDIAN__
38867__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38868  float64x1_t __ret;
38869  __ret = (float64x1_t)(__p0);
38870  return __ret;
38871}
38872#else
38873__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38874  float64x1_t __ret;
38875  __ret = (float64x1_t)(__p0);
38876  return __ret;
38877}
38878#endif
38879
38880#ifdef __LITTLE_ENDIAN__
38881__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38882  float64x1_t __ret;
38883  __ret = (float64x1_t)(__p0);
38884  return __ret;
38885}
38886#else
38887__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38888  float64x1_t __ret;
38889  __ret = (float64x1_t)(__p0);
38890  return __ret;
38891}
38892#endif
38893
38894#ifdef __LITTLE_ENDIAN__
38895__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38896  float64x1_t __ret;
38897  __ret = (float64x1_t)(__p0);
38898  return __ret;
38899}
38900#else
38901__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38902  float64x1_t __ret;
38903  __ret = (float64x1_t)(__p0);
38904  return __ret;
38905}
38906#endif
38907
38908#ifdef __LITTLE_ENDIAN__
38909__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38910  float64x1_t __ret;
38911  __ret = (float64x1_t)(__p0);
38912  return __ret;
38913}
38914#else
38915__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38916  float64x1_t __ret;
38917  __ret = (float64x1_t)(__p0);
38918  return __ret;
38919}
38920#endif
38921
38922#ifdef __LITTLE_ENDIAN__
38923__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38924  float32x2_t __ret;
38925  __ret = (float32x2_t)(__p0);
38926  return __ret;
38927}
38928#else
38929__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38930  float32x2_t __ret;
38931  __ret = (float32x2_t)(__p0);
38932  return __ret;
38933}
38934#endif
38935
38936#ifdef __LITTLE_ENDIAN__
38937__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38938  float32x2_t __ret;
38939  __ret = (float32x2_t)(__p0);
38940  return __ret;
38941}
38942#else
38943__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38944  float32x2_t __ret;
38945  __ret = (float32x2_t)(__p0);
38946  return __ret;
38947}
38948#endif
38949
38950#ifdef __LITTLE_ENDIAN__
38951__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38952  float32x2_t __ret;
38953  __ret = (float32x2_t)(__p0);
38954  return __ret;
38955}
38956#else
38957__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38958  float32x2_t __ret;
38959  __ret = (float32x2_t)(__p0);
38960  return __ret;
38961}
38962#endif
38963
38964#ifdef __LITTLE_ENDIAN__
38965__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
38966  float32x2_t __ret;
38967  __ret = (float32x2_t)(__p0);
38968  return __ret;
38969}
38970#else
38971__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
38972  float32x2_t __ret;
38973  __ret = (float32x2_t)(__p0);
38974  return __ret;
38975}
38976#endif
38977
38978#ifdef __LITTLE_ENDIAN__
38979__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
38980  float32x2_t __ret;
38981  __ret = (float32x2_t)(__p0);
38982  return __ret;
38983}
38984#else
38985__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
38986  float32x2_t __ret;
38987  __ret = (float32x2_t)(__p0);
38988  return __ret;
38989}
38990#endif
38991
38992#ifdef __LITTLE_ENDIAN__
38993__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
38994  float32x2_t __ret;
38995  __ret = (float32x2_t)(__p0);
38996  return __ret;
38997}
38998#else
38999__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
39000  float32x2_t __ret;
39001  __ret = (float32x2_t)(__p0);
39002  return __ret;
39003}
39004#endif
39005
39006#ifdef __LITTLE_ENDIAN__
39007__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39008  float32x2_t __ret;
39009  __ret = (float32x2_t)(__p0);
39010  return __ret;
39011}
39012#else
39013__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39014  float32x2_t __ret;
39015  __ret = (float32x2_t)(__p0);
39016  return __ret;
39017}
39018#endif
39019
39020#ifdef __LITTLE_ENDIAN__
39021__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39022  float32x2_t __ret;
39023  __ret = (float32x2_t)(__p0);
39024  return __ret;
39025}
39026#else
39027__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39028  float32x2_t __ret;
39029  __ret = (float32x2_t)(__p0);
39030  return __ret;
39031}
39032#endif
39033
39034#ifdef __LITTLE_ENDIAN__
39035__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39036  float32x2_t __ret;
39037  __ret = (float32x2_t)(__p0);
39038  return __ret;
39039}
39040#else
39041__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39042  float32x2_t __ret;
39043  __ret = (float32x2_t)(__p0);
39044  return __ret;
39045}
39046#endif
39047
39048#ifdef __LITTLE_ENDIAN__
39049__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39050  float32x2_t __ret;
39051  __ret = (float32x2_t)(__p0);
39052  return __ret;
39053}
39054#else
39055__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39056  float32x2_t __ret;
39057  __ret = (float32x2_t)(__p0);
39058  return __ret;
39059}
39060#endif
39061
39062#ifdef __LITTLE_ENDIAN__
39063__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39064  float32x2_t __ret;
39065  __ret = (float32x2_t)(__p0);
39066  return __ret;
39067}
39068#else
39069__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39070  float32x2_t __ret;
39071  __ret = (float32x2_t)(__p0);
39072  return __ret;
39073}
39074#endif
39075
39076#ifdef __LITTLE_ENDIAN__
39077__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39078  float32x2_t __ret;
39079  __ret = (float32x2_t)(__p0);
39080  return __ret;
39081}
39082#else
39083__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39084  float32x2_t __ret;
39085  __ret = (float32x2_t)(__p0);
39086  return __ret;
39087}
39088#endif
39089
39090#ifdef __LITTLE_ENDIAN__
39091__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39092  float32x2_t __ret;
39093  __ret = (float32x2_t)(__p0);
39094  return __ret;
39095}
39096#else
39097__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39098  float32x2_t __ret;
39099  __ret = (float32x2_t)(__p0);
39100  return __ret;
39101}
39102#endif
39103
39104#ifdef __LITTLE_ENDIAN__
39105__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39106  float16x4_t __ret;
39107  __ret = (float16x4_t)(__p0);
39108  return __ret;
39109}
39110#else
39111__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39112  float16x4_t __ret;
39113  __ret = (float16x4_t)(__p0);
39114  return __ret;
39115}
39116#endif
39117
39118#ifdef __LITTLE_ENDIAN__
39119__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39120  float16x4_t __ret;
39121  __ret = (float16x4_t)(__p0);
39122  return __ret;
39123}
39124#else
39125__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39126  float16x4_t __ret;
39127  __ret = (float16x4_t)(__p0);
39128  return __ret;
39129}
39130#endif
39131
39132#ifdef __LITTLE_ENDIAN__
39133__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39134  float16x4_t __ret;
39135  __ret = (float16x4_t)(__p0);
39136  return __ret;
39137}
39138#else
39139__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39140  float16x4_t __ret;
39141  __ret = (float16x4_t)(__p0);
39142  return __ret;
39143}
39144#endif
39145
39146#ifdef __LITTLE_ENDIAN__
39147__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39148  float16x4_t __ret;
39149  __ret = (float16x4_t)(__p0);
39150  return __ret;
39151}
39152#else
39153__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39154  float16x4_t __ret;
39155  __ret = (float16x4_t)(__p0);
39156  return __ret;
39157}
39158#endif
39159
39160#ifdef __LITTLE_ENDIAN__
39161__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39162  float16x4_t __ret;
39163  __ret = (float16x4_t)(__p0);
39164  return __ret;
39165}
39166#else
39167__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39168  float16x4_t __ret;
39169  __ret = (float16x4_t)(__p0);
39170  return __ret;
39171}
39172#endif
39173
39174#ifdef __LITTLE_ENDIAN__
39175__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39176  float16x4_t __ret;
39177  __ret = (float16x4_t)(__p0);
39178  return __ret;
39179}
39180#else
39181__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39182  float16x4_t __ret;
39183  __ret = (float16x4_t)(__p0);
39184  return __ret;
39185}
39186#endif
39187
39188#ifdef __LITTLE_ENDIAN__
39189__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39190  float16x4_t __ret;
39191  __ret = (float16x4_t)(__p0);
39192  return __ret;
39193}
39194#else
39195__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39196  float16x4_t __ret;
39197  __ret = (float16x4_t)(__p0);
39198  return __ret;
39199}
39200#endif
39201
39202#ifdef __LITTLE_ENDIAN__
39203__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39204  float16x4_t __ret;
39205  __ret = (float16x4_t)(__p0);
39206  return __ret;
39207}
39208#else
39209__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39210  float16x4_t __ret;
39211  __ret = (float16x4_t)(__p0);
39212  return __ret;
39213}
39214#endif
39215
39216#ifdef __LITTLE_ENDIAN__
39217__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39218  float16x4_t __ret;
39219  __ret = (float16x4_t)(__p0);
39220  return __ret;
39221}
39222#else
39223__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39224  float16x4_t __ret;
39225  __ret = (float16x4_t)(__p0);
39226  return __ret;
39227}
39228#endif
39229
39230#ifdef __LITTLE_ENDIAN__
39231__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39232  float16x4_t __ret;
39233  __ret = (float16x4_t)(__p0);
39234  return __ret;
39235}
39236#else
39237__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39238  float16x4_t __ret;
39239  __ret = (float16x4_t)(__p0);
39240  return __ret;
39241}
39242#endif
39243
39244#ifdef __LITTLE_ENDIAN__
39245__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39246  float16x4_t __ret;
39247  __ret = (float16x4_t)(__p0);
39248  return __ret;
39249}
39250#else
39251__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39252  float16x4_t __ret;
39253  __ret = (float16x4_t)(__p0);
39254  return __ret;
39255}
39256#endif
39257
39258#ifdef __LITTLE_ENDIAN__
39259__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39260  float16x4_t __ret;
39261  __ret = (float16x4_t)(__p0);
39262  return __ret;
39263}
39264#else
39265__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39266  float16x4_t __ret;
39267  __ret = (float16x4_t)(__p0);
39268  return __ret;
39269}
39270#endif
39271
39272#ifdef __LITTLE_ENDIAN__
39273__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39274  float16x4_t __ret;
39275  __ret = (float16x4_t)(__p0);
39276  return __ret;
39277}
39278#else
39279__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39280  float16x4_t __ret;
39281  __ret = (float16x4_t)(__p0);
39282  return __ret;
39283}
39284#endif
39285
39286#ifdef __LITTLE_ENDIAN__
39287__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39288  int32x2_t __ret;
39289  __ret = (int32x2_t)(__p0);
39290  return __ret;
39291}
39292#else
39293__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39294  int32x2_t __ret;
39295  __ret = (int32x2_t)(__p0);
39296  return __ret;
39297}
39298#endif
39299
39300#ifdef __LITTLE_ENDIAN__
39301__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39302  int32x2_t __ret;
39303  __ret = (int32x2_t)(__p0);
39304  return __ret;
39305}
39306#else
39307__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39308  int32x2_t __ret;
39309  __ret = (int32x2_t)(__p0);
39310  return __ret;
39311}
39312#endif
39313
39314#ifdef __LITTLE_ENDIAN__
39315__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39316  int32x2_t __ret;
39317  __ret = (int32x2_t)(__p0);
39318  return __ret;
39319}
39320#else
39321__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39322  int32x2_t __ret;
39323  __ret = (int32x2_t)(__p0);
39324  return __ret;
39325}
39326#endif
39327
39328#ifdef __LITTLE_ENDIAN__
39329__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39330  int32x2_t __ret;
39331  __ret = (int32x2_t)(__p0);
39332  return __ret;
39333}
39334#else
39335__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39336  int32x2_t __ret;
39337  __ret = (int32x2_t)(__p0);
39338  return __ret;
39339}
39340#endif
39341
39342#ifdef __LITTLE_ENDIAN__
39343__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39344  int32x2_t __ret;
39345  __ret = (int32x2_t)(__p0);
39346  return __ret;
39347}
39348#else
39349__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39350  int32x2_t __ret;
39351  __ret = (int32x2_t)(__p0);
39352  return __ret;
39353}
39354#endif
39355
39356#ifdef __LITTLE_ENDIAN__
39357__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39358  int32x2_t __ret;
39359  __ret = (int32x2_t)(__p0);
39360  return __ret;
39361}
39362#else
39363__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39364  int32x2_t __ret;
39365  __ret = (int32x2_t)(__p0);
39366  return __ret;
39367}
39368#endif
39369
39370#ifdef __LITTLE_ENDIAN__
39371__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39372  int32x2_t __ret;
39373  __ret = (int32x2_t)(__p0);
39374  return __ret;
39375}
39376#else
39377__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39378  int32x2_t __ret;
39379  __ret = (int32x2_t)(__p0);
39380  return __ret;
39381}
39382#endif
39383
39384#ifdef __LITTLE_ENDIAN__
39385__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39386  int32x2_t __ret;
39387  __ret = (int32x2_t)(__p0);
39388  return __ret;
39389}
39390#else
39391__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39392  int32x2_t __ret;
39393  __ret = (int32x2_t)(__p0);
39394  return __ret;
39395}
39396#endif
39397
39398#ifdef __LITTLE_ENDIAN__
39399__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39400  int32x2_t __ret;
39401  __ret = (int32x2_t)(__p0);
39402  return __ret;
39403}
39404#else
39405__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39406  int32x2_t __ret;
39407  __ret = (int32x2_t)(__p0);
39408  return __ret;
39409}
39410#endif
39411
39412#ifdef __LITTLE_ENDIAN__
39413__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39414  int32x2_t __ret;
39415  __ret = (int32x2_t)(__p0);
39416  return __ret;
39417}
39418#else
39419__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39420  int32x2_t __ret;
39421  __ret = (int32x2_t)(__p0);
39422  return __ret;
39423}
39424#endif
39425
39426#ifdef __LITTLE_ENDIAN__
39427__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39428  int32x2_t __ret;
39429  __ret = (int32x2_t)(__p0);
39430  return __ret;
39431}
39432#else
39433__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39434  int32x2_t __ret;
39435  __ret = (int32x2_t)(__p0);
39436  return __ret;
39437}
39438#endif
39439
39440#ifdef __LITTLE_ENDIAN__
39441__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39442  int32x2_t __ret;
39443  __ret = (int32x2_t)(__p0);
39444  return __ret;
39445}
39446#else
39447__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39448  int32x2_t __ret;
39449  __ret = (int32x2_t)(__p0);
39450  return __ret;
39451}
39452#endif
39453
39454#ifdef __LITTLE_ENDIAN__
39455__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39456  int32x2_t __ret;
39457  __ret = (int32x2_t)(__p0);
39458  return __ret;
39459}
39460#else
39461__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39462  int32x2_t __ret;
39463  __ret = (int32x2_t)(__p0);
39464  return __ret;
39465}
39466#endif
39467
39468#ifdef __LITTLE_ENDIAN__
39469__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39470  int64x1_t __ret;
39471  __ret = (int64x1_t)(__p0);
39472  return __ret;
39473}
39474#else
39475__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39476  int64x1_t __ret;
39477  __ret = (int64x1_t)(__p0);
39478  return __ret;
39479}
39480#endif
39481
39482#ifdef __LITTLE_ENDIAN__
39483__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39484  int64x1_t __ret;
39485  __ret = (int64x1_t)(__p0);
39486  return __ret;
39487}
39488#else
39489__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39490  int64x1_t __ret;
39491  __ret = (int64x1_t)(__p0);
39492  return __ret;
39493}
39494#endif
39495
39496#ifdef __LITTLE_ENDIAN__
39497__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39498  int64x1_t __ret;
39499  __ret = (int64x1_t)(__p0);
39500  return __ret;
39501}
39502#else
39503__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39504  int64x1_t __ret;
39505  __ret = (int64x1_t)(__p0);
39506  return __ret;
39507}
39508#endif
39509
39510#ifdef __LITTLE_ENDIAN__
39511__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39512  int64x1_t __ret;
39513  __ret = (int64x1_t)(__p0);
39514  return __ret;
39515}
39516#else
39517__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39518  int64x1_t __ret;
39519  __ret = (int64x1_t)(__p0);
39520  return __ret;
39521}
39522#endif
39523
39524#ifdef __LITTLE_ENDIAN__
39525__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39526  int64x1_t __ret;
39527  __ret = (int64x1_t)(__p0);
39528  return __ret;
39529}
39530#else
39531__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39532  int64x1_t __ret;
39533  __ret = (int64x1_t)(__p0);
39534  return __ret;
39535}
39536#endif
39537
39538#ifdef __LITTLE_ENDIAN__
39539__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39540  int64x1_t __ret;
39541  __ret = (int64x1_t)(__p0);
39542  return __ret;
39543}
39544#else
39545__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39546  int64x1_t __ret;
39547  __ret = (int64x1_t)(__p0);
39548  return __ret;
39549}
39550#endif
39551
39552#ifdef __LITTLE_ENDIAN__
39553__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39554  int64x1_t __ret;
39555  __ret = (int64x1_t)(__p0);
39556  return __ret;
39557}
39558#else
39559__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39560  int64x1_t __ret;
39561  __ret = (int64x1_t)(__p0);
39562  return __ret;
39563}
39564#endif
39565
39566#ifdef __LITTLE_ENDIAN__
39567__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39568  int64x1_t __ret;
39569  __ret = (int64x1_t)(__p0);
39570  return __ret;
39571}
39572#else
39573__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39574  int64x1_t __ret;
39575  __ret = (int64x1_t)(__p0);
39576  return __ret;
39577}
39578#endif
39579
39580#ifdef __LITTLE_ENDIAN__
39581__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39582  int64x1_t __ret;
39583  __ret = (int64x1_t)(__p0);
39584  return __ret;
39585}
39586#else
39587__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39588  int64x1_t __ret;
39589  __ret = (int64x1_t)(__p0);
39590  return __ret;
39591}
39592#endif
39593
39594#ifdef __LITTLE_ENDIAN__
39595__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39596  int64x1_t __ret;
39597  __ret = (int64x1_t)(__p0);
39598  return __ret;
39599}
39600#else
39601__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39602  int64x1_t __ret;
39603  __ret = (int64x1_t)(__p0);
39604  return __ret;
39605}
39606#endif
39607
39608#ifdef __LITTLE_ENDIAN__
39609__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39610  int64x1_t __ret;
39611  __ret = (int64x1_t)(__p0);
39612  return __ret;
39613}
39614#else
39615__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39616  int64x1_t __ret;
39617  __ret = (int64x1_t)(__p0);
39618  return __ret;
39619}
39620#endif
39621
39622#ifdef __LITTLE_ENDIAN__
39623__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39624  int64x1_t __ret;
39625  __ret = (int64x1_t)(__p0);
39626  return __ret;
39627}
39628#else
39629__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39630  int64x1_t __ret;
39631  __ret = (int64x1_t)(__p0);
39632  return __ret;
39633}
39634#endif
39635
39636#ifdef __LITTLE_ENDIAN__
39637__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39638  int64x1_t __ret;
39639  __ret = (int64x1_t)(__p0);
39640  return __ret;
39641}
39642#else
39643__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39644  int64x1_t __ret;
39645  __ret = (int64x1_t)(__p0);
39646  return __ret;
39647}
39648#endif
39649
39650#ifdef __LITTLE_ENDIAN__
39651__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39652  int16x4_t __ret;
39653  __ret = (int16x4_t)(__p0);
39654  return __ret;
39655}
39656#else
39657__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39658  int16x4_t __ret;
39659  __ret = (int16x4_t)(__p0);
39660  return __ret;
39661}
39662#endif
39663
39664#ifdef __LITTLE_ENDIAN__
39665__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39666  int16x4_t __ret;
39667  __ret = (int16x4_t)(__p0);
39668  return __ret;
39669}
39670#else
39671__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39672  int16x4_t __ret;
39673  __ret = (int16x4_t)(__p0);
39674  return __ret;
39675}
39676#endif
39677
39678#ifdef __LITTLE_ENDIAN__
39679__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39680  int16x4_t __ret;
39681  __ret = (int16x4_t)(__p0);
39682  return __ret;
39683}
39684#else
39685__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39686  int16x4_t __ret;
39687  __ret = (int16x4_t)(__p0);
39688  return __ret;
39689}
39690#endif
39691
39692#ifdef __LITTLE_ENDIAN__
39693__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39694  int16x4_t __ret;
39695  __ret = (int16x4_t)(__p0);
39696  return __ret;
39697}
39698#else
39699__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39700  int16x4_t __ret;
39701  __ret = (int16x4_t)(__p0);
39702  return __ret;
39703}
39704#endif
39705
39706#ifdef __LITTLE_ENDIAN__
39707__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39708  int16x4_t __ret;
39709  __ret = (int16x4_t)(__p0);
39710  return __ret;
39711}
39712#else
39713__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39714  int16x4_t __ret;
39715  __ret = (int16x4_t)(__p0);
39716  return __ret;
39717}
39718#endif
39719
39720#ifdef __LITTLE_ENDIAN__
39721__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39722  int16x4_t __ret;
39723  __ret = (int16x4_t)(__p0);
39724  return __ret;
39725}
39726#else
39727__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39728  int16x4_t __ret;
39729  __ret = (int16x4_t)(__p0);
39730  return __ret;
39731}
39732#endif
39733
39734#ifdef __LITTLE_ENDIAN__
39735__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39736  int16x4_t __ret;
39737  __ret = (int16x4_t)(__p0);
39738  return __ret;
39739}
39740#else
39741__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39742  int16x4_t __ret;
39743  __ret = (int16x4_t)(__p0);
39744  return __ret;
39745}
39746#endif
39747
39748#ifdef __LITTLE_ENDIAN__
39749__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39750  int16x4_t __ret;
39751  __ret = (int16x4_t)(__p0);
39752  return __ret;
39753}
39754#else
39755__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39756  int16x4_t __ret;
39757  __ret = (int16x4_t)(__p0);
39758  return __ret;
39759}
39760#endif
39761
39762#ifdef __LITTLE_ENDIAN__
39763__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39764  int16x4_t __ret;
39765  __ret = (int16x4_t)(__p0);
39766  return __ret;
39767}
39768#else
39769__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39770  int16x4_t __ret;
39771  __ret = (int16x4_t)(__p0);
39772  return __ret;
39773}
39774#endif
39775
39776#ifdef __LITTLE_ENDIAN__
39777__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39778  int16x4_t __ret;
39779  __ret = (int16x4_t)(__p0);
39780  return __ret;
39781}
39782#else
39783__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39784  int16x4_t __ret;
39785  __ret = (int16x4_t)(__p0);
39786  return __ret;
39787}
39788#endif
39789
39790#ifdef __LITTLE_ENDIAN__
39791__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39792  int16x4_t __ret;
39793  __ret = (int16x4_t)(__p0);
39794  return __ret;
39795}
39796#else
39797__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39798  int16x4_t __ret;
39799  __ret = (int16x4_t)(__p0);
39800  return __ret;
39801}
39802#endif
39803
39804#ifdef __LITTLE_ENDIAN__
39805__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39806  int16x4_t __ret;
39807  __ret = (int16x4_t)(__p0);
39808  return __ret;
39809}
39810#else
39811__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39812  int16x4_t __ret;
39813  __ret = (int16x4_t)(__p0);
39814  return __ret;
39815}
39816#endif
39817
39818#ifdef __LITTLE_ENDIAN__
39819__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39820  int16x4_t __ret;
39821  __ret = (int16x4_t)(__p0);
39822  return __ret;
39823}
39824#else
39825__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39826  int16x4_t __ret;
39827  __ret = (int16x4_t)(__p0);
39828  return __ret;
39829}
39830#endif
39831
39832#endif
39833#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
39834#ifdef __LITTLE_ENDIAN__
39835__ai float64x2_t vrndq_f64(float64x2_t __p0) {
39836  float64x2_t __ret;
39837  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
39838  return __ret;
39839}
39840#else
39841__ai float64x2_t vrndq_f64(float64x2_t __p0) {
39842  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39843  float64x2_t __ret;
39844  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
39845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39846  return __ret;
39847}
39848#endif
39849
39850#ifdef __LITTLE_ENDIAN__
39851__ai float64x1_t vrnd_f64(float64x1_t __p0) {
39852  float64x1_t __ret;
39853  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39854  return __ret;
39855}
39856#else
39857__ai float64x1_t vrnd_f64(float64x1_t __p0) {
39858  float64x1_t __ret;
39859  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39860  return __ret;
39861}
39862#endif
39863
39864#ifdef __LITTLE_ENDIAN__
39865__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39866  float64x2_t __ret;
39867  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
39868  return __ret;
39869}
39870#else
39871__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39872  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39873  float64x2_t __ret;
39874  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
39875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39876  return __ret;
39877}
39878#endif
39879
39880#ifdef __LITTLE_ENDIAN__
39881__ai float64x1_t vrnda_f64(float64x1_t __p0) {
39882  float64x1_t __ret;
39883  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39884  return __ret;
39885}
39886#else
39887__ai float64x1_t vrnda_f64(float64x1_t __p0) {
39888  float64x1_t __ret;
39889  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39890  return __ret;
39891}
39892#endif
39893
39894#ifdef __LITTLE_ENDIAN__
39895__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39896  float64x2_t __ret;
39897  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
39898  return __ret;
39899}
39900#else
39901__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39902  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39903  float64x2_t __ret;
39904  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
39905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39906  return __ret;
39907}
39908#endif
39909
39910#ifdef __LITTLE_ENDIAN__
39911__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39912  float32x4_t __ret;
39913  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
39914  return __ret;
39915}
39916#else
39917__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39918  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39919  float32x4_t __ret;
39920  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
39921  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39922  return __ret;
39923}
39924#endif
39925
39926#ifdef __LITTLE_ENDIAN__
39927__ai float64x1_t vrndi_f64(float64x1_t __p0) {
39928  float64x1_t __ret;
39929  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39930  return __ret;
39931}
39932#else
39933__ai float64x1_t vrndi_f64(float64x1_t __p0) {
39934  float64x1_t __ret;
39935  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39936  return __ret;
39937}
39938#endif
39939
39940#ifdef __LITTLE_ENDIAN__
39941__ai float32x2_t vrndi_f32(float32x2_t __p0) {
39942  float32x2_t __ret;
39943  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
39944  return __ret;
39945}
39946#else
39947__ai float32x2_t vrndi_f32(float32x2_t __p0) {
39948  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39949  float32x2_t __ret;
39950  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
39951  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39952  return __ret;
39953}
39954#endif
39955
39956#ifdef __LITTLE_ENDIAN__
39957__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
39958  float64x2_t __ret;
39959  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
39960  return __ret;
39961}
39962#else
39963__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
39964  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39965  float64x2_t __ret;
39966  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
39967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39968  return __ret;
39969}
39970#endif
39971
39972#ifdef __LITTLE_ENDIAN__
39973__ai float64x1_t vrndm_f64(float64x1_t __p0) {
39974  float64x1_t __ret;
39975  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
39976  return __ret;
39977}
39978#else
39979__ai float64x1_t vrndm_f64(float64x1_t __p0) {
39980  float64x1_t __ret;
39981  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
39982  return __ret;
39983}
39984#endif
39985
39986#ifdef __LITTLE_ENDIAN__
39987__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
39988  float64x2_t __ret;
39989  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
39990  return __ret;
39991}
39992#else
39993__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
39994  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39995  float64x2_t __ret;
39996  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
39997  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39998  return __ret;
39999}
40000#endif
40001
40002#ifdef __LITTLE_ENDIAN__
40003__ai float64x1_t vrndn_f64(float64x1_t __p0) {
40004  float64x1_t __ret;
40005  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40006  return __ret;
40007}
40008#else
40009__ai float64x1_t vrndn_f64(float64x1_t __p0) {
40010  float64x1_t __ret;
40011  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40012  return __ret;
40013}
40014#endif
40015
40016#ifdef __LITTLE_ENDIAN__
40017__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40018  float64x2_t __ret;
40019  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
40020  return __ret;
40021}
40022#else
40023__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40024  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40025  float64x2_t __ret;
40026  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
40027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40028  return __ret;
40029}
40030#endif
40031
40032#ifdef __LITTLE_ENDIAN__
40033__ai float64x1_t vrndp_f64(float64x1_t __p0) {
40034  float64x1_t __ret;
40035  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40036  return __ret;
40037}
40038#else
40039__ai float64x1_t vrndp_f64(float64x1_t __p0) {
40040  float64x1_t __ret;
40041  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40042  return __ret;
40043}
40044#endif
40045
40046#ifdef __LITTLE_ENDIAN__
40047__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40048  float64x2_t __ret;
40049  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
40050  return __ret;
40051}
40052#else
40053__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40054  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40055  float64x2_t __ret;
40056  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
40057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40058  return __ret;
40059}
40060#endif
40061
40062#ifdef __LITTLE_ENDIAN__
40063__ai float64x1_t vrndx_f64(float64x1_t __p0) {
40064  float64x1_t __ret;
40065  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40066  return __ret;
40067}
40068#else
40069__ai float64x1_t vrndx_f64(float64x1_t __p0) {
40070  float64x1_t __ret;
40071  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40072  return __ret;
40073}
40074#endif
40075
40076#endif
40077#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
40078#ifdef __LITTLE_ENDIAN__
40079__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40080  float64x2_t __ret;
40081  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40082  return __ret;
40083}
40084#else
40085__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40086  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40087  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40088  float64x2_t __ret;
40089  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40090  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40091  return __ret;
40092}
40093#endif
40094
40095#ifdef __LITTLE_ENDIAN__
40096__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40097  float64x1_t __ret;
40098  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40099  return __ret;
40100}
40101#else
40102__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40103  float64x1_t __ret;
40104  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40105  return __ret;
40106}
40107#endif
40108
40109#ifdef __LITTLE_ENDIAN__
40110__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40111  float64x2_t __ret;
40112  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40113  return __ret;
40114}
40115#else
40116__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40117  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40118  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40119  float64x2_t __ret;
40120  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40121  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40122  return __ret;
40123}
40124#endif
40125
40126#ifdef __LITTLE_ENDIAN__
40127__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40128  float64x1_t __ret;
40129  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40130  return __ret;
40131}
40132#else
40133__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40134  float64x1_t __ret;
40135  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40136  return __ret;
40137}
40138#endif
40139
40140#endif
40141#if __ARM_FEATURE_CRYPTO
40142#ifdef __LITTLE_ENDIAN__
40143__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40144  uint8x16_t __ret;
40145  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40146  return __ret;
40147}
40148#else
40149__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40150  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40151  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40152  uint8x16_t __ret;
40153  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40154  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40155  return __ret;
40156}
40157#endif
40158
40159#ifdef __LITTLE_ENDIAN__
40160__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40161  uint8x16_t __ret;
40162  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40163  return __ret;
40164}
40165#else
40166__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40167  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40168  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40169  uint8x16_t __ret;
40170  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40171  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40172  return __ret;
40173}
40174#endif
40175
40176#ifdef __LITTLE_ENDIAN__
40177__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40178  uint8x16_t __ret;
40179  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
40180  return __ret;
40181}
40182#else
40183__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40184  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40185  uint8x16_t __ret;
40186  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
40187  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40188  return __ret;
40189}
40190#endif
40191
40192#ifdef __LITTLE_ENDIAN__
40193__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40194  uint8x16_t __ret;
40195  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
40196  return __ret;
40197}
40198#else
40199__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40200  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40201  uint8x16_t __ret;
40202  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
40203  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40204  return __ret;
40205}
40206#endif
40207
40208#ifdef __LITTLE_ENDIAN__
40209__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40210  uint32x4_t __ret;
40211  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40212  return __ret;
40213}
40214#else
40215__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40216  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40217  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40218  uint32x4_t __ret;
40219  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40220  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40221  return __ret;
40222}
40223#endif
40224
40225#ifdef __LITTLE_ENDIAN__
40226__ai uint32_t vsha1h_u32(uint32_t __p0) {
40227  uint32_t __ret;
40228  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40229  return __ret;
40230}
40231#else
40232__ai uint32_t vsha1h_u32(uint32_t __p0) {
40233  uint32_t __ret;
40234  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40235  return __ret;
40236}
40237#endif
40238
40239#ifdef __LITTLE_ENDIAN__
40240__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40241  uint32x4_t __ret;
40242  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40243  return __ret;
40244}
40245#else
40246__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40247  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40248  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40249  uint32x4_t __ret;
40250  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40251  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40252  return __ret;
40253}
40254#endif
40255
40256#ifdef __LITTLE_ENDIAN__
40257__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40258  uint32x4_t __ret;
40259  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40260  return __ret;
40261}
40262#else
40263__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40264  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40265  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40266  uint32x4_t __ret;
40267  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40269  return __ret;
40270}
40271#endif
40272
40273#ifdef __LITTLE_ENDIAN__
40274__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40275  uint32x4_t __ret;
40276  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40277  return __ret;
40278}
40279#else
40280__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40281  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40282  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40283  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40284  uint32x4_t __ret;
40285  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40286  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40287  return __ret;
40288}
40289#endif
40290
40291#ifdef __LITTLE_ENDIAN__
40292__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40293  uint32x4_t __ret;
40294  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40295  return __ret;
40296}
40297#else
40298__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40299  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40300  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40301  uint32x4_t __ret;
40302  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40303  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40304  return __ret;
40305}
40306#endif
40307
40308#ifdef __LITTLE_ENDIAN__
40309__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40310  uint32x4_t __ret;
40311  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40312  return __ret;
40313}
40314#else
40315__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40316  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40317  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40318  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40319  uint32x4_t __ret;
40320  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40321  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40322  return __ret;
40323}
40324#endif
40325
40326#ifdef __LITTLE_ENDIAN__
40327__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40328  uint32x4_t __ret;
40329  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40330  return __ret;
40331}
40332#else
40333__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40334  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40335  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40336  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40337  uint32x4_t __ret;
40338  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40339  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40340  return __ret;
40341}
40342#endif
40343
40344#ifdef __LITTLE_ENDIAN__
40345__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40346  uint32x4_t __ret;
40347  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40348  return __ret;
40349}
40350#else
40351__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40352  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40353  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40354  uint32x4_t __ret;
40355  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40356  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40357  return __ret;
40358}
40359#endif
40360
40361#ifdef __LITTLE_ENDIAN__
40362__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40363  uint32x4_t __ret;
40364  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40365  return __ret;
40366}
40367#else
40368__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40369  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40370  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40371  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40372  uint32x4_t __ret;
40373  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40374  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40375  return __ret;
40376}
40377#endif
40378
40379#endif
40380#if defined(__ARM_FEATURE_FMA)
40381#ifdef __LITTLE_ENDIAN__
40382__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40383  float32x4_t __ret;
40384  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40385  return __ret;
40386}
40387#else
40388__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40389  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40390  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40391  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40392  float32x4_t __ret;
40393  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
40394  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40395  return __ret;
40396}
40397__ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40398  float32x4_t __ret;
40399  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40400  return __ret;
40401}
40402#endif
40403
40404#ifdef __LITTLE_ENDIAN__
40405__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40406  float32x2_t __ret;
40407  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
40408  return __ret;
40409}
40410#else
40411__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40412  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40413  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40414  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40415  float32x2_t __ret;
40416  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
40417  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40418  return __ret;
40419}
40420__ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40421  float32x2_t __ret;
40422  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
40423  return __ret;
40424}
40425#endif
40426
40427#ifdef __LITTLE_ENDIAN__
40428__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40429  float32x4_t __ret;
40430  __ret = vfmaq_f32(__p0, -__p1, __p2);
40431  return __ret;
40432}
40433#else
40434__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40435  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40436  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40437  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40438  float32x4_t __ret;
40439  __ret = __noswap_vfmaq_f32(__rev0, -__rev1, __rev2);
40440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40441  return __ret;
40442}
40443#endif
40444
40445#ifdef __LITTLE_ENDIAN__
40446__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40447  float32x2_t __ret;
40448  __ret = vfma_f32(__p0, -__p1, __p2);
40449  return __ret;
40450}
40451#else
40452__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40453  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40454  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40455  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40456  float32x2_t __ret;
40457  __ret = __noswap_vfma_f32(__rev0, -__rev1, __rev2);
40458  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40459  return __ret;
40460}
40461#endif
40462
40463#endif
40464#if defined(__ARM_FEATURE_QRDMX)
40465#ifdef __LITTLE_ENDIAN__
40466__ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40467  int32x4_t __ret;
40468  __ret = vqaddq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40469  return __ret;
40470}
40471#else
40472__ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40473  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40474  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40475  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40476  int32x4_t __ret;
40477  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40478  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40479  return __ret;
40480}
40481#endif
40482
40483#ifdef __LITTLE_ENDIAN__
40484__ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40485  int16x8_t __ret;
40486  __ret = vqaddq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40487  return __ret;
40488}
40489#else
40490__ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40491  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40492  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40493  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40494  int16x8_t __ret;
40495  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40496  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40497  return __ret;
40498}
40499#endif
40500
40501#ifdef __LITTLE_ENDIAN__
40502__ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40503  int32x2_t __ret;
40504  __ret = vqadd_s32(__p0, vqrdmulh_s32(__p1, __p2));
40505  return __ret;
40506}
40507#else
40508__ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40509  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40510  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40511  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40512  int32x2_t __ret;
40513  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40515  return __ret;
40516}
40517#endif
40518
40519#ifdef __LITTLE_ENDIAN__
40520__ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40521  int16x4_t __ret;
40522  __ret = vqadd_s16(__p0, vqrdmulh_s16(__p1, __p2));
40523  return __ret;
40524}
40525#else
40526__ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40527  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40528  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40529  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40530  int16x4_t __ret;
40531  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40533  return __ret;
40534}
40535#endif
40536
40537#ifdef __LITTLE_ENDIAN__
40538#define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40539  int32x4_t __s0 = __p0; \
40540  int32x4_t __s1 = __p1; \
40541  int32x2_t __s2 = __p2; \
40542  int32x4_t __ret; \
40543  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40544  __ret; \
40545})
40546#else
40547#define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40548  int32x4_t __s0 = __p0; \
40549  int32x4_t __s1 = __p1; \
40550  int32x2_t __s2 = __p2; \
40551  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40552  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40553  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40554  int32x4_t __ret; \
40555  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40556  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40557  __ret; \
40558})
40559#endif
40560
40561#ifdef __LITTLE_ENDIAN__
40562#define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40563  int16x8_t __s0 = __p0; \
40564  int16x8_t __s1 = __p1; \
40565  int16x4_t __s2 = __p2; \
40566  int16x8_t __ret; \
40567  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40568  __ret; \
40569})
40570#else
40571#define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40572  int16x8_t __s0 = __p0; \
40573  int16x8_t __s1 = __p1; \
40574  int16x4_t __s2 = __p2; \
40575  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40576  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40577  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40578  int16x8_t __ret; \
40579  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40581  __ret; \
40582})
40583#endif
40584
40585#ifdef __LITTLE_ENDIAN__
40586#define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40587  int32x2_t __s0 = __p0; \
40588  int32x2_t __s1 = __p1; \
40589  int32x2_t __s2 = __p2; \
40590  int32x2_t __ret; \
40591  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40592  __ret; \
40593})
40594#else
40595#define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40596  int32x2_t __s0 = __p0; \
40597  int32x2_t __s1 = __p1; \
40598  int32x2_t __s2 = __p2; \
40599  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40600  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40601  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40602  int32x2_t __ret; \
40603  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40605  __ret; \
40606})
40607#endif
40608
40609#ifdef __LITTLE_ENDIAN__
40610#define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40611  int16x4_t __s0 = __p0; \
40612  int16x4_t __s1 = __p1; \
40613  int16x4_t __s2 = __p2; \
40614  int16x4_t __ret; \
40615  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40616  __ret; \
40617})
40618#else
40619#define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40620  int16x4_t __s0 = __p0; \
40621  int16x4_t __s1 = __p1; \
40622  int16x4_t __s2 = __p2; \
40623  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40624  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40625  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40626  int16x4_t __ret; \
40627  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40628  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40629  __ret; \
40630})
40631#endif
40632
40633#ifdef __LITTLE_ENDIAN__
40634__ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40635  int32x4_t __ret;
40636  __ret = vqsubq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40637  return __ret;
40638}
40639#else
40640__ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40641  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40642  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40643  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40644  int32x4_t __ret;
40645  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40646  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40647  return __ret;
40648}
40649#endif
40650
40651#ifdef __LITTLE_ENDIAN__
40652__ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40653  int16x8_t __ret;
40654  __ret = vqsubq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40655  return __ret;
40656}
40657#else
40658__ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40659  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40660  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40661  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40662  int16x8_t __ret;
40663  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40665  return __ret;
40666}
40667#endif
40668
40669#ifdef __LITTLE_ENDIAN__
40670__ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40671  int32x2_t __ret;
40672  __ret = vqsub_s32(__p0, vqrdmulh_s32(__p1, __p2));
40673  return __ret;
40674}
40675#else
40676__ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40677  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40678  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40679  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40680  int32x2_t __ret;
40681  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40682  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40683  return __ret;
40684}
40685#endif
40686
40687#ifdef __LITTLE_ENDIAN__
40688__ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40689  int16x4_t __ret;
40690  __ret = vqsub_s16(__p0, vqrdmulh_s16(__p1, __p2));
40691  return __ret;
40692}
40693#else
40694__ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40695  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40696  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40697  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40698  int16x4_t __ret;
40699  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40700  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40701  return __ret;
40702}
40703#endif
40704
40705#ifdef __LITTLE_ENDIAN__
40706#define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40707  int32x4_t __s0 = __p0; \
40708  int32x4_t __s1 = __p1; \
40709  int32x2_t __s2 = __p2; \
40710  int32x4_t __ret; \
40711  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40712  __ret; \
40713})
40714#else
40715#define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40716  int32x4_t __s0 = __p0; \
40717  int32x4_t __s1 = __p1; \
40718  int32x2_t __s2 = __p2; \
40719  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40720  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40721  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40722  int32x4_t __ret; \
40723  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40724  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40725  __ret; \
40726})
40727#endif
40728
40729#ifdef __LITTLE_ENDIAN__
40730#define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40731  int16x8_t __s0 = __p0; \
40732  int16x8_t __s1 = __p1; \
40733  int16x4_t __s2 = __p2; \
40734  int16x8_t __ret; \
40735  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40736  __ret; \
40737})
40738#else
40739#define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40740  int16x8_t __s0 = __p0; \
40741  int16x8_t __s1 = __p1; \
40742  int16x4_t __s2 = __p2; \
40743  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40744  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40745  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40746  int16x8_t __ret; \
40747  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40749  __ret; \
40750})
40751#endif
40752
40753#ifdef __LITTLE_ENDIAN__
40754#define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40755  int32x2_t __s0 = __p0; \
40756  int32x2_t __s1 = __p1; \
40757  int32x2_t __s2 = __p2; \
40758  int32x2_t __ret; \
40759  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40760  __ret; \
40761})
40762#else
40763#define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40764  int32x2_t __s0 = __p0; \
40765  int32x2_t __s1 = __p1; \
40766  int32x2_t __s2 = __p2; \
40767  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40768  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40769  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40770  int32x2_t __ret; \
40771  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40772  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40773  __ret; \
40774})
40775#endif
40776
40777#ifdef __LITTLE_ENDIAN__
40778#define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40779  int16x4_t __s0 = __p0; \
40780  int16x4_t __s1 = __p1; \
40781  int16x4_t __s2 = __p2; \
40782  int16x4_t __ret; \
40783  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40784  __ret; \
40785})
40786#else
40787#define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40788  int16x4_t __s0 = __p0; \
40789  int16x4_t __s1 = __p1; \
40790  int16x4_t __s2 = __p2; \
40791  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40792  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40793  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40794  int16x4_t __ret; \
40795  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40796  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40797  __ret; \
40798})
40799#endif
40800
40801#endif
40802#if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
40803#ifdef __LITTLE_ENDIAN__
40804#define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40805  int32x4_t __s0 = __p0; \
40806  int32x4_t __s1 = __p1; \
40807  int32x4_t __s2 = __p2; \
40808  int32x4_t __ret; \
40809  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40810  __ret; \
40811})
40812#else
40813#define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40814  int32x4_t __s0 = __p0; \
40815  int32x4_t __s1 = __p1; \
40816  int32x4_t __s2 = __p2; \
40817  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40818  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40819  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40820  int32x4_t __ret; \
40821  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40822  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40823  __ret; \
40824})
40825#endif
40826
40827#ifdef __LITTLE_ENDIAN__
40828#define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40829  int16x8_t __s0 = __p0; \
40830  int16x8_t __s1 = __p1; \
40831  int16x8_t __s2 = __p2; \
40832  int16x8_t __ret; \
40833  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40834  __ret; \
40835})
40836#else
40837#define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40838  int16x8_t __s0 = __p0; \
40839  int16x8_t __s1 = __p1; \
40840  int16x8_t __s2 = __p2; \
40841  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40842  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40843  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40844  int16x8_t __ret; \
40845  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40846  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40847  __ret; \
40848})
40849#endif
40850
40851#ifdef __LITTLE_ENDIAN__
40852#define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40853  int32x2_t __s0 = __p0; \
40854  int32x2_t __s1 = __p1; \
40855  int32x4_t __s2 = __p2; \
40856  int32x2_t __ret; \
40857  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40858  __ret; \
40859})
40860#else
40861#define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40862  int32x2_t __s0 = __p0; \
40863  int32x2_t __s1 = __p1; \
40864  int32x4_t __s2 = __p2; \
40865  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40866  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40867  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40868  int32x2_t __ret; \
40869  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40870  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40871  __ret; \
40872})
40873#endif
40874
40875#ifdef __LITTLE_ENDIAN__
40876#define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40877  int16x4_t __s0 = __p0; \
40878  int16x4_t __s1 = __p1; \
40879  int16x8_t __s2 = __p2; \
40880  int16x4_t __ret; \
40881  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40882  __ret; \
40883})
40884#else
40885#define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40886  int16x4_t __s0 = __p0; \
40887  int16x4_t __s1 = __p1; \
40888  int16x8_t __s2 = __p2; \
40889  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40890  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40891  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40892  int16x4_t __ret; \
40893  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40894  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40895  __ret; \
40896})
40897#endif
40898
40899#ifdef __LITTLE_ENDIAN__
40900#define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40901  int32x4_t __s0 = __p0; \
40902  int32x4_t __s1 = __p1; \
40903  int32x4_t __s2 = __p2; \
40904  int32x4_t __ret; \
40905  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40906  __ret; \
40907})
40908#else
40909#define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40910  int32x4_t __s0 = __p0; \
40911  int32x4_t __s1 = __p1; \
40912  int32x4_t __s2 = __p2; \
40913  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40914  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40915  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40916  int32x4_t __ret; \
40917  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40918  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40919  __ret; \
40920})
40921#endif
40922
40923#ifdef __LITTLE_ENDIAN__
40924#define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40925  int16x8_t __s0 = __p0; \
40926  int16x8_t __s1 = __p1; \
40927  int16x8_t __s2 = __p2; \
40928  int16x8_t __ret; \
40929  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40930  __ret; \
40931})
40932#else
40933#define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40934  int16x8_t __s0 = __p0; \
40935  int16x8_t __s1 = __p1; \
40936  int16x8_t __s2 = __p2; \
40937  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40938  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40939  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40940  int16x8_t __ret; \
40941  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40942  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40943  __ret; \
40944})
40945#endif
40946
40947#ifdef __LITTLE_ENDIAN__
40948#define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40949  int32x2_t __s0 = __p0; \
40950  int32x2_t __s1 = __p1; \
40951  int32x4_t __s2 = __p2; \
40952  int32x2_t __ret; \
40953  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40954  __ret; \
40955})
40956#else
40957#define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40958  int32x2_t __s0 = __p0; \
40959  int32x2_t __s1 = __p1; \
40960  int32x4_t __s2 = __p2; \
40961  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40962  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40963  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40964  int32x2_t __ret; \
40965  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40966  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40967  __ret; \
40968})
40969#endif
40970
40971#ifdef __LITTLE_ENDIAN__
40972#define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40973  int16x4_t __s0 = __p0; \
40974  int16x4_t __s1 = __p1; \
40975  int16x8_t __s2 = __p2; \
40976  int16x4_t __ret; \
40977  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40978  __ret; \
40979})
40980#else
40981#define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40982  int16x4_t __s0 = __p0; \
40983  int16x4_t __s1 = __p1; \
40984  int16x8_t __s2 = __p2; \
40985  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40986  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40987  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40988  int16x4_t __ret; \
40989  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40990  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40991  __ret; \
40992})
40993#endif
40994
40995#endif
40996#if defined(__aarch64__)
40997#ifdef __LITTLE_ENDIAN__
40998__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
40999  float64x2_t __ret;
41000  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
41001  return __ret;
41002}
41003#else
41004__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
41005  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41006  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41007  float64x2_t __ret;
41008  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
41009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41010  return __ret;
41011}
41012#endif
41013
41014#ifdef __LITTLE_ENDIAN__
41015__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
41016  float64x1_t __ret;
41017  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
41018  return __ret;
41019}
41020#else
41021__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
41022  float64x1_t __ret;
41023  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
41024  return __ret;
41025}
41026#endif
41027
41028#ifdef __LITTLE_ENDIAN__
41029__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
41030  float64_t __ret;
41031  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
41032  return __ret;
41033}
41034#else
41035__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
41036  float64_t __ret;
41037  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
41038  return __ret;
41039}
41040#endif
41041
41042#ifdef __LITTLE_ENDIAN__
41043__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41044  float32_t __ret;
41045  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41046  return __ret;
41047}
41048#else
41049__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41050  float32_t __ret;
41051  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41052  return __ret;
41053}
41054#endif
41055
41056#ifdef __LITTLE_ENDIAN__
41057__ai float64x2_t vabsq_f64(float64x2_t __p0) {
41058  float64x2_t __ret;
41059  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
41060  return __ret;
41061}
41062#else
41063__ai float64x2_t vabsq_f64(float64x2_t __p0) {
41064  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41065  float64x2_t __ret;
41066  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
41067  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41068  return __ret;
41069}
41070#endif
41071
41072#ifdef __LITTLE_ENDIAN__
41073__ai int64x2_t vabsq_s64(int64x2_t __p0) {
41074  int64x2_t __ret;
41075  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
41076  return __ret;
41077}
41078#else
41079__ai int64x2_t vabsq_s64(int64x2_t __p0) {
41080  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41081  int64x2_t __ret;
41082  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
41083  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41084  return __ret;
41085}
41086#endif
41087
41088#ifdef __LITTLE_ENDIAN__
41089__ai float64x1_t vabs_f64(float64x1_t __p0) {
41090  float64x1_t __ret;
41091  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41092  return __ret;
41093}
41094#else
41095__ai float64x1_t vabs_f64(float64x1_t __p0) {
41096  float64x1_t __ret;
41097  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41098  return __ret;
41099}
41100#endif
41101
41102#ifdef __LITTLE_ENDIAN__
41103__ai int64x1_t vabs_s64(int64x1_t __p0) {
41104  int64x1_t __ret;
41105  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41106  return __ret;
41107}
41108#else
41109__ai int64x1_t vabs_s64(int64x1_t __p0) {
41110  int64x1_t __ret;
41111  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41112  return __ret;
41113}
41114#endif
41115
41116#ifdef __LITTLE_ENDIAN__
41117__ai int64_t vabsd_s64(int64_t __p0) {
41118  int64_t __ret;
41119  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41120  return __ret;
41121}
41122#else
41123__ai int64_t vabsd_s64(int64_t __p0) {
41124  int64_t __ret;
41125  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41126  return __ret;
41127}
41128#endif
41129
41130#ifdef __LITTLE_ENDIAN__
41131__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41132  float64x2_t __ret;
41133  __ret = __p0 + __p1;
41134  return __ret;
41135}
41136#else
41137__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41138  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41139  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41140  float64x2_t __ret;
41141  __ret = __rev0 + __rev1;
41142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41143  return __ret;
41144}
41145#endif
41146
41147#ifdef __LITTLE_ENDIAN__
41148__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41149  float64x1_t __ret;
41150  __ret = __p0 + __p1;
41151  return __ret;
41152}
41153#else
41154__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41155  float64x1_t __ret;
41156  __ret = __p0 + __p1;
41157  return __ret;
41158}
41159#endif
41160
41161#ifdef __LITTLE_ENDIAN__
41162__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41163  uint64_t __ret;
41164  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41165  return __ret;
41166}
41167#else
41168__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41169  uint64_t __ret;
41170  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41171  return __ret;
41172}
41173#endif
41174
41175#ifdef __LITTLE_ENDIAN__
41176__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41177  int64_t __ret;
41178  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41179  return __ret;
41180}
41181#else
41182__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41183  int64_t __ret;
41184  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41185  return __ret;
41186}
41187#endif
41188
41189#ifdef __LITTLE_ENDIAN__
41190__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41191  uint16x8_t __ret;
41192  __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
41193  return __ret;
41194}
41195#else
41196__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41197  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41198  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41199  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41200  uint16x8_t __ret;
41201  __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
41202  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41203  return __ret;
41204}
41205#endif
41206
41207#ifdef __LITTLE_ENDIAN__
41208__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41209  uint32x4_t __ret;
41210  __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
41211  return __ret;
41212}
41213#else
41214__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41215  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41216  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41217  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41218  uint32x4_t __ret;
41219  __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
41220  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41221  return __ret;
41222}
41223#endif
41224
41225#ifdef __LITTLE_ENDIAN__
41226__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41227  uint8x16_t __ret;
41228  __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
41229  return __ret;
41230}
41231#else
41232__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41233  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41234  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41235  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41236  uint8x16_t __ret;
41237  __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
41238  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41239  return __ret;
41240}
41241#endif
41242
41243#ifdef __LITTLE_ENDIAN__
41244__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41245  int16x8_t __ret;
41246  __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
41247  return __ret;
41248}
41249#else
41250__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41251  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41252  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41253  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41254  int16x8_t __ret;
41255  __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
41256  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41257  return __ret;
41258}
41259#endif
41260
41261#ifdef __LITTLE_ENDIAN__
41262__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41263  int32x4_t __ret;
41264  __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
41265  return __ret;
41266}
41267#else
41268__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41269  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41270  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41271  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41272  int32x4_t __ret;
41273  __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
41274  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41275  return __ret;
41276}
41277#endif
41278
41279#ifdef __LITTLE_ENDIAN__
41280__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41281  int8x16_t __ret;
41282  __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
41283  return __ret;
41284}
41285#else
41286__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41287  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41288  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41289  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41290  int8x16_t __ret;
41291  __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
41292  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41293  return __ret;
41294}
41295#endif
41296
41297#ifdef __LITTLE_ENDIAN__
41298__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41299  uint16_t __ret;
41300  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__p0);
41301  return __ret;
41302}
41303#else
41304__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41305  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41306  uint16_t __ret;
41307  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__rev0);
41308  return __ret;
41309}
41310#endif
41311
41312#ifdef __LITTLE_ENDIAN__
41313__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41314  uint64_t __ret;
41315  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__p0);
41316  return __ret;
41317}
41318#else
41319__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41320  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41321  uint64_t __ret;
41322  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__rev0);
41323  return __ret;
41324}
41325#endif
41326
41327#ifdef __LITTLE_ENDIAN__
41328__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41329  uint32_t __ret;
41330  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__p0);
41331  return __ret;
41332}
41333#else
41334__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41335  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41336  uint32_t __ret;
41337  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__rev0);
41338  return __ret;
41339}
41340#endif
41341
41342#ifdef __LITTLE_ENDIAN__
41343__ai int16_t vaddlvq_s8(int8x16_t __p0) {
41344  int16_t __ret;
41345  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__p0);
41346  return __ret;
41347}
41348#else
41349__ai int16_t vaddlvq_s8(int8x16_t __p0) {
41350  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41351  int16_t __ret;
41352  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__rev0);
41353  return __ret;
41354}
41355#endif
41356
41357#ifdef __LITTLE_ENDIAN__
41358__ai int64_t vaddlvq_s32(int32x4_t __p0) {
41359  int64_t __ret;
41360  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__p0);
41361  return __ret;
41362}
41363#else
41364__ai int64_t vaddlvq_s32(int32x4_t __p0) {
41365  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41366  int64_t __ret;
41367  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__rev0);
41368  return __ret;
41369}
41370#endif
41371
41372#ifdef __LITTLE_ENDIAN__
41373__ai int32_t vaddlvq_s16(int16x8_t __p0) {
41374  int32_t __ret;
41375  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__p0);
41376  return __ret;
41377}
41378#else
41379__ai int32_t vaddlvq_s16(int16x8_t __p0) {
41380  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41381  int32_t __ret;
41382  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__rev0);
41383  return __ret;
41384}
41385#endif
41386
41387#ifdef __LITTLE_ENDIAN__
41388__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41389  uint16_t __ret;
41390  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__p0);
41391  return __ret;
41392}
41393#else
41394__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41395  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41396  uint16_t __ret;
41397  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__rev0);
41398  return __ret;
41399}
41400#endif
41401
41402#ifdef __LITTLE_ENDIAN__
41403__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41404  uint64_t __ret;
41405  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__p0);
41406  return __ret;
41407}
41408#else
41409__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41410  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41411  uint64_t __ret;
41412  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__rev0);
41413  return __ret;
41414}
41415#endif
41416
41417#ifdef __LITTLE_ENDIAN__
41418__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41419  uint32_t __ret;
41420  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__p0);
41421  return __ret;
41422}
41423#else
41424__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41425  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41426  uint32_t __ret;
41427  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__rev0);
41428  return __ret;
41429}
41430#endif
41431
41432#ifdef __LITTLE_ENDIAN__
41433__ai int16_t vaddlv_s8(int8x8_t __p0) {
41434  int16_t __ret;
41435  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__p0);
41436  return __ret;
41437}
41438#else
41439__ai int16_t vaddlv_s8(int8x8_t __p0) {
41440  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41441  int16_t __ret;
41442  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__rev0);
41443  return __ret;
41444}
41445#endif
41446
41447#ifdef __LITTLE_ENDIAN__
41448__ai int64_t vaddlv_s32(int32x2_t __p0) {
41449  int64_t __ret;
41450  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__p0);
41451  return __ret;
41452}
41453#else
41454__ai int64_t vaddlv_s32(int32x2_t __p0) {
41455  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41456  int64_t __ret;
41457  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__rev0);
41458  return __ret;
41459}
41460#endif
41461
41462#ifdef __LITTLE_ENDIAN__
41463__ai int32_t vaddlv_s16(int16x4_t __p0) {
41464  int32_t __ret;
41465  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__p0);
41466  return __ret;
41467}
41468#else
41469__ai int32_t vaddlv_s16(int16x4_t __p0) {
41470  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41471  int32_t __ret;
41472  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__rev0);
41473  return __ret;
41474}
41475#endif
41476
41477#ifdef __LITTLE_ENDIAN__
41478__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41479  uint8_t __ret;
41480  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__p0);
41481  return __ret;
41482}
41483#else
41484__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41485  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41486  uint8_t __ret;
41487  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__rev0);
41488  return __ret;
41489}
41490#endif
41491
41492#ifdef __LITTLE_ENDIAN__
41493__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41494  uint32_t __ret;
41495  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__p0);
41496  return __ret;
41497}
41498#else
41499__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41500  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41501  uint32_t __ret;
41502  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__rev0);
41503  return __ret;
41504}
41505#endif
41506
41507#ifdef __LITTLE_ENDIAN__
41508__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41509  uint64_t __ret;
41510  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__p0);
41511  return __ret;
41512}
41513#else
41514__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41515  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41516  uint64_t __ret;
41517  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__rev0);
41518  return __ret;
41519}
41520#endif
41521
41522#ifdef __LITTLE_ENDIAN__
41523__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41524  uint16_t __ret;
41525  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__p0);
41526  return __ret;
41527}
41528#else
41529__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41530  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41531  uint16_t __ret;
41532  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__rev0);
41533  return __ret;
41534}
41535#endif
41536
41537#ifdef __LITTLE_ENDIAN__
41538__ai int8_t vaddvq_s8(int8x16_t __p0) {
41539  int8_t __ret;
41540  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__p0);
41541  return __ret;
41542}
41543#else
41544__ai int8_t vaddvq_s8(int8x16_t __p0) {
41545  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41546  int8_t __ret;
41547  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__rev0);
41548  return __ret;
41549}
41550#endif
41551
41552#ifdef __LITTLE_ENDIAN__
41553__ai float64_t vaddvq_f64(float64x2_t __p0) {
41554  float64_t __ret;
41555  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__p0);
41556  return __ret;
41557}
41558#else
41559__ai float64_t vaddvq_f64(float64x2_t __p0) {
41560  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41561  float64_t __ret;
41562  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__rev0);
41563  return __ret;
41564}
41565#endif
41566
41567#ifdef __LITTLE_ENDIAN__
41568__ai float32_t vaddvq_f32(float32x4_t __p0) {
41569  float32_t __ret;
41570  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__p0);
41571  return __ret;
41572}
41573#else
41574__ai float32_t vaddvq_f32(float32x4_t __p0) {
41575  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41576  float32_t __ret;
41577  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__rev0);
41578  return __ret;
41579}
41580#endif
41581
41582#ifdef __LITTLE_ENDIAN__
41583__ai int32_t vaddvq_s32(int32x4_t __p0) {
41584  int32_t __ret;
41585  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__p0);
41586  return __ret;
41587}
41588#else
41589__ai int32_t vaddvq_s32(int32x4_t __p0) {
41590  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41591  int32_t __ret;
41592  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__rev0);
41593  return __ret;
41594}
41595#endif
41596
41597#ifdef __LITTLE_ENDIAN__
41598__ai int64_t vaddvq_s64(int64x2_t __p0) {
41599  int64_t __ret;
41600  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__p0);
41601  return __ret;
41602}
41603#else
41604__ai int64_t vaddvq_s64(int64x2_t __p0) {
41605  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41606  int64_t __ret;
41607  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__rev0);
41608  return __ret;
41609}
41610#endif
41611
41612#ifdef __LITTLE_ENDIAN__
41613__ai int16_t vaddvq_s16(int16x8_t __p0) {
41614  int16_t __ret;
41615  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__p0);
41616  return __ret;
41617}
41618#else
41619__ai int16_t vaddvq_s16(int16x8_t __p0) {
41620  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41621  int16_t __ret;
41622  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__rev0);
41623  return __ret;
41624}
41625#endif
41626
41627#ifdef __LITTLE_ENDIAN__
41628__ai uint8_t vaddv_u8(uint8x8_t __p0) {
41629  uint8_t __ret;
41630  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__p0);
41631  return __ret;
41632}
41633#else
41634__ai uint8_t vaddv_u8(uint8x8_t __p0) {
41635  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41636  uint8_t __ret;
41637  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__rev0);
41638  return __ret;
41639}
41640#endif
41641
41642#ifdef __LITTLE_ENDIAN__
41643__ai uint32_t vaddv_u32(uint32x2_t __p0) {
41644  uint32_t __ret;
41645  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__p0);
41646  return __ret;
41647}
41648#else
41649__ai uint32_t vaddv_u32(uint32x2_t __p0) {
41650  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41651  uint32_t __ret;
41652  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__rev0);
41653  return __ret;
41654}
41655#endif
41656
41657#ifdef __LITTLE_ENDIAN__
41658__ai uint16_t vaddv_u16(uint16x4_t __p0) {
41659  uint16_t __ret;
41660  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__p0);
41661  return __ret;
41662}
41663#else
41664__ai uint16_t vaddv_u16(uint16x4_t __p0) {
41665  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41666  uint16_t __ret;
41667  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__rev0);
41668  return __ret;
41669}
41670#endif
41671
41672#ifdef __LITTLE_ENDIAN__
41673__ai int8_t vaddv_s8(int8x8_t __p0) {
41674  int8_t __ret;
41675  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__p0);
41676  return __ret;
41677}
41678#else
41679__ai int8_t vaddv_s8(int8x8_t __p0) {
41680  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41681  int8_t __ret;
41682  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__rev0);
41683  return __ret;
41684}
41685#endif
41686
41687#ifdef __LITTLE_ENDIAN__
41688__ai float32_t vaddv_f32(float32x2_t __p0) {
41689  float32_t __ret;
41690  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__p0);
41691  return __ret;
41692}
41693#else
41694__ai float32_t vaddv_f32(float32x2_t __p0) {
41695  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41696  float32_t __ret;
41697  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__rev0);
41698  return __ret;
41699}
41700#endif
41701
41702#ifdef __LITTLE_ENDIAN__
41703__ai int32_t vaddv_s32(int32x2_t __p0) {
41704  int32_t __ret;
41705  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__p0);
41706  return __ret;
41707}
41708#else
41709__ai int32_t vaddv_s32(int32x2_t __p0) {
41710  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41711  int32_t __ret;
41712  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__rev0);
41713  return __ret;
41714}
41715#endif
41716
41717#ifdef __LITTLE_ENDIAN__
41718__ai int16_t vaddv_s16(int16x4_t __p0) {
41719  int16_t __ret;
41720  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__p0);
41721  return __ret;
41722}
41723#else
41724__ai int16_t vaddv_s16(int16x4_t __p0) {
41725  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41726  int16_t __ret;
41727  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__rev0);
41728  return __ret;
41729}
41730#endif
41731
41732#ifdef __LITTLE_ENDIAN__
41733__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41734  poly64x1_t __ret;
41735  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41736  return __ret;
41737}
41738#else
41739__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41740  poly64x1_t __ret;
41741  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41742  return __ret;
41743}
41744#endif
41745
41746#ifdef __LITTLE_ENDIAN__
41747__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41748  poly64x2_t __ret;
41749  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
41750  return __ret;
41751}
41752#else
41753__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41754  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41755  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41756  poly64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41757  poly64x2_t __ret;
41758  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
41759  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41760  return __ret;
41761}
41762#endif
41763
41764#ifdef __LITTLE_ENDIAN__
41765__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41766  float64x2_t __ret;
41767  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
41768  return __ret;
41769}
41770#else
41771__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41772  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41773  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41774  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41775  float64x2_t __ret;
41776  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
41777  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41778  return __ret;
41779}
41780#endif
41781
41782#ifdef __LITTLE_ENDIAN__
41783__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41784  float64x1_t __ret;
41785  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41786  return __ret;
41787}
41788#else
41789__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41790  float64x1_t __ret;
41791  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41792  return __ret;
41793}
41794#endif
41795
41796#ifdef __LITTLE_ENDIAN__
41797__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41798  uint64x2_t __ret;
41799  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41800  return __ret;
41801}
41802#else
41803__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41804  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41805  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41806  uint64x2_t __ret;
41807  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41808  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41809  return __ret;
41810}
41811#endif
41812
41813#ifdef __LITTLE_ENDIAN__
41814__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41815  uint64x1_t __ret;
41816  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41817  return __ret;
41818}
41819#else
41820__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41821  uint64x1_t __ret;
41822  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41823  return __ret;
41824}
41825#endif
41826
41827#ifdef __LITTLE_ENDIAN__
41828__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41829  uint64_t __ret;
41830  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41831  return __ret;
41832}
41833#else
41834__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41835  uint64_t __ret;
41836  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41837  return __ret;
41838}
41839#endif
41840
41841#ifdef __LITTLE_ENDIAN__
41842__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41843  uint32_t __ret;
41844  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41845  return __ret;
41846}
41847#else
41848__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41849  uint32_t __ret;
41850  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41851  return __ret;
41852}
41853#endif
41854
41855#ifdef __LITTLE_ENDIAN__
41856__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41857  uint64x2_t __ret;
41858  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41859  return __ret;
41860}
41861#else
41862__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41863  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41864  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41865  uint64x2_t __ret;
41866  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41868  return __ret;
41869}
41870#endif
41871
41872#ifdef __LITTLE_ENDIAN__
41873__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41874  uint64x1_t __ret;
41875  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41876  return __ret;
41877}
41878#else
41879__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41880  uint64x1_t __ret;
41881  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41882  return __ret;
41883}
41884#endif
41885
41886#ifdef __LITTLE_ENDIAN__
41887__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41888  uint64_t __ret;
41889  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41890  return __ret;
41891}
41892#else
41893__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41894  uint64_t __ret;
41895  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41896  return __ret;
41897}
41898#endif
41899
41900#ifdef __LITTLE_ENDIAN__
41901__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41902  uint32_t __ret;
41903  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41904  return __ret;
41905}
41906#else
41907__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41908  uint32_t __ret;
41909  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41910  return __ret;
41911}
41912#endif
41913
41914#ifdef __LITTLE_ENDIAN__
41915__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41916  uint64x2_t __ret;
41917  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41918  return __ret;
41919}
41920#else
41921__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41922  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41923  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41924  uint64x2_t __ret;
41925  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41927  return __ret;
41928}
41929#endif
41930
41931#ifdef __LITTLE_ENDIAN__
41932__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41933  uint64x1_t __ret;
41934  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41935  return __ret;
41936}
41937#else
41938__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41939  uint64x1_t __ret;
41940  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41941  return __ret;
41942}
41943#endif
41944
41945#ifdef __LITTLE_ENDIAN__
41946__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41947  uint64_t __ret;
41948  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41949  return __ret;
41950}
41951#else
41952__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41953  uint64_t __ret;
41954  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41955  return __ret;
41956}
41957#endif
41958
41959#ifdef __LITTLE_ENDIAN__
41960__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41961  uint32_t __ret;
41962  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41963  return __ret;
41964}
41965#else
41966__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41967  uint32_t __ret;
41968  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41969  return __ret;
41970}
41971#endif
41972
41973#ifdef __LITTLE_ENDIAN__
41974__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41975  uint64x2_t __ret;
41976  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41977  return __ret;
41978}
41979#else
41980__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41981  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41982  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41983  uint64x2_t __ret;
41984  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41985  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41986  return __ret;
41987}
41988#endif
41989
41990#ifdef __LITTLE_ENDIAN__
41991__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41992  uint64x1_t __ret;
41993  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41994  return __ret;
41995}
41996#else
41997__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41998  uint64x1_t __ret;
41999  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
42000  return __ret;
42001}
42002#endif
42003
42004#ifdef __LITTLE_ENDIAN__
42005__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
42006  uint64_t __ret;
42007  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
42008  return __ret;
42009}
42010#else
42011__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
42012  uint64_t __ret;
42013  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
42014  return __ret;
42015}
42016#endif
42017
42018#ifdef __LITTLE_ENDIAN__
42019__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
42020  uint32_t __ret;
42021  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
42022  return __ret;
42023}
42024#else
42025__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
42026  uint32_t __ret;
42027  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
42028  return __ret;
42029}
42030#endif
42031
42032#ifdef __LITTLE_ENDIAN__
42033__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
42034  uint64x1_t __ret;
42035  __ret = (uint64x1_t)(__p0 == __p1);
42036  return __ret;
42037}
42038#else
42039__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
42040  uint64x1_t __ret;
42041  __ret = (uint64x1_t)(__p0 == __p1);
42042  return __ret;
42043}
42044#endif
42045
42046#ifdef __LITTLE_ENDIAN__
42047__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42048  uint64x2_t __ret;
42049  __ret = (uint64x2_t)(__p0 == __p1);
42050  return __ret;
42051}
42052#else
42053__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42054  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42055  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42056  uint64x2_t __ret;
42057  __ret = (uint64x2_t)(__rev0 == __rev1);
42058  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42059  return __ret;
42060}
42061#endif
42062
42063#ifdef __LITTLE_ENDIAN__
42064__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42065  uint64x2_t __ret;
42066  __ret = (uint64x2_t)(__p0 == __p1);
42067  return __ret;
42068}
42069#else
42070__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42071  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42072  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42073  uint64x2_t __ret;
42074  __ret = (uint64x2_t)(__rev0 == __rev1);
42075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42076  return __ret;
42077}
42078#endif
42079
42080#ifdef __LITTLE_ENDIAN__
42081__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42082  uint64x2_t __ret;
42083  __ret = (uint64x2_t)(__p0 == __p1);
42084  return __ret;
42085}
42086#else
42087__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42088  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42089  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42090  uint64x2_t __ret;
42091  __ret = (uint64x2_t)(__rev0 == __rev1);
42092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42093  return __ret;
42094}
42095#endif
42096
42097#ifdef __LITTLE_ENDIAN__
42098__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42099  uint64x2_t __ret;
42100  __ret = (uint64x2_t)(__p0 == __p1);
42101  return __ret;
42102}
42103#else
42104__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42105  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42106  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42107  uint64x2_t __ret;
42108  __ret = (uint64x2_t)(__rev0 == __rev1);
42109  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42110  return __ret;
42111}
42112#endif
42113
42114#ifdef __LITTLE_ENDIAN__
42115__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42116  uint64x1_t __ret;
42117  __ret = (uint64x1_t)(__p0 == __p1);
42118  return __ret;
42119}
42120#else
42121__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42122  uint64x1_t __ret;
42123  __ret = (uint64x1_t)(__p0 == __p1);
42124  return __ret;
42125}
42126#endif
42127
42128#ifdef __LITTLE_ENDIAN__
42129__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42130  uint64x1_t __ret;
42131  __ret = (uint64x1_t)(__p0 == __p1);
42132  return __ret;
42133}
42134#else
42135__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42136  uint64x1_t __ret;
42137  __ret = (uint64x1_t)(__p0 == __p1);
42138  return __ret;
42139}
42140#endif
42141
42142#ifdef __LITTLE_ENDIAN__
42143__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42144  uint64x1_t __ret;
42145  __ret = (uint64x1_t)(__p0 == __p1);
42146  return __ret;
42147}
42148#else
42149__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42150  uint64x1_t __ret;
42151  __ret = (uint64x1_t)(__p0 == __p1);
42152  return __ret;
42153}
42154#endif
42155
42156#ifdef __LITTLE_ENDIAN__
42157__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42158  uint64_t __ret;
42159  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42160  return __ret;
42161}
42162#else
42163__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42164  uint64_t __ret;
42165  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42166  return __ret;
42167}
42168#endif
42169
42170#ifdef __LITTLE_ENDIAN__
42171__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42172  int64_t __ret;
42173  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42174  return __ret;
42175}
42176#else
42177__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42178  int64_t __ret;
42179  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42180  return __ret;
42181}
42182#endif
42183
42184#ifdef __LITTLE_ENDIAN__
42185__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42186  uint64_t __ret;
42187  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42188  return __ret;
42189}
42190#else
42191__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42192  uint64_t __ret;
42193  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42194  return __ret;
42195}
42196#endif
42197
42198#ifdef __LITTLE_ENDIAN__
42199__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42200  uint32_t __ret;
42201  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42202  return __ret;
42203}
42204#else
42205__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42206  uint32_t __ret;
42207  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42208  return __ret;
42209}
42210#endif
42211
42212#ifdef __LITTLE_ENDIAN__
42213__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42214  uint8x8_t __ret;
42215  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42216  return __ret;
42217}
42218#else
42219__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42220  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42221  uint8x8_t __ret;
42222  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42223  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42224  return __ret;
42225}
42226#endif
42227
42228#ifdef __LITTLE_ENDIAN__
42229__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42230  uint64x1_t __ret;
42231  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42232  return __ret;
42233}
42234#else
42235__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42236  uint64x1_t __ret;
42237  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42238  return __ret;
42239}
42240#endif
42241
42242#ifdef __LITTLE_ENDIAN__
42243__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42244  uint16x4_t __ret;
42245  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42246  return __ret;
42247}
42248#else
42249__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42250  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42251  uint16x4_t __ret;
42252  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42253  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42254  return __ret;
42255}
42256#endif
42257
42258#ifdef __LITTLE_ENDIAN__
42259__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42260  uint8x16_t __ret;
42261  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42262  return __ret;
42263}
42264#else
42265__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42266  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42267  uint8x16_t __ret;
42268  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42269  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42270  return __ret;
42271}
42272#endif
42273
42274#ifdef __LITTLE_ENDIAN__
42275__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42276  uint64x2_t __ret;
42277  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42278  return __ret;
42279}
42280#else
42281__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42282  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42283  uint64x2_t __ret;
42284  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42285  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42286  return __ret;
42287}
42288#endif
42289
42290#ifdef __LITTLE_ENDIAN__
42291__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42292  uint16x8_t __ret;
42293  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42294  return __ret;
42295}
42296#else
42297__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42298  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42299  uint16x8_t __ret;
42300  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42302  return __ret;
42303}
42304#endif
42305
42306#ifdef __LITTLE_ENDIAN__
42307__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42308  uint8x16_t __ret;
42309  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42310  return __ret;
42311}
42312#else
42313__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42314  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42315  uint8x16_t __ret;
42316  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42317  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42318  return __ret;
42319}
42320#endif
42321
42322#ifdef __LITTLE_ENDIAN__
42323__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42324  uint32x4_t __ret;
42325  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42326  return __ret;
42327}
42328#else
42329__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42330  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42331  uint32x4_t __ret;
42332  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42333  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42334  return __ret;
42335}
42336#endif
42337
42338#ifdef __LITTLE_ENDIAN__
42339__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42340  uint64x2_t __ret;
42341  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42342  return __ret;
42343}
42344#else
42345__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42346  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42347  uint64x2_t __ret;
42348  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42349  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42350  return __ret;
42351}
42352#endif
42353
42354#ifdef __LITTLE_ENDIAN__
42355__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42356  uint16x8_t __ret;
42357  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42358  return __ret;
42359}
42360#else
42361__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42362  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42363  uint16x8_t __ret;
42364  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42365  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42366  return __ret;
42367}
42368#endif
42369
42370#ifdef __LITTLE_ENDIAN__
42371__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42372  uint8x16_t __ret;
42373  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42374  return __ret;
42375}
42376#else
42377__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42378  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42379  uint8x16_t __ret;
42380  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42381  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42382  return __ret;
42383}
42384#endif
42385
42386#ifdef __LITTLE_ENDIAN__
42387__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42388  uint64x2_t __ret;
42389  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42390  return __ret;
42391}
42392#else
42393__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42394  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42395  uint64x2_t __ret;
42396  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42397  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42398  return __ret;
42399}
42400#endif
42401
42402#ifdef __LITTLE_ENDIAN__
42403__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42404  uint32x4_t __ret;
42405  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42406  return __ret;
42407}
42408#else
42409__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42410  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42411  uint32x4_t __ret;
42412  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42413  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42414  return __ret;
42415}
42416#endif
42417
42418#ifdef __LITTLE_ENDIAN__
42419__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42420  uint32x4_t __ret;
42421  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42422  return __ret;
42423}
42424#else
42425__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42426  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42427  uint32x4_t __ret;
42428  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42429  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42430  return __ret;
42431}
42432#endif
42433
42434#ifdef __LITTLE_ENDIAN__
42435__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42436  uint64x2_t __ret;
42437  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42438  return __ret;
42439}
42440#else
42441__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42442  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42443  uint64x2_t __ret;
42444  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42445  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42446  return __ret;
42447}
42448#endif
42449
42450#ifdef __LITTLE_ENDIAN__
42451__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42452  uint16x8_t __ret;
42453  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42454  return __ret;
42455}
42456#else
42457__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42458  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42459  uint16x8_t __ret;
42460  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42461  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42462  return __ret;
42463}
42464#endif
42465
42466#ifdef __LITTLE_ENDIAN__
42467__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42468  uint8x8_t __ret;
42469  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42470  return __ret;
42471}
42472#else
42473__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42474  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42475  uint8x8_t __ret;
42476  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42477  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42478  return __ret;
42479}
42480#endif
42481
42482#ifdef __LITTLE_ENDIAN__
42483__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42484  uint32x2_t __ret;
42485  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42486  return __ret;
42487}
42488#else
42489__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42490  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42491  uint32x2_t __ret;
42492  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42493  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42494  return __ret;
42495}
42496#endif
42497
42498#ifdef __LITTLE_ENDIAN__
42499__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42500  uint64x1_t __ret;
42501  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42502  return __ret;
42503}
42504#else
42505__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42506  uint64x1_t __ret;
42507  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42508  return __ret;
42509}
42510#endif
42511
42512#ifdef __LITTLE_ENDIAN__
42513__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42514  uint16x4_t __ret;
42515  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42516  return __ret;
42517}
42518#else
42519__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42520  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42521  uint16x4_t __ret;
42522  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42523  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42524  return __ret;
42525}
42526#endif
42527
42528#ifdef __LITTLE_ENDIAN__
42529__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42530  uint8x8_t __ret;
42531  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42532  return __ret;
42533}
42534#else
42535__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42536  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42537  uint8x8_t __ret;
42538  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42539  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42540  return __ret;
42541}
42542#endif
42543
42544#ifdef __LITTLE_ENDIAN__
42545__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42546  uint64x1_t __ret;
42547  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42548  return __ret;
42549}
42550#else
42551__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42552  uint64x1_t __ret;
42553  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42554  return __ret;
42555}
42556#endif
42557
42558#ifdef __LITTLE_ENDIAN__
42559__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42560  uint32x2_t __ret;
42561  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42562  return __ret;
42563}
42564#else
42565__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42566  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42567  uint32x2_t __ret;
42568  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42569  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42570  return __ret;
42571}
42572#endif
42573
42574#ifdef __LITTLE_ENDIAN__
42575__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42576  uint32x2_t __ret;
42577  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42578  return __ret;
42579}
42580#else
42581__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42582  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42583  uint32x2_t __ret;
42584  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42585  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42586  return __ret;
42587}
42588#endif
42589
42590#ifdef __LITTLE_ENDIAN__
42591__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42592  uint64x1_t __ret;
42593  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42594  return __ret;
42595}
42596#else
42597__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42598  uint64x1_t __ret;
42599  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42600  return __ret;
42601}
42602#endif
42603
42604#ifdef __LITTLE_ENDIAN__
42605__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42606  uint16x4_t __ret;
42607  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42608  return __ret;
42609}
42610#else
42611__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42612  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42613  uint16x4_t __ret;
42614  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42615  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42616  return __ret;
42617}
42618#endif
42619
42620#ifdef __LITTLE_ENDIAN__
42621__ai uint64_t vceqzd_u64(uint64_t __p0) {
42622  uint64_t __ret;
42623  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42624  return __ret;
42625}
42626#else
42627__ai uint64_t vceqzd_u64(uint64_t __p0) {
42628  uint64_t __ret;
42629  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42630  return __ret;
42631}
42632#endif
42633
42634#ifdef __LITTLE_ENDIAN__
42635__ai int64_t vceqzd_s64(int64_t __p0) {
42636  int64_t __ret;
42637  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42638  return __ret;
42639}
42640#else
42641__ai int64_t vceqzd_s64(int64_t __p0) {
42642  int64_t __ret;
42643  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42644  return __ret;
42645}
42646#endif
42647
42648#ifdef __LITTLE_ENDIAN__
42649__ai uint64_t vceqzd_f64(float64_t __p0) {
42650  uint64_t __ret;
42651  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42652  return __ret;
42653}
42654#else
42655__ai uint64_t vceqzd_f64(float64_t __p0) {
42656  uint64_t __ret;
42657  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42658  return __ret;
42659}
42660#endif
42661
42662#ifdef __LITTLE_ENDIAN__
42663__ai uint32_t vceqzs_f32(float32_t __p0) {
42664  uint32_t __ret;
42665  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42666  return __ret;
42667}
42668#else
42669__ai uint32_t vceqzs_f32(float32_t __p0) {
42670  uint32_t __ret;
42671  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42672  return __ret;
42673}
42674#endif
42675
42676#ifdef __LITTLE_ENDIAN__
42677__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42678  uint64x2_t __ret;
42679  __ret = (uint64x2_t)(__p0 >= __p1);
42680  return __ret;
42681}
42682#else
42683__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42684  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42685  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42686  uint64x2_t __ret;
42687  __ret = (uint64x2_t)(__rev0 >= __rev1);
42688  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42689  return __ret;
42690}
42691#endif
42692
42693#ifdef __LITTLE_ENDIAN__
42694__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42695  uint64x2_t __ret;
42696  __ret = (uint64x2_t)(__p0 >= __p1);
42697  return __ret;
42698}
42699#else
42700__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42701  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42702  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42703  uint64x2_t __ret;
42704  __ret = (uint64x2_t)(__rev0 >= __rev1);
42705  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42706  return __ret;
42707}
42708#endif
42709
42710#ifdef __LITTLE_ENDIAN__
42711__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42712  uint64x2_t __ret;
42713  __ret = (uint64x2_t)(__p0 >= __p1);
42714  return __ret;
42715}
42716#else
42717__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42718  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42719  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42720  uint64x2_t __ret;
42721  __ret = (uint64x2_t)(__rev0 >= __rev1);
42722  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42723  return __ret;
42724}
42725#endif
42726
42727#ifdef __LITTLE_ENDIAN__
42728__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42729  uint64x1_t __ret;
42730  __ret = (uint64x1_t)(__p0 >= __p1);
42731  return __ret;
42732}
42733#else
42734__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42735  uint64x1_t __ret;
42736  __ret = (uint64x1_t)(__p0 >= __p1);
42737  return __ret;
42738}
42739#endif
42740
42741#ifdef __LITTLE_ENDIAN__
42742__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42743  uint64x1_t __ret;
42744  __ret = (uint64x1_t)(__p0 >= __p1);
42745  return __ret;
42746}
42747#else
42748__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42749  uint64x1_t __ret;
42750  __ret = (uint64x1_t)(__p0 >= __p1);
42751  return __ret;
42752}
42753#endif
42754
42755#ifdef __LITTLE_ENDIAN__
42756__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42757  uint64x1_t __ret;
42758  __ret = (uint64x1_t)(__p0 >= __p1);
42759  return __ret;
42760}
42761#else
42762__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42763  uint64x1_t __ret;
42764  __ret = (uint64x1_t)(__p0 >= __p1);
42765  return __ret;
42766}
42767#endif
42768
42769#ifdef __LITTLE_ENDIAN__
42770__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42771  int64_t __ret;
42772  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42773  return __ret;
42774}
42775#else
42776__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42777  int64_t __ret;
42778  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42779  return __ret;
42780}
42781#endif
42782
42783#ifdef __LITTLE_ENDIAN__
42784__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42785  uint64_t __ret;
42786  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42787  return __ret;
42788}
42789#else
42790__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42791  uint64_t __ret;
42792  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42793  return __ret;
42794}
42795#endif
42796
42797#ifdef __LITTLE_ENDIAN__
42798__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42799  uint64_t __ret;
42800  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42801  return __ret;
42802}
42803#else
42804__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42805  uint64_t __ret;
42806  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42807  return __ret;
42808}
42809#endif
42810
42811#ifdef __LITTLE_ENDIAN__
42812__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42813  uint32_t __ret;
42814  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42815  return __ret;
42816}
42817#else
42818__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42819  uint32_t __ret;
42820  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42821  return __ret;
42822}
42823#endif
42824
42825#ifdef __LITTLE_ENDIAN__
42826__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42827  uint8x16_t __ret;
42828  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
42829  return __ret;
42830}
42831#else
42832__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42833  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42834  uint8x16_t __ret;
42835  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
42836  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42837  return __ret;
42838}
42839#endif
42840
42841#ifdef __LITTLE_ENDIAN__
42842__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42843  uint64x2_t __ret;
42844  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42845  return __ret;
42846}
42847#else
42848__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42849  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42850  uint64x2_t __ret;
42851  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42852  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42853  return __ret;
42854}
42855#endif
42856
42857#ifdef __LITTLE_ENDIAN__
42858__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42859  uint32x4_t __ret;
42860  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42861  return __ret;
42862}
42863#else
42864__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42865  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42866  uint32x4_t __ret;
42867  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42869  return __ret;
42870}
42871#endif
42872
42873#ifdef __LITTLE_ENDIAN__
42874__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42875  uint32x4_t __ret;
42876  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42877  return __ret;
42878}
42879#else
42880__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42881  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42882  uint32x4_t __ret;
42883  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42884  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42885  return __ret;
42886}
42887#endif
42888
42889#ifdef __LITTLE_ENDIAN__
42890__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42891  uint64x2_t __ret;
42892  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42893  return __ret;
42894}
42895#else
42896__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42897  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42898  uint64x2_t __ret;
42899  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42900  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42901  return __ret;
42902}
42903#endif
42904
42905#ifdef __LITTLE_ENDIAN__
42906__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42907  uint16x8_t __ret;
42908  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
42909  return __ret;
42910}
42911#else
42912__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42913  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42914  uint16x8_t __ret;
42915  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
42916  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42917  return __ret;
42918}
42919#endif
42920
42921#ifdef __LITTLE_ENDIAN__
42922__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42923  uint8x8_t __ret;
42924  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
42925  return __ret;
42926}
42927#else
42928__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42929  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42930  uint8x8_t __ret;
42931  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
42932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42933  return __ret;
42934}
42935#endif
42936
42937#ifdef __LITTLE_ENDIAN__
42938__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42939  uint64x1_t __ret;
42940  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42941  return __ret;
42942}
42943#else
42944__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42945  uint64x1_t __ret;
42946  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42947  return __ret;
42948}
42949#endif
42950
42951#ifdef __LITTLE_ENDIAN__
42952__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42953  uint32x2_t __ret;
42954  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42955  return __ret;
42956}
42957#else
42958__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42959  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42960  uint32x2_t __ret;
42961  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42962  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42963  return __ret;
42964}
42965#endif
42966
42967#ifdef __LITTLE_ENDIAN__
42968__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42969  uint32x2_t __ret;
42970  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42971  return __ret;
42972}
42973#else
42974__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42975  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42976  uint32x2_t __ret;
42977  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42978  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42979  return __ret;
42980}
42981#endif
42982
42983#ifdef __LITTLE_ENDIAN__
42984__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42985  uint64x1_t __ret;
42986  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42987  return __ret;
42988}
42989#else
42990__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42991  uint64x1_t __ret;
42992  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42993  return __ret;
42994}
42995#endif
42996
42997#ifdef __LITTLE_ENDIAN__
42998__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42999  uint16x4_t __ret;
43000  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
43001  return __ret;
43002}
43003#else
43004__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
43005  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43006  uint16x4_t __ret;
43007  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
43008  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43009  return __ret;
43010}
43011#endif
43012
43013#ifdef __LITTLE_ENDIAN__
43014__ai int64_t vcgezd_s64(int64_t __p0) {
43015  int64_t __ret;
43016  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
43017  return __ret;
43018}
43019#else
43020__ai int64_t vcgezd_s64(int64_t __p0) {
43021  int64_t __ret;
43022  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
43023  return __ret;
43024}
43025#endif
43026
43027#ifdef __LITTLE_ENDIAN__
43028__ai uint64_t vcgezd_f64(float64_t __p0) {
43029  uint64_t __ret;
43030  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
43031  return __ret;
43032}
43033#else
43034__ai uint64_t vcgezd_f64(float64_t __p0) {
43035  uint64_t __ret;
43036  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
43037  return __ret;
43038}
43039#endif
43040
43041#ifdef __LITTLE_ENDIAN__
43042__ai uint32_t vcgezs_f32(float32_t __p0) {
43043  uint32_t __ret;
43044  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43045  return __ret;
43046}
43047#else
43048__ai uint32_t vcgezs_f32(float32_t __p0) {
43049  uint32_t __ret;
43050  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43051  return __ret;
43052}
43053#endif
43054
43055#ifdef __LITTLE_ENDIAN__
43056__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43057  uint64x2_t __ret;
43058  __ret = (uint64x2_t)(__p0 > __p1);
43059  return __ret;
43060}
43061#else
43062__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43063  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43064  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43065  uint64x2_t __ret;
43066  __ret = (uint64x2_t)(__rev0 > __rev1);
43067  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43068  return __ret;
43069}
43070#endif
43071
43072#ifdef __LITTLE_ENDIAN__
43073__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43074  uint64x2_t __ret;
43075  __ret = (uint64x2_t)(__p0 > __p1);
43076  return __ret;
43077}
43078#else
43079__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43080  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43081  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43082  uint64x2_t __ret;
43083  __ret = (uint64x2_t)(__rev0 > __rev1);
43084  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43085  return __ret;
43086}
43087#endif
43088
43089#ifdef __LITTLE_ENDIAN__
43090__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43091  uint64x2_t __ret;
43092  __ret = (uint64x2_t)(__p0 > __p1);
43093  return __ret;
43094}
43095#else
43096__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43097  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43098  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43099  uint64x2_t __ret;
43100  __ret = (uint64x2_t)(__rev0 > __rev1);
43101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43102  return __ret;
43103}
43104#endif
43105
43106#ifdef __LITTLE_ENDIAN__
43107__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43108  uint64x1_t __ret;
43109  __ret = (uint64x1_t)(__p0 > __p1);
43110  return __ret;
43111}
43112#else
43113__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43114  uint64x1_t __ret;
43115  __ret = (uint64x1_t)(__p0 > __p1);
43116  return __ret;
43117}
43118#endif
43119
43120#ifdef __LITTLE_ENDIAN__
43121__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43122  uint64x1_t __ret;
43123  __ret = (uint64x1_t)(__p0 > __p1);
43124  return __ret;
43125}
43126#else
43127__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43128  uint64x1_t __ret;
43129  __ret = (uint64x1_t)(__p0 > __p1);
43130  return __ret;
43131}
43132#endif
43133
43134#ifdef __LITTLE_ENDIAN__
43135__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43136  uint64x1_t __ret;
43137  __ret = (uint64x1_t)(__p0 > __p1);
43138  return __ret;
43139}
43140#else
43141__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43142  uint64x1_t __ret;
43143  __ret = (uint64x1_t)(__p0 > __p1);
43144  return __ret;
43145}
43146#endif
43147
43148#ifdef __LITTLE_ENDIAN__
43149__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43150  int64_t __ret;
43151  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43152  return __ret;
43153}
43154#else
43155__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43156  int64_t __ret;
43157  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43158  return __ret;
43159}
43160#endif
43161
43162#ifdef __LITTLE_ENDIAN__
43163__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43164  uint64_t __ret;
43165  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43166  return __ret;
43167}
43168#else
43169__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43170  uint64_t __ret;
43171  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43172  return __ret;
43173}
43174#endif
43175
43176#ifdef __LITTLE_ENDIAN__
43177__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43178  uint64_t __ret;
43179  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43180  return __ret;
43181}
43182#else
43183__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43184  uint64_t __ret;
43185  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43186  return __ret;
43187}
43188#endif
43189
43190#ifdef __LITTLE_ENDIAN__
43191__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43192  uint32_t __ret;
43193  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43194  return __ret;
43195}
43196#else
43197__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43198  uint32_t __ret;
43199  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43200  return __ret;
43201}
43202#endif
43203
43204#ifdef __LITTLE_ENDIAN__
43205__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43206  uint8x16_t __ret;
43207  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
43208  return __ret;
43209}
43210#else
43211__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43212  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43213  uint8x16_t __ret;
43214  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
43215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43216  return __ret;
43217}
43218#endif
43219
43220#ifdef __LITTLE_ENDIAN__
43221__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43222  uint64x2_t __ret;
43223  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43224  return __ret;
43225}
43226#else
43227__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43228  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43229  uint64x2_t __ret;
43230  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43232  return __ret;
43233}
43234#endif
43235
43236#ifdef __LITTLE_ENDIAN__
43237__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43238  uint32x4_t __ret;
43239  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43240  return __ret;
43241}
43242#else
43243__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43244  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43245  uint32x4_t __ret;
43246  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43247  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43248  return __ret;
43249}
43250#endif
43251
43252#ifdef __LITTLE_ENDIAN__
43253__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43254  uint32x4_t __ret;
43255  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43256  return __ret;
43257}
43258#else
43259__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43260  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43261  uint32x4_t __ret;
43262  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43263  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43264  return __ret;
43265}
43266#endif
43267
43268#ifdef __LITTLE_ENDIAN__
43269__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43270  uint64x2_t __ret;
43271  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43272  return __ret;
43273}
43274#else
43275__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43276  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43277  uint64x2_t __ret;
43278  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43279  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43280  return __ret;
43281}
43282#endif
43283
43284#ifdef __LITTLE_ENDIAN__
43285__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43286  uint16x8_t __ret;
43287  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
43288  return __ret;
43289}
43290#else
43291__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43292  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43293  uint16x8_t __ret;
43294  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
43295  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43296  return __ret;
43297}
43298#endif
43299
43300#ifdef __LITTLE_ENDIAN__
43301__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43302  uint8x8_t __ret;
43303  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
43304  return __ret;
43305}
43306#else
43307__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43308  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43309  uint8x8_t __ret;
43310  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
43311  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43312  return __ret;
43313}
43314#endif
43315
43316#ifdef __LITTLE_ENDIAN__
43317__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43318  uint64x1_t __ret;
43319  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43320  return __ret;
43321}
43322#else
43323__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43324  uint64x1_t __ret;
43325  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43326  return __ret;
43327}
43328#endif
43329
43330#ifdef __LITTLE_ENDIAN__
43331__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43332  uint32x2_t __ret;
43333  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43334  return __ret;
43335}
43336#else
43337__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43338  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43339  uint32x2_t __ret;
43340  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43342  return __ret;
43343}
43344#endif
43345
43346#ifdef __LITTLE_ENDIAN__
43347__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43348  uint32x2_t __ret;
43349  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43350  return __ret;
43351}
43352#else
43353__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43354  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43355  uint32x2_t __ret;
43356  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43357  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43358  return __ret;
43359}
43360#endif
43361
43362#ifdef __LITTLE_ENDIAN__
43363__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43364  uint64x1_t __ret;
43365  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43366  return __ret;
43367}
43368#else
43369__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43370  uint64x1_t __ret;
43371  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43372  return __ret;
43373}
43374#endif
43375
43376#ifdef __LITTLE_ENDIAN__
43377__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43378  uint16x4_t __ret;
43379  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
43380  return __ret;
43381}
43382#else
43383__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43384  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43385  uint16x4_t __ret;
43386  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
43387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43388  return __ret;
43389}
43390#endif
43391
43392#ifdef __LITTLE_ENDIAN__
43393__ai int64_t vcgtzd_s64(int64_t __p0) {
43394  int64_t __ret;
43395  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43396  return __ret;
43397}
43398#else
43399__ai int64_t vcgtzd_s64(int64_t __p0) {
43400  int64_t __ret;
43401  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43402  return __ret;
43403}
43404#endif
43405
43406#ifdef __LITTLE_ENDIAN__
43407__ai uint64_t vcgtzd_f64(float64_t __p0) {
43408  uint64_t __ret;
43409  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43410  return __ret;
43411}
43412#else
43413__ai uint64_t vcgtzd_f64(float64_t __p0) {
43414  uint64_t __ret;
43415  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43416  return __ret;
43417}
43418#endif
43419
43420#ifdef __LITTLE_ENDIAN__
43421__ai uint32_t vcgtzs_f32(float32_t __p0) {
43422  uint32_t __ret;
43423  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43424  return __ret;
43425}
43426#else
43427__ai uint32_t vcgtzs_f32(float32_t __p0) {
43428  uint32_t __ret;
43429  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43430  return __ret;
43431}
43432#endif
43433
43434#ifdef __LITTLE_ENDIAN__
43435__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43436  uint64x2_t __ret;
43437  __ret = (uint64x2_t)(__p0 <= __p1);
43438  return __ret;
43439}
43440#else
43441__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43442  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43443  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43444  uint64x2_t __ret;
43445  __ret = (uint64x2_t)(__rev0 <= __rev1);
43446  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43447  return __ret;
43448}
43449#endif
43450
43451#ifdef __LITTLE_ENDIAN__
43452__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43453  uint64x2_t __ret;
43454  __ret = (uint64x2_t)(__p0 <= __p1);
43455  return __ret;
43456}
43457#else
43458__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43459  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43460  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43461  uint64x2_t __ret;
43462  __ret = (uint64x2_t)(__rev0 <= __rev1);
43463  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43464  return __ret;
43465}
43466#endif
43467
43468#ifdef __LITTLE_ENDIAN__
43469__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43470  uint64x2_t __ret;
43471  __ret = (uint64x2_t)(__p0 <= __p1);
43472  return __ret;
43473}
43474#else
43475__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43476  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43477  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43478  uint64x2_t __ret;
43479  __ret = (uint64x2_t)(__rev0 <= __rev1);
43480  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43481  return __ret;
43482}
43483#endif
43484
43485#ifdef __LITTLE_ENDIAN__
43486__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43487  uint64x1_t __ret;
43488  __ret = (uint64x1_t)(__p0 <= __p1);
43489  return __ret;
43490}
43491#else
43492__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43493  uint64x1_t __ret;
43494  __ret = (uint64x1_t)(__p0 <= __p1);
43495  return __ret;
43496}
43497#endif
43498
43499#ifdef __LITTLE_ENDIAN__
43500__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43501  uint64x1_t __ret;
43502  __ret = (uint64x1_t)(__p0 <= __p1);
43503  return __ret;
43504}
43505#else
43506__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43507  uint64x1_t __ret;
43508  __ret = (uint64x1_t)(__p0 <= __p1);
43509  return __ret;
43510}
43511#endif
43512
43513#ifdef __LITTLE_ENDIAN__
43514__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43515  uint64x1_t __ret;
43516  __ret = (uint64x1_t)(__p0 <= __p1);
43517  return __ret;
43518}
43519#else
43520__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43521  uint64x1_t __ret;
43522  __ret = (uint64x1_t)(__p0 <= __p1);
43523  return __ret;
43524}
43525#endif
43526
43527#ifdef __LITTLE_ENDIAN__
43528__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43529  uint64_t __ret;
43530  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43531  return __ret;
43532}
43533#else
43534__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43535  uint64_t __ret;
43536  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43537  return __ret;
43538}
43539#endif
43540
43541#ifdef __LITTLE_ENDIAN__
43542__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43543  int64_t __ret;
43544  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43545  return __ret;
43546}
43547#else
43548__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43549  int64_t __ret;
43550  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43551  return __ret;
43552}
43553#endif
43554
43555#ifdef __LITTLE_ENDIAN__
43556__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43557  uint64_t __ret;
43558  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43559  return __ret;
43560}
43561#else
43562__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43563  uint64_t __ret;
43564  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43565  return __ret;
43566}
43567#endif
43568
43569#ifdef __LITTLE_ENDIAN__
43570__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43571  uint32_t __ret;
43572  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43573  return __ret;
43574}
43575#else
43576__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43577  uint32_t __ret;
43578  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43579  return __ret;
43580}
43581#endif
43582
43583#ifdef __LITTLE_ENDIAN__
43584__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43585  uint8x16_t __ret;
43586  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
43587  return __ret;
43588}
43589#else
43590__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43591  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43592  uint8x16_t __ret;
43593  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
43594  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43595  return __ret;
43596}
43597#endif
43598
43599#ifdef __LITTLE_ENDIAN__
43600__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43601  uint64x2_t __ret;
43602  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43603  return __ret;
43604}
43605#else
43606__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43607  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43608  uint64x2_t __ret;
43609  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43610  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43611  return __ret;
43612}
43613#endif
43614
43615#ifdef __LITTLE_ENDIAN__
43616__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43617  uint32x4_t __ret;
43618  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43619  return __ret;
43620}
43621#else
43622__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43623  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43624  uint32x4_t __ret;
43625  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43627  return __ret;
43628}
43629#endif
43630
43631#ifdef __LITTLE_ENDIAN__
43632__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43633  uint32x4_t __ret;
43634  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43635  return __ret;
43636}
43637#else
43638__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43639  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43640  uint32x4_t __ret;
43641  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43642  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43643  return __ret;
43644}
43645#endif
43646
43647#ifdef __LITTLE_ENDIAN__
43648__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43649  uint64x2_t __ret;
43650  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43651  return __ret;
43652}
43653#else
43654__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43655  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43656  uint64x2_t __ret;
43657  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43659  return __ret;
43660}
43661#endif
43662
43663#ifdef __LITTLE_ENDIAN__
43664__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43665  uint16x8_t __ret;
43666  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
43667  return __ret;
43668}
43669#else
43670__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43671  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43672  uint16x8_t __ret;
43673  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
43674  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43675  return __ret;
43676}
43677#endif
43678
43679#ifdef __LITTLE_ENDIAN__
43680__ai uint8x8_t vclez_s8(int8x8_t __p0) {
43681  uint8x8_t __ret;
43682  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
43683  return __ret;
43684}
43685#else
43686__ai uint8x8_t vclez_s8(int8x8_t __p0) {
43687  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43688  uint8x8_t __ret;
43689  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
43690  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43691  return __ret;
43692}
43693#endif
43694
43695#ifdef __LITTLE_ENDIAN__
43696__ai uint64x1_t vclez_f64(float64x1_t __p0) {
43697  uint64x1_t __ret;
43698  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43699  return __ret;
43700}
43701#else
43702__ai uint64x1_t vclez_f64(float64x1_t __p0) {
43703  uint64x1_t __ret;
43704  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43705  return __ret;
43706}
43707#endif
43708
43709#ifdef __LITTLE_ENDIAN__
43710__ai uint32x2_t vclez_f32(float32x2_t __p0) {
43711  uint32x2_t __ret;
43712  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43713  return __ret;
43714}
43715#else
43716__ai uint32x2_t vclez_f32(float32x2_t __p0) {
43717  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43718  uint32x2_t __ret;
43719  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43720  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43721  return __ret;
43722}
43723#endif
43724
43725#ifdef __LITTLE_ENDIAN__
43726__ai uint32x2_t vclez_s32(int32x2_t __p0) {
43727  uint32x2_t __ret;
43728  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43729  return __ret;
43730}
43731#else
43732__ai uint32x2_t vclez_s32(int32x2_t __p0) {
43733  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43734  uint32x2_t __ret;
43735  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43736  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43737  return __ret;
43738}
43739#endif
43740
43741#ifdef __LITTLE_ENDIAN__
43742__ai uint64x1_t vclez_s64(int64x1_t __p0) {
43743  uint64x1_t __ret;
43744  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43745  return __ret;
43746}
43747#else
43748__ai uint64x1_t vclez_s64(int64x1_t __p0) {
43749  uint64x1_t __ret;
43750  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43751  return __ret;
43752}
43753#endif
43754
43755#ifdef __LITTLE_ENDIAN__
43756__ai uint16x4_t vclez_s16(int16x4_t __p0) {
43757  uint16x4_t __ret;
43758  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
43759  return __ret;
43760}
43761#else
43762__ai uint16x4_t vclez_s16(int16x4_t __p0) {
43763  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43764  uint16x4_t __ret;
43765  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
43766  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43767  return __ret;
43768}
43769#endif
43770
43771#ifdef __LITTLE_ENDIAN__
43772__ai int64_t vclezd_s64(int64_t __p0) {
43773  int64_t __ret;
43774  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43775  return __ret;
43776}
43777#else
43778__ai int64_t vclezd_s64(int64_t __p0) {
43779  int64_t __ret;
43780  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43781  return __ret;
43782}
43783#endif
43784
43785#ifdef __LITTLE_ENDIAN__
43786__ai uint64_t vclezd_f64(float64_t __p0) {
43787  uint64_t __ret;
43788  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43789  return __ret;
43790}
43791#else
43792__ai uint64_t vclezd_f64(float64_t __p0) {
43793  uint64_t __ret;
43794  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43795  return __ret;
43796}
43797#endif
43798
43799#ifdef __LITTLE_ENDIAN__
43800__ai uint32_t vclezs_f32(float32_t __p0) {
43801  uint32_t __ret;
43802  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43803  return __ret;
43804}
43805#else
43806__ai uint32_t vclezs_f32(float32_t __p0) {
43807  uint32_t __ret;
43808  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43809  return __ret;
43810}
43811#endif
43812
43813#ifdef __LITTLE_ENDIAN__
43814__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43815  uint64x2_t __ret;
43816  __ret = (uint64x2_t)(__p0 < __p1);
43817  return __ret;
43818}
43819#else
43820__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43821  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43822  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43823  uint64x2_t __ret;
43824  __ret = (uint64x2_t)(__rev0 < __rev1);
43825  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43826  return __ret;
43827}
43828#endif
43829
43830#ifdef __LITTLE_ENDIAN__
43831__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43832  uint64x2_t __ret;
43833  __ret = (uint64x2_t)(__p0 < __p1);
43834  return __ret;
43835}
43836#else
43837__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43838  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43839  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43840  uint64x2_t __ret;
43841  __ret = (uint64x2_t)(__rev0 < __rev1);
43842  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43843  return __ret;
43844}
43845#endif
43846
43847#ifdef __LITTLE_ENDIAN__
43848__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43849  uint64x2_t __ret;
43850  __ret = (uint64x2_t)(__p0 < __p1);
43851  return __ret;
43852}
43853#else
43854__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43855  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43856  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43857  uint64x2_t __ret;
43858  __ret = (uint64x2_t)(__rev0 < __rev1);
43859  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43860  return __ret;
43861}
43862#endif
43863
43864#ifdef __LITTLE_ENDIAN__
43865__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43866  uint64x1_t __ret;
43867  __ret = (uint64x1_t)(__p0 < __p1);
43868  return __ret;
43869}
43870#else
43871__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43872  uint64x1_t __ret;
43873  __ret = (uint64x1_t)(__p0 < __p1);
43874  return __ret;
43875}
43876#endif
43877
43878#ifdef __LITTLE_ENDIAN__
43879__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43880  uint64x1_t __ret;
43881  __ret = (uint64x1_t)(__p0 < __p1);
43882  return __ret;
43883}
43884#else
43885__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43886  uint64x1_t __ret;
43887  __ret = (uint64x1_t)(__p0 < __p1);
43888  return __ret;
43889}
43890#endif
43891
43892#ifdef __LITTLE_ENDIAN__
43893__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43894  uint64x1_t __ret;
43895  __ret = (uint64x1_t)(__p0 < __p1);
43896  return __ret;
43897}
43898#else
43899__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43900  uint64x1_t __ret;
43901  __ret = (uint64x1_t)(__p0 < __p1);
43902  return __ret;
43903}
43904#endif
43905
43906#ifdef __LITTLE_ENDIAN__
43907__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43908  uint64_t __ret;
43909  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43910  return __ret;
43911}
43912#else
43913__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43914  uint64_t __ret;
43915  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43916  return __ret;
43917}
43918#endif
43919
43920#ifdef __LITTLE_ENDIAN__
43921__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43922  int64_t __ret;
43923  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43924  return __ret;
43925}
43926#else
43927__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43928  int64_t __ret;
43929  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43930  return __ret;
43931}
43932#endif
43933
43934#ifdef __LITTLE_ENDIAN__
43935__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43936  uint64_t __ret;
43937  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43938  return __ret;
43939}
43940#else
43941__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43942  uint64_t __ret;
43943  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43944  return __ret;
43945}
43946#endif
43947
43948#ifdef __LITTLE_ENDIAN__
43949__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43950  uint32_t __ret;
43951  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43952  return __ret;
43953}
43954#else
43955__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43956  uint32_t __ret;
43957  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43958  return __ret;
43959}
43960#endif
43961
43962#ifdef __LITTLE_ENDIAN__
43963__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43964  uint8x16_t __ret;
43965  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
43966  return __ret;
43967}
43968#else
43969__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43970  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43971  uint8x16_t __ret;
43972  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
43973  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43974  return __ret;
43975}
43976#endif
43977
43978#ifdef __LITTLE_ENDIAN__
43979__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43980  uint64x2_t __ret;
43981  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43982  return __ret;
43983}
43984#else
43985__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43986  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43987  uint64x2_t __ret;
43988  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43990  return __ret;
43991}
43992#endif
43993
43994#ifdef __LITTLE_ENDIAN__
43995__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43996  uint32x4_t __ret;
43997  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43998  return __ret;
43999}
44000#else
44001__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
44002  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44003  uint32x4_t __ret;
44004  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
44005  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44006  return __ret;
44007}
44008#endif
44009
44010#ifdef __LITTLE_ENDIAN__
44011__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
44012  uint32x4_t __ret;
44013  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
44014  return __ret;
44015}
44016#else
44017__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
44018  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44019  uint32x4_t __ret;
44020  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
44021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44022  return __ret;
44023}
44024#endif
44025
44026#ifdef __LITTLE_ENDIAN__
44027__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
44028  uint64x2_t __ret;
44029  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
44030  return __ret;
44031}
44032#else
44033__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
44034  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44035  uint64x2_t __ret;
44036  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
44037  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44038  return __ret;
44039}
44040#endif
44041
44042#ifdef __LITTLE_ENDIAN__
44043__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44044  uint16x8_t __ret;
44045  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
44046  return __ret;
44047}
44048#else
44049__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44050  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44051  uint16x8_t __ret;
44052  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
44053  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44054  return __ret;
44055}
44056#endif
44057
44058#ifdef __LITTLE_ENDIAN__
44059__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44060  uint8x8_t __ret;
44061  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
44062  return __ret;
44063}
44064#else
44065__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44066  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44067  uint8x8_t __ret;
44068  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
44069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44070  return __ret;
44071}
44072#endif
44073
44074#ifdef __LITTLE_ENDIAN__
44075__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44076  uint64x1_t __ret;
44077  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44078  return __ret;
44079}
44080#else
44081__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44082  uint64x1_t __ret;
44083  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44084  return __ret;
44085}
44086#endif
44087
44088#ifdef __LITTLE_ENDIAN__
44089__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44090  uint32x2_t __ret;
44091  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44092  return __ret;
44093}
44094#else
44095__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44096  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44097  uint32x2_t __ret;
44098  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44099  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44100  return __ret;
44101}
44102#endif
44103
44104#ifdef __LITTLE_ENDIAN__
44105__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44106  uint32x2_t __ret;
44107  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44108  return __ret;
44109}
44110#else
44111__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44112  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44113  uint32x2_t __ret;
44114  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44115  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44116  return __ret;
44117}
44118#endif
44119
44120#ifdef __LITTLE_ENDIAN__
44121__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44122  uint64x1_t __ret;
44123  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44124  return __ret;
44125}
44126#else
44127__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44128  uint64x1_t __ret;
44129  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44130  return __ret;
44131}
44132#endif
44133
44134#ifdef __LITTLE_ENDIAN__
44135__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44136  uint16x4_t __ret;
44137  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
44138  return __ret;
44139}
44140#else
44141__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44142  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44143  uint16x4_t __ret;
44144  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
44145  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44146  return __ret;
44147}
44148#endif
44149
44150#ifdef __LITTLE_ENDIAN__
44151__ai int64_t vcltzd_s64(int64_t __p0) {
44152  int64_t __ret;
44153  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44154  return __ret;
44155}
44156#else
44157__ai int64_t vcltzd_s64(int64_t __p0) {
44158  int64_t __ret;
44159  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44160  return __ret;
44161}
44162#endif
44163
44164#ifdef __LITTLE_ENDIAN__
44165__ai uint64_t vcltzd_f64(float64_t __p0) {
44166  uint64_t __ret;
44167  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44168  return __ret;
44169}
44170#else
44171__ai uint64_t vcltzd_f64(float64_t __p0) {
44172  uint64_t __ret;
44173  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44174  return __ret;
44175}
44176#endif
44177
44178#ifdef __LITTLE_ENDIAN__
44179__ai uint32_t vcltzs_f32(float32_t __p0) {
44180  uint32_t __ret;
44181  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44182  return __ret;
44183}
44184#else
44185__ai uint32_t vcltzs_f32(float32_t __p0) {
44186  uint32_t __ret;
44187  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44188  return __ret;
44189}
44190#endif
44191
44192#ifdef __LITTLE_ENDIAN__
44193__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44194  poly64x2_t __ret;
44195  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44196  return __ret;
44197}
44198#else
44199__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44200  poly64x2_t __ret;
44201  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44203  return __ret;
44204}
44205#endif
44206
44207#ifdef __LITTLE_ENDIAN__
44208__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44209  float64x2_t __ret;
44210  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44211  return __ret;
44212}
44213#else
44214__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44215  float64x2_t __ret;
44216  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44217  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44218  return __ret;
44219}
44220#endif
44221
44222#ifdef __LITTLE_ENDIAN__
44223#define vcopyq_lane_p8(__p0_0, __p1_0, __p2_0, __p3_0) __extension__ ({ \
44224  poly8x16_t __s0_0 = __p0_0; \
44225  poly8x8_t __s2_0 = __p2_0; \
44226  poly8x16_t __ret_0; \
44227  __ret_0 = vsetq_lane_p8(vget_lane_p8(__s2_0, __p3_0), __s0_0, __p1_0); \
44228  __ret_0; \
44229})
44230#else
44231#define vcopyq_lane_p8(__p0_1, __p1_1, __p2_1, __p3_1) __extension__ ({ \
44232  poly8x16_t __s0_1 = __p0_1; \
44233  poly8x8_t __s2_1 = __p2_1; \
44234  poly8x16_t __rev0_1;  __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44235  poly8x8_t __rev2_1;  __rev2_1 = __builtin_shufflevector(__s2_1, __s2_1, 7, 6, 5, 4, 3, 2, 1, 0); \
44236  poly8x16_t __ret_1; \
44237  __ret_1 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_1, __p3_1), __rev0_1, __p1_1); \
44238  __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44239  __ret_1; \
44240})
44241#endif
44242
44243#ifdef __LITTLE_ENDIAN__
44244#define vcopyq_lane_p16(__p0_2, __p1_2, __p2_2, __p3_2) __extension__ ({ \
44245  poly16x8_t __s0_2 = __p0_2; \
44246  poly16x4_t __s2_2 = __p2_2; \
44247  poly16x8_t __ret_2; \
44248  __ret_2 = vsetq_lane_p16(vget_lane_p16(__s2_2, __p3_2), __s0_2, __p1_2); \
44249  __ret_2; \
44250})
44251#else
44252#define vcopyq_lane_p16(__p0_3, __p1_3, __p2_3, __p3_3) __extension__ ({ \
44253  poly16x8_t __s0_3 = __p0_3; \
44254  poly16x4_t __s2_3 = __p2_3; \
44255  poly16x8_t __rev0_3;  __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44256  poly16x4_t __rev2_3;  __rev2_3 = __builtin_shufflevector(__s2_3, __s2_3, 3, 2, 1, 0); \
44257  poly16x8_t __ret_3; \
44258  __ret_3 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_3, __p3_3), __rev0_3, __p1_3); \
44259  __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44260  __ret_3; \
44261})
44262#endif
44263
44264#ifdef __LITTLE_ENDIAN__
44265#define vcopyq_lane_u8(__p0_4, __p1_4, __p2_4, __p3_4) __extension__ ({ \
44266  uint8x16_t __s0_4 = __p0_4; \
44267  uint8x8_t __s2_4 = __p2_4; \
44268  uint8x16_t __ret_4; \
44269  __ret_4 = vsetq_lane_u8(vget_lane_u8(__s2_4, __p3_4), __s0_4, __p1_4); \
44270  __ret_4; \
44271})
44272#else
44273#define vcopyq_lane_u8(__p0_5, __p1_5, __p2_5, __p3_5) __extension__ ({ \
44274  uint8x16_t __s0_5 = __p0_5; \
44275  uint8x8_t __s2_5 = __p2_5; \
44276  uint8x16_t __rev0_5;  __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44277  uint8x8_t __rev2_5;  __rev2_5 = __builtin_shufflevector(__s2_5, __s2_5, 7, 6, 5, 4, 3, 2, 1, 0); \
44278  uint8x16_t __ret_5; \
44279  __ret_5 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_5, __p3_5), __rev0_5, __p1_5); \
44280  __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44281  __ret_5; \
44282})
44283#endif
44284
44285#ifdef __LITTLE_ENDIAN__
44286#define vcopyq_lane_u32(__p0_6, __p1_6, __p2_6, __p3_6) __extension__ ({ \
44287  uint32x4_t __s0_6 = __p0_6; \
44288  uint32x2_t __s2_6 = __p2_6; \
44289  uint32x4_t __ret_6; \
44290  __ret_6 = vsetq_lane_u32(vget_lane_u32(__s2_6, __p3_6), __s0_6, __p1_6); \
44291  __ret_6; \
44292})
44293#else
44294#define vcopyq_lane_u32(__p0_7, __p1_7, __p2_7, __p3_7) __extension__ ({ \
44295  uint32x4_t __s0_7 = __p0_7; \
44296  uint32x2_t __s2_7 = __p2_7; \
44297  uint32x4_t __rev0_7;  __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
44298  uint32x2_t __rev2_7;  __rev2_7 = __builtin_shufflevector(__s2_7, __s2_7, 1, 0); \
44299  uint32x4_t __ret_7; \
44300  __ret_7 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_7, __p3_7), __rev0_7, __p1_7); \
44301  __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 3, 2, 1, 0); \
44302  __ret_7; \
44303})
44304#endif
44305
44306#ifdef __LITTLE_ENDIAN__
44307#define vcopyq_lane_u64(__p0_8, __p1_8, __p2_8, __p3_8) __extension__ ({ \
44308  uint64x2_t __s0_8 = __p0_8; \
44309  uint64x1_t __s2_8 = __p2_8; \
44310  uint64x2_t __ret_8; \
44311  __ret_8 = vsetq_lane_u64(vget_lane_u64(__s2_8, __p3_8), __s0_8, __p1_8); \
44312  __ret_8; \
44313})
44314#else
44315#define vcopyq_lane_u64(__p0_9, __p1_9, __p2_9, __p3_9) __extension__ ({ \
44316  uint64x2_t __s0_9 = __p0_9; \
44317  uint64x1_t __s2_9 = __p2_9; \
44318  uint64x2_t __rev0_9;  __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 1, 0); \
44319  uint64x2_t __ret_9; \
44320  __ret_9 = __noswap_vsetq_lane_u64(__noswap_vget_lane_u64(__s2_9, __p3_9), __rev0_9, __p1_9); \
44321  __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 1, 0); \
44322  __ret_9; \
44323})
44324#endif
44325
44326#ifdef __LITTLE_ENDIAN__
44327#define vcopyq_lane_u16(__p0_10, __p1_10, __p2_10, __p3_10) __extension__ ({ \
44328  uint16x8_t __s0_10 = __p0_10; \
44329  uint16x4_t __s2_10 = __p2_10; \
44330  uint16x8_t __ret_10; \
44331  __ret_10 = vsetq_lane_u16(vget_lane_u16(__s2_10, __p3_10), __s0_10, __p1_10); \
44332  __ret_10; \
44333})
44334#else
44335#define vcopyq_lane_u16(__p0_11, __p1_11, __p2_11, __p3_11) __extension__ ({ \
44336  uint16x8_t __s0_11 = __p0_11; \
44337  uint16x4_t __s2_11 = __p2_11; \
44338  uint16x8_t __rev0_11;  __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44339  uint16x4_t __rev2_11;  __rev2_11 = __builtin_shufflevector(__s2_11, __s2_11, 3, 2, 1, 0); \
44340  uint16x8_t __ret_11; \
44341  __ret_11 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_11, __p3_11), __rev0_11, __p1_11); \
44342  __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44343  __ret_11; \
44344})
44345#endif
44346
44347#ifdef __LITTLE_ENDIAN__
44348#define vcopyq_lane_s8(__p0_12, __p1_12, __p2_12, __p3_12) __extension__ ({ \
44349  int8x16_t __s0_12 = __p0_12; \
44350  int8x8_t __s2_12 = __p2_12; \
44351  int8x16_t __ret_12; \
44352  __ret_12 = vsetq_lane_s8(vget_lane_s8(__s2_12, __p3_12), __s0_12, __p1_12); \
44353  __ret_12; \
44354})
44355#else
44356#define vcopyq_lane_s8(__p0_13, __p1_13, __p2_13, __p3_13) __extension__ ({ \
44357  int8x16_t __s0_13 = __p0_13; \
44358  int8x8_t __s2_13 = __p2_13; \
44359  int8x16_t __rev0_13;  __rev0_13 = __builtin_shufflevector(__s0_13, __s0_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44360  int8x8_t __rev2_13;  __rev2_13 = __builtin_shufflevector(__s2_13, __s2_13, 7, 6, 5, 4, 3, 2, 1, 0); \
44361  int8x16_t __ret_13; \
44362  __ret_13 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_13, __p3_13), __rev0_13, __p1_13); \
44363  __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44364  __ret_13; \
44365})
44366#endif
44367
44368#ifdef __LITTLE_ENDIAN__
44369#define vcopyq_lane_f32(__p0_14, __p1_14, __p2_14, __p3_14) __extension__ ({ \
44370  float32x4_t __s0_14 = __p0_14; \
44371  float32x2_t __s2_14 = __p2_14; \
44372  float32x4_t __ret_14; \
44373  __ret_14 = vsetq_lane_f32(vget_lane_f32(__s2_14, __p3_14), __s0_14, __p1_14); \
44374  __ret_14; \
44375})
44376#else
44377#define vcopyq_lane_f32(__p0_15, __p1_15, __p2_15, __p3_15) __extension__ ({ \
44378  float32x4_t __s0_15 = __p0_15; \
44379  float32x2_t __s2_15 = __p2_15; \
44380  float32x4_t __rev0_15;  __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
44381  float32x2_t __rev2_15;  __rev2_15 = __builtin_shufflevector(__s2_15, __s2_15, 1, 0); \
44382  float32x4_t __ret_15; \
44383  __ret_15 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_15, __p3_15), __rev0_15, __p1_15); \
44384  __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 3, 2, 1, 0); \
44385  __ret_15; \
44386})
44387#endif
44388
44389#ifdef __LITTLE_ENDIAN__
44390#define vcopyq_lane_s32(__p0_16, __p1_16, __p2_16, __p3_16) __extension__ ({ \
44391  int32x4_t __s0_16 = __p0_16; \
44392  int32x2_t __s2_16 = __p2_16; \
44393  int32x4_t __ret_16; \
44394  __ret_16 = vsetq_lane_s32(vget_lane_s32(__s2_16, __p3_16), __s0_16, __p1_16); \
44395  __ret_16; \
44396})
44397#else
44398#define vcopyq_lane_s32(__p0_17, __p1_17, __p2_17, __p3_17) __extension__ ({ \
44399  int32x4_t __s0_17 = __p0_17; \
44400  int32x2_t __s2_17 = __p2_17; \
44401  int32x4_t __rev0_17;  __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 3, 2, 1, 0); \
44402  int32x2_t __rev2_17;  __rev2_17 = __builtin_shufflevector(__s2_17, __s2_17, 1, 0); \
44403  int32x4_t __ret_17; \
44404  __ret_17 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_17, __p3_17), __rev0_17, __p1_17); \
44405  __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 3, 2, 1, 0); \
44406  __ret_17; \
44407})
44408#endif
44409
44410#ifdef __LITTLE_ENDIAN__
44411#define vcopyq_lane_s64(__p0_18, __p1_18, __p2_18, __p3_18) __extension__ ({ \
44412  int64x2_t __s0_18 = __p0_18; \
44413  int64x1_t __s2_18 = __p2_18; \
44414  int64x2_t __ret_18; \
44415  __ret_18 = vsetq_lane_s64(vget_lane_s64(__s2_18, __p3_18), __s0_18, __p1_18); \
44416  __ret_18; \
44417})
44418#else
44419#define vcopyq_lane_s64(__p0_19, __p1_19, __p2_19, __p3_19) __extension__ ({ \
44420  int64x2_t __s0_19 = __p0_19; \
44421  int64x1_t __s2_19 = __p2_19; \
44422  int64x2_t __rev0_19;  __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
44423  int64x2_t __ret_19; \
44424  __ret_19 = __noswap_vsetq_lane_s64(__noswap_vget_lane_s64(__s2_19, __p3_19), __rev0_19, __p1_19); \
44425  __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 1, 0); \
44426  __ret_19; \
44427})
44428#endif
44429
44430#ifdef __LITTLE_ENDIAN__
44431#define vcopyq_lane_s16(__p0_20, __p1_20, __p2_20, __p3_20) __extension__ ({ \
44432  int16x8_t __s0_20 = __p0_20; \
44433  int16x4_t __s2_20 = __p2_20; \
44434  int16x8_t __ret_20; \
44435  __ret_20 = vsetq_lane_s16(vget_lane_s16(__s2_20, __p3_20), __s0_20, __p1_20); \
44436  __ret_20; \
44437})
44438#else
44439#define vcopyq_lane_s16(__p0_21, __p1_21, __p2_21, __p3_21) __extension__ ({ \
44440  int16x8_t __s0_21 = __p0_21; \
44441  int16x4_t __s2_21 = __p2_21; \
44442  int16x8_t __rev0_21;  __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44443  int16x4_t __rev2_21;  __rev2_21 = __builtin_shufflevector(__s2_21, __s2_21, 3, 2, 1, 0); \
44444  int16x8_t __ret_21; \
44445  __ret_21 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_21, __p3_21), __rev0_21, __p1_21); \
44446  __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44447  __ret_21; \
44448})
44449#endif
44450
44451#ifdef __LITTLE_ENDIAN__
44452#define vcopy_lane_p8(__p0_22, __p1_22, __p2_22, __p3_22) __extension__ ({ \
44453  poly8x8_t __s0_22 = __p0_22; \
44454  poly8x8_t __s2_22 = __p2_22; \
44455  poly8x8_t __ret_22; \
44456  __ret_22 = vset_lane_p8(vget_lane_p8(__s2_22, __p3_22), __s0_22, __p1_22); \
44457  __ret_22; \
44458})
44459#else
44460#define vcopy_lane_p8(__p0_23, __p1_23, __p2_23, __p3_23) __extension__ ({ \
44461  poly8x8_t __s0_23 = __p0_23; \
44462  poly8x8_t __s2_23 = __p2_23; \
44463  poly8x8_t __rev0_23;  __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44464  poly8x8_t __rev2_23;  __rev2_23 = __builtin_shufflevector(__s2_23, __s2_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44465  poly8x8_t __ret_23; \
44466  __ret_23 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_23, __p3_23), __rev0_23, __p1_23); \
44467  __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44468  __ret_23; \
44469})
44470#endif
44471
44472#ifdef __LITTLE_ENDIAN__
44473#define vcopy_lane_p16(__p0_24, __p1_24, __p2_24, __p3_24) __extension__ ({ \
44474  poly16x4_t __s0_24 = __p0_24; \
44475  poly16x4_t __s2_24 = __p2_24; \
44476  poly16x4_t __ret_24; \
44477  __ret_24 = vset_lane_p16(vget_lane_p16(__s2_24, __p3_24), __s0_24, __p1_24); \
44478  __ret_24; \
44479})
44480#else
44481#define vcopy_lane_p16(__p0_25, __p1_25, __p2_25, __p3_25) __extension__ ({ \
44482  poly16x4_t __s0_25 = __p0_25; \
44483  poly16x4_t __s2_25 = __p2_25; \
44484  poly16x4_t __rev0_25;  __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
44485  poly16x4_t __rev2_25;  __rev2_25 = __builtin_shufflevector(__s2_25, __s2_25, 3, 2, 1, 0); \
44486  poly16x4_t __ret_25; \
44487  __ret_25 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_25, __p3_25), __rev0_25, __p1_25); \
44488  __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 3, 2, 1, 0); \
44489  __ret_25; \
44490})
44491#endif
44492
44493#ifdef __LITTLE_ENDIAN__
44494#define vcopy_lane_u8(__p0_26, __p1_26, __p2_26, __p3_26) __extension__ ({ \
44495  uint8x8_t __s0_26 = __p0_26; \
44496  uint8x8_t __s2_26 = __p2_26; \
44497  uint8x8_t __ret_26; \
44498  __ret_26 = vset_lane_u8(vget_lane_u8(__s2_26, __p3_26), __s0_26, __p1_26); \
44499  __ret_26; \
44500})
44501#else
44502#define vcopy_lane_u8(__p0_27, __p1_27, __p2_27, __p3_27) __extension__ ({ \
44503  uint8x8_t __s0_27 = __p0_27; \
44504  uint8x8_t __s2_27 = __p2_27; \
44505  uint8x8_t __rev0_27;  __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44506  uint8x8_t __rev2_27;  __rev2_27 = __builtin_shufflevector(__s2_27, __s2_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44507  uint8x8_t __ret_27; \
44508  __ret_27 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_27, __p3_27), __rev0_27, __p1_27); \
44509  __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44510  __ret_27; \
44511})
44512#endif
44513
44514#ifdef __LITTLE_ENDIAN__
44515#define vcopy_lane_u32(__p0_28, __p1_28, __p2_28, __p3_28) __extension__ ({ \
44516  uint32x2_t __s0_28 = __p0_28; \
44517  uint32x2_t __s2_28 = __p2_28; \
44518  uint32x2_t __ret_28; \
44519  __ret_28 = vset_lane_u32(vget_lane_u32(__s2_28, __p3_28), __s0_28, __p1_28); \
44520  __ret_28; \
44521})
44522#else
44523#define vcopy_lane_u32(__p0_29, __p1_29, __p2_29, __p3_29) __extension__ ({ \
44524  uint32x2_t __s0_29 = __p0_29; \
44525  uint32x2_t __s2_29 = __p2_29; \
44526  uint32x2_t __rev0_29;  __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
44527  uint32x2_t __rev2_29;  __rev2_29 = __builtin_shufflevector(__s2_29, __s2_29, 1, 0); \
44528  uint32x2_t __ret_29; \
44529  __ret_29 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_29, __p3_29), __rev0_29, __p1_29); \
44530  __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
44531  __ret_29; \
44532})
44533#endif
44534
44535#ifdef __LITTLE_ENDIAN__
44536#define vcopy_lane_u64(__p0_30, __p1_30, __p2_30, __p3_30) __extension__ ({ \
44537  uint64x1_t __s0_30 = __p0_30; \
44538  uint64x1_t __s2_30 = __p2_30; \
44539  uint64x1_t __ret_30; \
44540  __ret_30 = vset_lane_u64(vget_lane_u64(__s2_30, __p3_30), __s0_30, __p1_30); \
44541  __ret_30; \
44542})
44543#else
44544#define vcopy_lane_u64(__p0_31, __p1_31, __p2_31, __p3_31) __extension__ ({ \
44545  uint64x1_t __s0_31 = __p0_31; \
44546  uint64x1_t __s2_31 = __p2_31; \
44547  uint64x1_t __ret_31; \
44548  __ret_31 = __noswap_vset_lane_u64(__noswap_vget_lane_u64(__s2_31, __p3_31), __s0_31, __p1_31); \
44549  __ret_31; \
44550})
44551#endif
44552
44553#ifdef __LITTLE_ENDIAN__
44554#define vcopy_lane_u16(__p0_32, __p1_32, __p2_32, __p3_32) __extension__ ({ \
44555  uint16x4_t __s0_32 = __p0_32; \
44556  uint16x4_t __s2_32 = __p2_32; \
44557  uint16x4_t __ret_32; \
44558  __ret_32 = vset_lane_u16(vget_lane_u16(__s2_32, __p3_32), __s0_32, __p1_32); \
44559  __ret_32; \
44560})
44561#else
44562#define vcopy_lane_u16(__p0_33, __p1_33, __p2_33, __p3_33) __extension__ ({ \
44563  uint16x4_t __s0_33 = __p0_33; \
44564  uint16x4_t __s2_33 = __p2_33; \
44565  uint16x4_t __rev0_33;  __rev0_33 = __builtin_shufflevector(__s0_33, __s0_33, 3, 2, 1, 0); \
44566  uint16x4_t __rev2_33;  __rev2_33 = __builtin_shufflevector(__s2_33, __s2_33, 3, 2, 1, 0); \
44567  uint16x4_t __ret_33; \
44568  __ret_33 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_33, __p3_33), __rev0_33, __p1_33); \
44569  __ret_33 = __builtin_shufflevector(__ret_33, __ret_33, 3, 2, 1, 0); \
44570  __ret_33; \
44571})
44572#endif
44573
44574#ifdef __LITTLE_ENDIAN__
44575#define vcopy_lane_s8(__p0_34, __p1_34, __p2_34, __p3_34) __extension__ ({ \
44576  int8x8_t __s0_34 = __p0_34; \
44577  int8x8_t __s2_34 = __p2_34; \
44578  int8x8_t __ret_34; \
44579  __ret_34 = vset_lane_s8(vget_lane_s8(__s2_34, __p3_34), __s0_34, __p1_34); \
44580  __ret_34; \
44581})
44582#else
44583#define vcopy_lane_s8(__p0_35, __p1_35, __p2_35, __p3_35) __extension__ ({ \
44584  int8x8_t __s0_35 = __p0_35; \
44585  int8x8_t __s2_35 = __p2_35; \
44586  int8x8_t __rev0_35;  __rev0_35 = __builtin_shufflevector(__s0_35, __s0_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44587  int8x8_t __rev2_35;  __rev2_35 = __builtin_shufflevector(__s2_35, __s2_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44588  int8x8_t __ret_35; \
44589  __ret_35 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_35, __p3_35), __rev0_35, __p1_35); \
44590  __ret_35 = __builtin_shufflevector(__ret_35, __ret_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44591  __ret_35; \
44592})
44593#endif
44594
44595#ifdef __LITTLE_ENDIAN__
44596#define vcopy_lane_f32(__p0_36, __p1_36, __p2_36, __p3_36) __extension__ ({ \
44597  float32x2_t __s0_36 = __p0_36; \
44598  float32x2_t __s2_36 = __p2_36; \
44599  float32x2_t __ret_36; \
44600  __ret_36 = vset_lane_f32(vget_lane_f32(__s2_36, __p3_36), __s0_36, __p1_36); \
44601  __ret_36; \
44602})
44603#else
44604#define vcopy_lane_f32(__p0_37, __p1_37, __p2_37, __p3_37) __extension__ ({ \
44605  float32x2_t __s0_37 = __p0_37; \
44606  float32x2_t __s2_37 = __p2_37; \
44607  float32x2_t __rev0_37;  __rev0_37 = __builtin_shufflevector(__s0_37, __s0_37, 1, 0); \
44608  float32x2_t __rev2_37;  __rev2_37 = __builtin_shufflevector(__s2_37, __s2_37, 1, 0); \
44609  float32x2_t __ret_37; \
44610  __ret_37 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_37, __p3_37), __rev0_37, __p1_37); \
44611  __ret_37 = __builtin_shufflevector(__ret_37, __ret_37, 1, 0); \
44612  __ret_37; \
44613})
44614#endif
44615
44616#ifdef __LITTLE_ENDIAN__
44617#define vcopy_lane_s32(__p0_38, __p1_38, __p2_38, __p3_38) __extension__ ({ \
44618  int32x2_t __s0_38 = __p0_38; \
44619  int32x2_t __s2_38 = __p2_38; \
44620  int32x2_t __ret_38; \
44621  __ret_38 = vset_lane_s32(vget_lane_s32(__s2_38, __p3_38), __s0_38, __p1_38); \
44622  __ret_38; \
44623})
44624#else
44625#define vcopy_lane_s32(__p0_39, __p1_39, __p2_39, __p3_39) __extension__ ({ \
44626  int32x2_t __s0_39 = __p0_39; \
44627  int32x2_t __s2_39 = __p2_39; \
44628  int32x2_t __rev0_39;  __rev0_39 = __builtin_shufflevector(__s0_39, __s0_39, 1, 0); \
44629  int32x2_t __rev2_39;  __rev2_39 = __builtin_shufflevector(__s2_39, __s2_39, 1, 0); \
44630  int32x2_t __ret_39; \
44631  __ret_39 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_39, __p3_39), __rev0_39, __p1_39); \
44632  __ret_39 = __builtin_shufflevector(__ret_39, __ret_39, 1, 0); \
44633  __ret_39; \
44634})
44635#endif
44636
44637#ifdef __LITTLE_ENDIAN__
44638#define vcopy_lane_s64(__p0_40, __p1_40, __p2_40, __p3_40) __extension__ ({ \
44639  int64x1_t __s0_40 = __p0_40; \
44640  int64x1_t __s2_40 = __p2_40; \
44641  int64x1_t __ret_40; \
44642  __ret_40 = vset_lane_s64(vget_lane_s64(__s2_40, __p3_40), __s0_40, __p1_40); \
44643  __ret_40; \
44644})
44645#else
44646#define vcopy_lane_s64(__p0_41, __p1_41, __p2_41, __p3_41) __extension__ ({ \
44647  int64x1_t __s0_41 = __p0_41; \
44648  int64x1_t __s2_41 = __p2_41; \
44649  int64x1_t __ret_41; \
44650  __ret_41 = __noswap_vset_lane_s64(__noswap_vget_lane_s64(__s2_41, __p3_41), __s0_41, __p1_41); \
44651  __ret_41; \
44652})
44653#endif
44654
44655#ifdef __LITTLE_ENDIAN__
44656#define vcopy_lane_s16(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
44657  int16x4_t __s0_42 = __p0_42; \
44658  int16x4_t __s2_42 = __p2_42; \
44659  int16x4_t __ret_42; \
44660  __ret_42 = vset_lane_s16(vget_lane_s16(__s2_42, __p3_42), __s0_42, __p1_42); \
44661  __ret_42; \
44662})
44663#else
44664#define vcopy_lane_s16(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
44665  int16x4_t __s0_43 = __p0_43; \
44666  int16x4_t __s2_43 = __p2_43; \
44667  int16x4_t __rev0_43;  __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
44668  int16x4_t __rev2_43;  __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 3, 2, 1, 0); \
44669  int16x4_t __ret_43; \
44670  __ret_43 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_43, __p3_43), __rev0_43, __p1_43); \
44671  __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
44672  __ret_43; \
44673})
44674#endif
44675
44676#ifdef __LITTLE_ENDIAN__
44677#define vcopyq_laneq_p8(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
44678  poly8x16_t __s0_44 = __p0_44; \
44679  poly8x16_t __s2_44 = __p2_44; \
44680  poly8x16_t __ret_44; \
44681  __ret_44 = vsetq_lane_p8(vgetq_lane_p8(__s2_44, __p3_44), __s0_44, __p1_44); \
44682  __ret_44; \
44683})
44684#else
44685#define vcopyq_laneq_p8(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
44686  poly8x16_t __s0_45 = __p0_45; \
44687  poly8x16_t __s2_45 = __p2_45; \
44688  poly8x16_t __rev0_45;  __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44689  poly8x16_t __rev2_45;  __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44690  poly8x16_t __ret_45; \
44691  __ret_45 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_45, __p3_45), __rev0_45, __p1_45); \
44692  __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44693  __ret_45; \
44694})
44695#endif
44696
44697#ifdef __LITTLE_ENDIAN__
44698#define vcopyq_laneq_p16(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
44699  poly16x8_t __s0_46 = __p0_46; \
44700  poly16x8_t __s2_46 = __p2_46; \
44701  poly16x8_t __ret_46; \
44702  __ret_46 = vsetq_lane_p16(vgetq_lane_p16(__s2_46, __p3_46), __s0_46, __p1_46); \
44703  __ret_46; \
44704})
44705#else
44706#define vcopyq_laneq_p16(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
44707  poly16x8_t __s0_47 = __p0_47; \
44708  poly16x8_t __s2_47 = __p2_47; \
44709  poly16x8_t __rev0_47;  __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44710  poly16x8_t __rev2_47;  __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44711  poly16x8_t __ret_47; \
44712  __ret_47 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_47, __p3_47), __rev0_47, __p1_47); \
44713  __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44714  __ret_47; \
44715})
44716#endif
44717
44718#ifdef __LITTLE_ENDIAN__
44719#define vcopyq_laneq_u8(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
44720  uint8x16_t __s0_48 = __p0_48; \
44721  uint8x16_t __s2_48 = __p2_48; \
44722  uint8x16_t __ret_48; \
44723  __ret_48 = vsetq_lane_u8(vgetq_lane_u8(__s2_48, __p3_48), __s0_48, __p1_48); \
44724  __ret_48; \
44725})
44726#else
44727#define vcopyq_laneq_u8(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
44728  uint8x16_t __s0_49 = __p0_49; \
44729  uint8x16_t __s2_49 = __p2_49; \
44730  uint8x16_t __rev0_49;  __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44731  uint8x16_t __rev2_49;  __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44732  uint8x16_t __ret_49; \
44733  __ret_49 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_49, __p3_49), __rev0_49, __p1_49); \
44734  __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44735  __ret_49; \
44736})
44737#endif
44738
44739#ifdef __LITTLE_ENDIAN__
44740#define vcopyq_laneq_u32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
44741  uint32x4_t __s0_50 = __p0_50; \
44742  uint32x4_t __s2_50 = __p2_50; \
44743  uint32x4_t __ret_50; \
44744  __ret_50 = vsetq_lane_u32(vgetq_lane_u32(__s2_50, __p3_50), __s0_50, __p1_50); \
44745  __ret_50; \
44746})
44747#else
44748#define vcopyq_laneq_u32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
44749  uint32x4_t __s0_51 = __p0_51; \
44750  uint32x4_t __s2_51 = __p2_51; \
44751  uint32x4_t __rev0_51;  __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
44752  uint32x4_t __rev2_51;  __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
44753  uint32x4_t __ret_51; \
44754  __ret_51 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_51, __p3_51), __rev0_51, __p1_51); \
44755  __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
44756  __ret_51; \
44757})
44758#endif
44759
44760#ifdef __LITTLE_ENDIAN__
44761#define vcopyq_laneq_u64(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
44762  uint64x2_t __s0_52 = __p0_52; \
44763  uint64x2_t __s2_52 = __p2_52; \
44764  uint64x2_t __ret_52; \
44765  __ret_52 = vsetq_lane_u64(vgetq_lane_u64(__s2_52, __p3_52), __s0_52, __p1_52); \
44766  __ret_52; \
44767})
44768#else
44769#define vcopyq_laneq_u64(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
44770  uint64x2_t __s0_53 = __p0_53; \
44771  uint64x2_t __s2_53 = __p2_53; \
44772  uint64x2_t __rev0_53;  __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
44773  uint64x2_t __rev2_53;  __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
44774  uint64x2_t __ret_53; \
44775  __ret_53 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_53, __p3_53), __rev0_53, __p1_53); \
44776  __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
44777  __ret_53; \
44778})
44779#endif
44780
44781#ifdef __LITTLE_ENDIAN__
44782#define vcopyq_laneq_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
44783  uint16x8_t __s0_54 = __p0_54; \
44784  uint16x8_t __s2_54 = __p2_54; \
44785  uint16x8_t __ret_54; \
44786  __ret_54 = vsetq_lane_u16(vgetq_lane_u16(__s2_54, __p3_54), __s0_54, __p1_54); \
44787  __ret_54; \
44788})
44789#else
44790#define vcopyq_laneq_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
44791  uint16x8_t __s0_55 = __p0_55; \
44792  uint16x8_t __s2_55 = __p2_55; \
44793  uint16x8_t __rev0_55;  __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44794  uint16x8_t __rev2_55;  __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44795  uint16x8_t __ret_55; \
44796  __ret_55 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_55, __p3_55), __rev0_55, __p1_55); \
44797  __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44798  __ret_55; \
44799})
44800#endif
44801
44802#ifdef __LITTLE_ENDIAN__
44803#define vcopyq_laneq_s8(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
44804  int8x16_t __s0_56 = __p0_56; \
44805  int8x16_t __s2_56 = __p2_56; \
44806  int8x16_t __ret_56; \
44807  __ret_56 = vsetq_lane_s8(vgetq_lane_s8(__s2_56, __p3_56), __s0_56, __p1_56); \
44808  __ret_56; \
44809})
44810#else
44811#define vcopyq_laneq_s8(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
44812  int8x16_t __s0_57 = __p0_57; \
44813  int8x16_t __s2_57 = __p2_57; \
44814  int8x16_t __rev0_57;  __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44815  int8x16_t __rev2_57;  __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44816  int8x16_t __ret_57; \
44817  __ret_57 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_57, __p3_57), __rev0_57, __p1_57); \
44818  __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44819  __ret_57; \
44820})
44821#endif
44822
44823#ifdef __LITTLE_ENDIAN__
44824#define vcopyq_laneq_f32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
44825  float32x4_t __s0_58 = __p0_58; \
44826  float32x4_t __s2_58 = __p2_58; \
44827  float32x4_t __ret_58; \
44828  __ret_58 = vsetq_lane_f32(vgetq_lane_f32(__s2_58, __p3_58), __s0_58, __p1_58); \
44829  __ret_58; \
44830})
44831#else
44832#define vcopyq_laneq_f32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
44833  float32x4_t __s0_59 = __p0_59; \
44834  float32x4_t __s2_59 = __p2_59; \
44835  float32x4_t __rev0_59;  __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
44836  float32x4_t __rev2_59;  __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
44837  float32x4_t __ret_59; \
44838  __ret_59 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_59, __p3_59), __rev0_59, __p1_59); \
44839  __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
44840  __ret_59; \
44841})
44842#endif
44843
44844#ifdef __LITTLE_ENDIAN__
44845#define vcopyq_laneq_s32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
44846  int32x4_t __s0_60 = __p0_60; \
44847  int32x4_t __s2_60 = __p2_60; \
44848  int32x4_t __ret_60; \
44849  __ret_60 = vsetq_lane_s32(vgetq_lane_s32(__s2_60, __p3_60), __s0_60, __p1_60); \
44850  __ret_60; \
44851})
44852#else
44853#define vcopyq_laneq_s32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
44854  int32x4_t __s0_61 = __p0_61; \
44855  int32x4_t __s2_61 = __p2_61; \
44856  int32x4_t __rev0_61;  __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
44857  int32x4_t __rev2_61;  __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
44858  int32x4_t __ret_61; \
44859  __ret_61 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_61, __p3_61), __rev0_61, __p1_61); \
44860  __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
44861  __ret_61; \
44862})
44863#endif
44864
44865#ifdef __LITTLE_ENDIAN__
44866#define vcopyq_laneq_s64(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
44867  int64x2_t __s0_62 = __p0_62; \
44868  int64x2_t __s2_62 = __p2_62; \
44869  int64x2_t __ret_62; \
44870  __ret_62 = vsetq_lane_s64(vgetq_lane_s64(__s2_62, __p3_62), __s0_62, __p1_62); \
44871  __ret_62; \
44872})
44873#else
44874#define vcopyq_laneq_s64(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
44875  int64x2_t __s0_63 = __p0_63; \
44876  int64x2_t __s2_63 = __p2_63; \
44877  int64x2_t __rev0_63;  __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
44878  int64x2_t __rev2_63;  __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
44879  int64x2_t __ret_63; \
44880  __ret_63 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_63, __p3_63), __rev0_63, __p1_63); \
44881  __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
44882  __ret_63; \
44883})
44884#endif
44885
44886#ifdef __LITTLE_ENDIAN__
44887#define vcopyq_laneq_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
44888  int16x8_t __s0_64 = __p0_64; \
44889  int16x8_t __s2_64 = __p2_64; \
44890  int16x8_t __ret_64; \
44891  __ret_64 = vsetq_lane_s16(vgetq_lane_s16(__s2_64, __p3_64), __s0_64, __p1_64); \
44892  __ret_64; \
44893})
44894#else
44895#define vcopyq_laneq_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
44896  int16x8_t __s0_65 = __p0_65; \
44897  int16x8_t __s2_65 = __p2_65; \
44898  int16x8_t __rev0_65;  __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44899  int16x8_t __rev2_65;  __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44900  int16x8_t __ret_65; \
44901  __ret_65 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_65, __p3_65), __rev0_65, __p1_65); \
44902  __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44903  __ret_65; \
44904})
44905#endif
44906
44907#ifdef __LITTLE_ENDIAN__
44908#define vcopy_laneq_p8(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
44909  poly8x8_t __s0_66 = __p0_66; \
44910  poly8x16_t __s2_66 = __p2_66; \
44911  poly8x8_t __ret_66; \
44912  __ret_66 = vset_lane_p8(vgetq_lane_p8(__s2_66, __p3_66), __s0_66, __p1_66); \
44913  __ret_66; \
44914})
44915#else
44916#define vcopy_laneq_p8(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
44917  poly8x8_t __s0_67 = __p0_67; \
44918  poly8x16_t __s2_67 = __p2_67; \
44919  poly8x8_t __rev0_67;  __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44920  poly8x16_t __rev2_67;  __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44921  poly8x8_t __ret_67; \
44922  __ret_67 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_67, __p3_67), __rev0_67, __p1_67); \
44923  __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44924  __ret_67; \
44925})
44926#endif
44927
44928#ifdef __LITTLE_ENDIAN__
44929#define vcopy_laneq_p16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
44930  poly16x4_t __s0_68 = __p0_68; \
44931  poly16x8_t __s2_68 = __p2_68; \
44932  poly16x4_t __ret_68; \
44933  __ret_68 = vset_lane_p16(vgetq_lane_p16(__s2_68, __p3_68), __s0_68, __p1_68); \
44934  __ret_68; \
44935})
44936#else
44937#define vcopy_laneq_p16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
44938  poly16x4_t __s0_69 = __p0_69; \
44939  poly16x8_t __s2_69 = __p2_69; \
44940  poly16x4_t __rev0_69;  __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
44941  poly16x8_t __rev2_69;  __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 7, 6, 5, 4, 3, 2, 1, 0); \
44942  poly16x4_t __ret_69; \
44943  __ret_69 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_69, __p3_69), __rev0_69, __p1_69); \
44944  __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
44945  __ret_69; \
44946})
44947#endif
44948
44949#ifdef __LITTLE_ENDIAN__
44950#define vcopy_laneq_u8(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
44951  uint8x8_t __s0_70 = __p0_70; \
44952  uint8x16_t __s2_70 = __p2_70; \
44953  uint8x8_t __ret_70; \
44954  __ret_70 = vset_lane_u8(vgetq_lane_u8(__s2_70, __p3_70), __s0_70, __p1_70); \
44955  __ret_70; \
44956})
44957#else
44958#define vcopy_laneq_u8(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
44959  uint8x8_t __s0_71 = __p0_71; \
44960  uint8x16_t __s2_71 = __p2_71; \
44961  uint8x8_t __rev0_71;  __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44962  uint8x16_t __rev2_71;  __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44963  uint8x8_t __ret_71; \
44964  __ret_71 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_71, __p3_71), __rev0_71, __p1_71); \
44965  __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44966  __ret_71; \
44967})
44968#endif
44969
44970#ifdef __LITTLE_ENDIAN__
44971#define vcopy_laneq_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
44972  uint32x2_t __s0_72 = __p0_72; \
44973  uint32x4_t __s2_72 = __p2_72; \
44974  uint32x2_t __ret_72; \
44975  __ret_72 = vset_lane_u32(vgetq_lane_u32(__s2_72, __p3_72), __s0_72, __p1_72); \
44976  __ret_72; \
44977})
44978#else
44979#define vcopy_laneq_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
44980  uint32x2_t __s0_73 = __p0_73; \
44981  uint32x4_t __s2_73 = __p2_73; \
44982  uint32x2_t __rev0_73;  __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
44983  uint32x4_t __rev2_73;  __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 3, 2, 1, 0); \
44984  uint32x2_t __ret_73; \
44985  __ret_73 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_73, __p3_73), __rev0_73, __p1_73); \
44986  __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
44987  __ret_73; \
44988})
44989#endif
44990
44991#ifdef __LITTLE_ENDIAN__
44992#define vcopy_laneq_u64(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
44993  uint64x1_t __s0_74 = __p0_74; \
44994  uint64x2_t __s2_74 = __p2_74; \
44995  uint64x1_t __ret_74; \
44996  __ret_74 = vset_lane_u64(vgetq_lane_u64(__s2_74, __p3_74), __s0_74, __p1_74); \
44997  __ret_74; \
44998})
44999#else
45000#define vcopy_laneq_u64(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
45001  uint64x1_t __s0_75 = __p0_75; \
45002  uint64x2_t __s2_75 = __p2_75; \
45003  uint64x2_t __rev2_75;  __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 1, 0); \
45004  uint64x1_t __ret_75; \
45005  __ret_75 = __noswap_vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_75, __p3_75), __s0_75, __p1_75); \
45006  __ret_75; \
45007})
45008#endif
45009
45010#ifdef __LITTLE_ENDIAN__
45011#define vcopy_laneq_u16(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
45012  uint16x4_t __s0_76 = __p0_76; \
45013  uint16x8_t __s2_76 = __p2_76; \
45014  uint16x4_t __ret_76; \
45015  __ret_76 = vset_lane_u16(vgetq_lane_u16(__s2_76, __p3_76), __s0_76, __p1_76); \
45016  __ret_76; \
45017})
45018#else
45019#define vcopy_laneq_u16(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
45020  uint16x4_t __s0_77 = __p0_77; \
45021  uint16x8_t __s2_77 = __p2_77; \
45022  uint16x4_t __rev0_77;  __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 3, 2, 1, 0); \
45023  uint16x8_t __rev2_77;  __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 7, 6, 5, 4, 3, 2, 1, 0); \
45024  uint16x4_t __ret_77; \
45025  __ret_77 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_77, __p3_77), __rev0_77, __p1_77); \
45026  __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 3, 2, 1, 0); \
45027  __ret_77; \
45028})
45029#endif
45030
45031#ifdef __LITTLE_ENDIAN__
45032#define vcopy_laneq_s8(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
45033  int8x8_t __s0_78 = __p0_78; \
45034  int8x16_t __s2_78 = __p2_78; \
45035  int8x8_t __ret_78; \
45036  __ret_78 = vset_lane_s8(vgetq_lane_s8(__s2_78, __p3_78), __s0_78, __p1_78); \
45037  __ret_78; \
45038})
45039#else
45040#define vcopy_laneq_s8(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
45041  int8x8_t __s0_79 = __p0_79; \
45042  int8x16_t __s2_79 = __p2_79; \
45043  int8x8_t __rev0_79;  __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45044  int8x16_t __rev2_79;  __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45045  int8x8_t __ret_79; \
45046  __ret_79 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_79, __p3_79), __rev0_79, __p1_79); \
45047  __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45048  __ret_79; \
45049})
45050#endif
45051
45052#ifdef __LITTLE_ENDIAN__
45053#define vcopy_laneq_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
45054  float32x2_t __s0_80 = __p0_80; \
45055  float32x4_t __s2_80 = __p2_80; \
45056  float32x2_t __ret_80; \
45057  __ret_80 = vset_lane_f32(vgetq_lane_f32(__s2_80, __p3_80), __s0_80, __p1_80); \
45058  __ret_80; \
45059})
45060#else
45061#define vcopy_laneq_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
45062  float32x2_t __s0_81 = __p0_81; \
45063  float32x4_t __s2_81 = __p2_81; \
45064  float32x2_t __rev0_81;  __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
45065  float32x4_t __rev2_81;  __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
45066  float32x2_t __ret_81; \
45067  __ret_81 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_81, __p3_81), __rev0_81, __p1_81); \
45068  __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
45069  __ret_81; \
45070})
45071#endif
45072
45073#ifdef __LITTLE_ENDIAN__
45074#define vcopy_laneq_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
45075  int32x2_t __s0_82 = __p0_82; \
45076  int32x4_t __s2_82 = __p2_82; \
45077  int32x2_t __ret_82; \
45078  __ret_82 = vset_lane_s32(vgetq_lane_s32(__s2_82, __p3_82), __s0_82, __p1_82); \
45079  __ret_82; \
45080})
45081#else
45082#define vcopy_laneq_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
45083  int32x2_t __s0_83 = __p0_83; \
45084  int32x4_t __s2_83 = __p2_83; \
45085  int32x2_t __rev0_83;  __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
45086  int32x4_t __rev2_83;  __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 3, 2, 1, 0); \
45087  int32x2_t __ret_83; \
45088  __ret_83 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_83, __p3_83), __rev0_83, __p1_83); \
45089  __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
45090  __ret_83; \
45091})
45092#endif
45093
45094#ifdef __LITTLE_ENDIAN__
45095#define vcopy_laneq_s64(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
45096  int64x1_t __s0_84 = __p0_84; \
45097  int64x2_t __s2_84 = __p2_84; \
45098  int64x1_t __ret_84; \
45099  __ret_84 = vset_lane_s64(vgetq_lane_s64(__s2_84, __p3_84), __s0_84, __p1_84); \
45100  __ret_84; \
45101})
45102#else
45103#define vcopy_laneq_s64(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
45104  int64x1_t __s0_85 = __p0_85; \
45105  int64x2_t __s2_85 = __p2_85; \
45106  int64x2_t __rev2_85;  __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 1, 0); \
45107  int64x1_t __ret_85; \
45108  __ret_85 = __noswap_vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_85, __p3_85), __s0_85, __p1_85); \
45109  __ret_85; \
45110})
45111#endif
45112
45113#ifdef __LITTLE_ENDIAN__
45114#define vcopy_laneq_s16(__p0_86, __p1_86, __p2_86, __p3_86) __extension__ ({ \
45115  int16x4_t __s0_86 = __p0_86; \
45116  int16x8_t __s2_86 = __p2_86; \
45117  int16x4_t __ret_86; \
45118  __ret_86 = vset_lane_s16(vgetq_lane_s16(__s2_86, __p3_86), __s0_86, __p1_86); \
45119  __ret_86; \
45120})
45121#else
45122#define vcopy_laneq_s16(__p0_87, __p1_87, __p2_87, __p3_87) __extension__ ({ \
45123  int16x4_t __s0_87 = __p0_87; \
45124  int16x8_t __s2_87 = __p2_87; \
45125  int16x4_t __rev0_87;  __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
45126  int16x8_t __rev2_87;  __rev2_87 = __builtin_shufflevector(__s2_87, __s2_87, 7, 6, 5, 4, 3, 2, 1, 0); \
45127  int16x4_t __ret_87; \
45128  __ret_87 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_87, __p3_87), __rev0_87, __p1_87); \
45129  __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
45130  __ret_87; \
45131})
45132#endif
45133
45134#ifdef __LITTLE_ENDIAN__
45135__ai poly64x1_t vcreate_p64(uint64_t __p0) {
45136  poly64x1_t __ret;
45137  __ret = (poly64x1_t)(__p0);
45138  return __ret;
45139}
45140#else
45141__ai poly64x1_t vcreate_p64(uint64_t __p0) {
45142  poly64x1_t __ret;
45143  __ret = (poly64x1_t)(__p0);
45144  return __ret;
45145}
45146#endif
45147
45148#ifdef __LITTLE_ENDIAN__
45149__ai float64x1_t vcreate_f64(uint64_t __p0) {
45150  float64x1_t __ret;
45151  __ret = (float64x1_t)(__p0);
45152  return __ret;
45153}
45154#else
45155__ai float64x1_t vcreate_f64(uint64_t __p0) {
45156  float64x1_t __ret;
45157  __ret = (float64x1_t)(__p0);
45158  return __ret;
45159}
45160#endif
45161
45162#ifdef __LITTLE_ENDIAN__
45163__ai float32_t vcvts_f32_s32(int32_t __p0) {
45164  float32_t __ret;
45165  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45166  return __ret;
45167}
45168#else
45169__ai float32_t vcvts_f32_s32(int32_t __p0) {
45170  float32_t __ret;
45171  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45172  return __ret;
45173}
45174#endif
45175
45176#ifdef __LITTLE_ENDIAN__
45177__ai float32_t vcvts_f32_u32(uint32_t __p0) {
45178  float32_t __ret;
45179  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45180  return __ret;
45181}
45182#else
45183__ai float32_t vcvts_f32_u32(uint32_t __p0) {
45184  float32_t __ret;
45185  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45186  return __ret;
45187}
45188#endif
45189
45190#ifdef __LITTLE_ENDIAN__
45191__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45192  float32x2_t __ret;
45193  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45194  return __ret;
45195}
45196#else
45197__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45198  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45199  float32x2_t __ret;
45200  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
45201  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45202  return __ret;
45203}
45204__ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
45205  float32x2_t __ret;
45206  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45207  return __ret;
45208}
45209#endif
45210
45211#ifdef __LITTLE_ENDIAN__
45212__ai float64_t vcvtd_f64_s64(int64_t __p0) {
45213  float64_t __ret;
45214  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45215  return __ret;
45216}
45217#else
45218__ai float64_t vcvtd_f64_s64(int64_t __p0) {
45219  float64_t __ret;
45220  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45221  return __ret;
45222}
45223#endif
45224
45225#ifdef __LITTLE_ENDIAN__
45226__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45227  float64_t __ret;
45228  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45229  return __ret;
45230}
45231#else
45232__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45233  float64_t __ret;
45234  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45235  return __ret;
45236}
45237#endif
45238
45239#ifdef __LITTLE_ENDIAN__
45240__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45241  float64x2_t __ret;
45242  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
45243  return __ret;
45244}
45245#else
45246__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45247  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45248  float64x2_t __ret;
45249  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
45250  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45251  return __ret;
45252}
45253#endif
45254
45255#ifdef __LITTLE_ENDIAN__
45256__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45257  float64x2_t __ret;
45258  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
45259  return __ret;
45260}
45261#else
45262__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45263  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45264  float64x2_t __ret;
45265  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
45266  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45267  return __ret;
45268}
45269#endif
45270
45271#ifdef __LITTLE_ENDIAN__
45272__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45273  float64x1_t __ret;
45274  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45275  return __ret;
45276}
45277#else
45278__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45279  float64x1_t __ret;
45280  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45281  return __ret;
45282}
45283#endif
45284
45285#ifdef __LITTLE_ENDIAN__
45286__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45287  float64x1_t __ret;
45288  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45289  return __ret;
45290}
45291#else
45292__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45293  float64x1_t __ret;
45294  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45295  return __ret;
45296}
45297#endif
45298
45299#ifdef __LITTLE_ENDIAN__
45300__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45301  float64x2_t __ret;
45302  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45303  return __ret;
45304}
45305#else
45306__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45307  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45308  float64x2_t __ret;
45309  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
45310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45311  return __ret;
45312}
45313__ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
45314  float64x2_t __ret;
45315  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45316  return __ret;
45317}
45318#endif
45319
45320#ifdef __LITTLE_ENDIAN__
45321__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45322  float16x8_t __ret;
45323  __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
45324  return __ret;
45325}
45326#else
45327__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45328  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45329  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45330  float16x8_t __ret;
45331  __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
45332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45333  return __ret;
45334}
45335#endif
45336
45337#ifdef __LITTLE_ENDIAN__
45338__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45339  float32x4_t __ret;
45340  __ret = vcvt_f32_f16(vget_high_f16(__p0));
45341  return __ret;
45342}
45343#else
45344__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45345  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45346  float32x4_t __ret;
45347  __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
45348  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45349  return __ret;
45350}
45351#endif
45352
45353#ifdef __LITTLE_ENDIAN__
45354__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45355  float32x4_t __ret;
45356  __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
45357  return __ret;
45358}
45359#else
45360__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45361  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45362  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45363  float32x4_t __ret;
45364  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
45365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45366  return __ret;
45367}
45368#endif
45369
45370#ifdef __LITTLE_ENDIAN__
45371__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45372  float64x2_t __ret;
45373  __ret = vcvt_f64_f32(vget_high_f32(__p0));
45374  return __ret;
45375}
45376#else
45377__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45378  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45379  float64x2_t __ret;
45380  __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
45381  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45382  return __ret;
45383}
45384#endif
45385
45386#ifdef __LITTLE_ENDIAN__
45387#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45388  uint32_t __s0 = __p0; \
45389  float32_t __ret; \
45390  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45391  __ret; \
45392})
45393#else
45394#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45395  uint32_t __s0 = __p0; \
45396  float32_t __ret; \
45397  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45398  __ret; \
45399})
45400#endif
45401
45402#ifdef __LITTLE_ENDIAN__
45403#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45404  int32_t __s0 = __p0; \
45405  float32_t __ret; \
45406  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45407  __ret; \
45408})
45409#else
45410#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45411  int32_t __s0 = __p0; \
45412  float32_t __ret; \
45413  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45414  __ret; \
45415})
45416#endif
45417
45418#ifdef __LITTLE_ENDIAN__
45419#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45420  uint64x2_t __s0 = __p0; \
45421  float64x2_t __ret; \
45422  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
45423  __ret; \
45424})
45425#else
45426#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45427  uint64x2_t __s0 = __p0; \
45428  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45429  float64x2_t __ret; \
45430  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
45431  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45432  __ret; \
45433})
45434#endif
45435
45436#ifdef __LITTLE_ENDIAN__
45437#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45438  int64x2_t __s0 = __p0; \
45439  float64x2_t __ret; \
45440  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
45441  __ret; \
45442})
45443#else
45444#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45445  int64x2_t __s0 = __p0; \
45446  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45447  float64x2_t __ret; \
45448  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
45449  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45450  __ret; \
45451})
45452#endif
45453
45454#ifdef __LITTLE_ENDIAN__
45455#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45456  uint64x1_t __s0 = __p0; \
45457  float64x1_t __ret; \
45458  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45459  __ret; \
45460})
45461#else
45462#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45463  uint64x1_t __s0 = __p0; \
45464  float64x1_t __ret; \
45465  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45466  __ret; \
45467})
45468#endif
45469
45470#ifdef __LITTLE_ENDIAN__
45471#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45472  int64x1_t __s0 = __p0; \
45473  float64x1_t __ret; \
45474  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45475  __ret; \
45476})
45477#else
45478#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45479  int64x1_t __s0 = __p0; \
45480  float64x1_t __ret; \
45481  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45482  __ret; \
45483})
45484#endif
45485
45486#ifdef __LITTLE_ENDIAN__
45487#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45488  uint64_t __s0 = __p0; \
45489  float64_t __ret; \
45490  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45491  __ret; \
45492})
45493#else
45494#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45495  uint64_t __s0 = __p0; \
45496  float64_t __ret; \
45497  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45498  __ret; \
45499})
45500#endif
45501
45502#ifdef __LITTLE_ENDIAN__
45503#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45504  int64_t __s0 = __p0; \
45505  float64_t __ret; \
45506  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45507  __ret; \
45508})
45509#else
45510#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45511  int64_t __s0 = __p0; \
45512  float64_t __ret; \
45513  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45514  __ret; \
45515})
45516#endif
45517
45518#ifdef __LITTLE_ENDIAN__
45519#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45520  float32_t __s0 = __p0; \
45521  int32_t __ret; \
45522  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45523  __ret; \
45524})
45525#else
45526#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45527  float32_t __s0 = __p0; \
45528  int32_t __ret; \
45529  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45530  __ret; \
45531})
45532#endif
45533
45534#ifdef __LITTLE_ENDIAN__
45535#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45536  float64x2_t __s0 = __p0; \
45537  int64x2_t __ret; \
45538  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
45539  __ret; \
45540})
45541#else
45542#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45543  float64x2_t __s0 = __p0; \
45544  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45545  int64x2_t __ret; \
45546  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
45547  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45548  __ret; \
45549})
45550#endif
45551
45552#ifdef __LITTLE_ENDIAN__
45553#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45554  float64x1_t __s0 = __p0; \
45555  int64x1_t __ret; \
45556  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45557  __ret; \
45558})
45559#else
45560#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45561  float64x1_t __s0 = __p0; \
45562  int64x1_t __ret; \
45563  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45564  __ret; \
45565})
45566#endif
45567
45568#ifdef __LITTLE_ENDIAN__
45569#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45570  float64_t __s0 = __p0; \
45571  int64_t __ret; \
45572  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45573  __ret; \
45574})
45575#else
45576#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45577  float64_t __s0 = __p0; \
45578  int64_t __ret; \
45579  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45580  __ret; \
45581})
45582#endif
45583
45584#ifdef __LITTLE_ENDIAN__
45585#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45586  float32_t __s0 = __p0; \
45587  uint32_t __ret; \
45588  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45589  __ret; \
45590})
45591#else
45592#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45593  float32_t __s0 = __p0; \
45594  uint32_t __ret; \
45595  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45596  __ret; \
45597})
45598#endif
45599
45600#ifdef __LITTLE_ENDIAN__
45601#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45602  float64x2_t __s0 = __p0; \
45603  uint64x2_t __ret; \
45604  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
45605  __ret; \
45606})
45607#else
45608#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45609  float64x2_t __s0 = __p0; \
45610  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45611  uint64x2_t __ret; \
45612  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
45613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45614  __ret; \
45615})
45616#endif
45617
45618#ifdef __LITTLE_ENDIAN__
45619#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45620  float64x1_t __s0 = __p0; \
45621  uint64x1_t __ret; \
45622  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45623  __ret; \
45624})
45625#else
45626#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45627  float64x1_t __s0 = __p0; \
45628  uint64x1_t __ret; \
45629  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45630  __ret; \
45631})
45632#endif
45633
45634#ifdef __LITTLE_ENDIAN__
45635#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45636  float64_t __s0 = __p0; \
45637  uint64_t __ret; \
45638  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45639  __ret; \
45640})
45641#else
45642#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45643  float64_t __s0 = __p0; \
45644  uint64_t __ret; \
45645  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45646  __ret; \
45647})
45648#endif
45649
45650#ifdef __LITTLE_ENDIAN__
45651__ai int32_t vcvts_s32_f32(float32_t __p0) {
45652  int32_t __ret;
45653  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45654  return __ret;
45655}
45656#else
45657__ai int32_t vcvts_s32_f32(float32_t __p0) {
45658  int32_t __ret;
45659  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45660  return __ret;
45661}
45662#endif
45663
45664#ifdef __LITTLE_ENDIAN__
45665__ai int64_t vcvtd_s64_f64(float64_t __p0) {
45666  int64_t __ret;
45667  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45668  return __ret;
45669}
45670#else
45671__ai int64_t vcvtd_s64_f64(float64_t __p0) {
45672  int64_t __ret;
45673  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45674  return __ret;
45675}
45676#endif
45677
45678#ifdef __LITTLE_ENDIAN__
45679__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45680  int64x2_t __ret;
45681  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
45682  return __ret;
45683}
45684#else
45685__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45686  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45687  int64x2_t __ret;
45688  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
45689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45690  return __ret;
45691}
45692#endif
45693
45694#ifdef __LITTLE_ENDIAN__
45695__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45696  int64x1_t __ret;
45697  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45698  return __ret;
45699}
45700#else
45701__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45702  int64x1_t __ret;
45703  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45704  return __ret;
45705}
45706#endif
45707
45708#ifdef __LITTLE_ENDIAN__
45709__ai uint32_t vcvts_u32_f32(float32_t __p0) {
45710  uint32_t __ret;
45711  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45712  return __ret;
45713}
45714#else
45715__ai uint32_t vcvts_u32_f32(float32_t __p0) {
45716  uint32_t __ret;
45717  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45718  return __ret;
45719}
45720#endif
45721
45722#ifdef __LITTLE_ENDIAN__
45723__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45724  uint64_t __ret;
45725  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45726  return __ret;
45727}
45728#else
45729__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45730  uint64_t __ret;
45731  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45732  return __ret;
45733}
45734#endif
45735
45736#ifdef __LITTLE_ENDIAN__
45737__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45738  uint64x2_t __ret;
45739  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
45740  return __ret;
45741}
45742#else
45743__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45744  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45745  uint64x2_t __ret;
45746  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
45747  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45748  return __ret;
45749}
45750#endif
45751
45752#ifdef __LITTLE_ENDIAN__
45753__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45754  uint64x1_t __ret;
45755  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45756  return __ret;
45757}
45758#else
45759__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45760  uint64x1_t __ret;
45761  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45762  return __ret;
45763}
45764#endif
45765
45766#ifdef __LITTLE_ENDIAN__
45767__ai int32_t vcvtas_s32_f32(float32_t __p0) {
45768  int32_t __ret;
45769  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45770  return __ret;
45771}
45772#else
45773__ai int32_t vcvtas_s32_f32(float32_t __p0) {
45774  int32_t __ret;
45775  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45776  return __ret;
45777}
45778#endif
45779
45780#ifdef __LITTLE_ENDIAN__
45781__ai int64_t vcvtad_s64_f64(float64_t __p0) {
45782  int64_t __ret;
45783  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45784  return __ret;
45785}
45786#else
45787__ai int64_t vcvtad_s64_f64(float64_t __p0) {
45788  int64_t __ret;
45789  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45790  return __ret;
45791}
45792#endif
45793
45794#ifdef __LITTLE_ENDIAN__
45795__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45796  uint32_t __ret;
45797  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45798  return __ret;
45799}
45800#else
45801__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45802  uint32_t __ret;
45803  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45804  return __ret;
45805}
45806#endif
45807
45808#ifdef __LITTLE_ENDIAN__
45809__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45810  uint64_t __ret;
45811  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45812  return __ret;
45813}
45814#else
45815__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45816  uint64_t __ret;
45817  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45818  return __ret;
45819}
45820#endif
45821
45822#ifdef __LITTLE_ENDIAN__
45823__ai int32_t vcvtms_s32_f32(float32_t __p0) {
45824  int32_t __ret;
45825  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45826  return __ret;
45827}
45828#else
45829__ai int32_t vcvtms_s32_f32(float32_t __p0) {
45830  int32_t __ret;
45831  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45832  return __ret;
45833}
45834#endif
45835
45836#ifdef __LITTLE_ENDIAN__
45837__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45838  int64_t __ret;
45839  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45840  return __ret;
45841}
45842#else
45843__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45844  int64_t __ret;
45845  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45846  return __ret;
45847}
45848#endif
45849
45850#ifdef __LITTLE_ENDIAN__
45851__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45852  uint32_t __ret;
45853  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45854  return __ret;
45855}
45856#else
45857__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45858  uint32_t __ret;
45859  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45860  return __ret;
45861}
45862#endif
45863
45864#ifdef __LITTLE_ENDIAN__
45865__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45866  uint64_t __ret;
45867  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45868  return __ret;
45869}
45870#else
45871__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45872  uint64_t __ret;
45873  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45874  return __ret;
45875}
45876#endif
45877
45878#ifdef __LITTLE_ENDIAN__
45879__ai int32_t vcvtns_s32_f32(float32_t __p0) {
45880  int32_t __ret;
45881  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45882  return __ret;
45883}
45884#else
45885__ai int32_t vcvtns_s32_f32(float32_t __p0) {
45886  int32_t __ret;
45887  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45888  return __ret;
45889}
45890#endif
45891
45892#ifdef __LITTLE_ENDIAN__
45893__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45894  int64_t __ret;
45895  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45896  return __ret;
45897}
45898#else
45899__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45900  int64_t __ret;
45901  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45902  return __ret;
45903}
45904#endif
45905
45906#ifdef __LITTLE_ENDIAN__
45907__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45908  uint32_t __ret;
45909  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45910  return __ret;
45911}
45912#else
45913__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45914  uint32_t __ret;
45915  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45916  return __ret;
45917}
45918#endif
45919
45920#ifdef __LITTLE_ENDIAN__
45921__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45922  uint64_t __ret;
45923  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45924  return __ret;
45925}
45926#else
45927__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45928  uint64_t __ret;
45929  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45930  return __ret;
45931}
45932#endif
45933
45934#ifdef __LITTLE_ENDIAN__
45935__ai int32_t vcvtps_s32_f32(float32_t __p0) {
45936  int32_t __ret;
45937  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45938  return __ret;
45939}
45940#else
45941__ai int32_t vcvtps_s32_f32(float32_t __p0) {
45942  int32_t __ret;
45943  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45944  return __ret;
45945}
45946#endif
45947
45948#ifdef __LITTLE_ENDIAN__
45949__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45950  int64_t __ret;
45951  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45952  return __ret;
45953}
45954#else
45955__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45956  int64_t __ret;
45957  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45958  return __ret;
45959}
45960#endif
45961
45962#ifdef __LITTLE_ENDIAN__
45963__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45964  uint32_t __ret;
45965  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45966  return __ret;
45967}
45968#else
45969__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45970  uint32_t __ret;
45971  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45972  return __ret;
45973}
45974#endif
45975
45976#ifdef __LITTLE_ENDIAN__
45977__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45978  uint64_t __ret;
45979  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45980  return __ret;
45981}
45982#else
45983__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45984  uint64_t __ret;
45985  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45986  return __ret;
45987}
45988#endif
45989
45990#ifdef __LITTLE_ENDIAN__
45991__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45992  float32_t __ret;
45993  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45994  return __ret;
45995}
45996#else
45997__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45998  float32_t __ret;
45999  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
46000  return __ret;
46001}
46002#endif
46003
46004#ifdef __LITTLE_ENDIAN__
46005__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
46006  float32x2_t __ret;
46007  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
46008  return __ret;
46009}
46010#else
46011__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
46012  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46013  float32x2_t __ret;
46014  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
46015  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46016  return __ret;
46017}
46018__ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
46019  float32x2_t __ret;
46020  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
46021  return __ret;
46022}
46023#endif
46024
46025#ifdef __LITTLE_ENDIAN__
46026__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
46027  float32x4_t __ret;
46028  __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
46029  return __ret;
46030}
46031#else
46032__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
46033  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46034  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46035  float32x4_t __ret;
46036  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
46037  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46038  return __ret;
46039}
46040#endif
46041
46042#ifdef __LITTLE_ENDIAN__
46043__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46044  float64x2_t __ret;
46045  __ret = __p0 / __p1;
46046  return __ret;
46047}
46048#else
46049__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46050  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46051  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46052  float64x2_t __ret;
46053  __ret = __rev0 / __rev1;
46054  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46055  return __ret;
46056}
46057#endif
46058
46059#ifdef __LITTLE_ENDIAN__
46060__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46061  float32x4_t __ret;
46062  __ret = __p0 / __p1;
46063  return __ret;
46064}
46065#else
46066__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46067  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46068  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46069  float32x4_t __ret;
46070  __ret = __rev0 / __rev1;
46071  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46072  return __ret;
46073}
46074#endif
46075
46076#ifdef __LITTLE_ENDIAN__
46077__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46078  float64x1_t __ret;
46079  __ret = __p0 / __p1;
46080  return __ret;
46081}
46082#else
46083__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46084  float64x1_t __ret;
46085  __ret = __p0 / __p1;
46086  return __ret;
46087}
46088#endif
46089
46090#ifdef __LITTLE_ENDIAN__
46091__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46092  float32x2_t __ret;
46093  __ret = __p0 / __p1;
46094  return __ret;
46095}
46096#else
46097__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46098  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46099  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46100  float32x2_t __ret;
46101  __ret = __rev0 / __rev1;
46102  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46103  return __ret;
46104}
46105#endif
46106
46107#ifdef __LITTLE_ENDIAN__
46108#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46109  poly8x8_t __s0 = __p0; \
46110  poly8_t __ret; \
46111  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46112  __ret; \
46113})
46114#else
46115#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46116  poly8x8_t __s0 = __p0; \
46117  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46118  poly8_t __ret; \
46119  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46120  __ret; \
46121})
46122#endif
46123
46124#ifdef __LITTLE_ENDIAN__
46125#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46126  poly16x4_t __s0 = __p0; \
46127  poly16_t __ret; \
46128  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46129  __ret; \
46130})
46131#else
46132#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46133  poly16x4_t __s0 = __p0; \
46134  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46135  poly16_t __ret; \
46136  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46137  __ret; \
46138})
46139#endif
46140
46141#ifdef __LITTLE_ENDIAN__
46142#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46143  uint8x8_t __s0 = __p0; \
46144  uint8_t __ret; \
46145  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46146  __ret; \
46147})
46148#else
46149#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46150  uint8x8_t __s0 = __p0; \
46151  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46152  uint8_t __ret; \
46153  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46154  __ret; \
46155})
46156#endif
46157
46158#ifdef __LITTLE_ENDIAN__
46159#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46160  uint32x2_t __s0 = __p0; \
46161  uint32_t __ret; \
46162  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46163  __ret; \
46164})
46165#else
46166#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46167  uint32x2_t __s0 = __p0; \
46168  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46169  uint32_t __ret; \
46170  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46171  __ret; \
46172})
46173#endif
46174
46175#ifdef __LITTLE_ENDIAN__
46176#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46177  uint64x1_t __s0 = __p0; \
46178  uint64_t __ret; \
46179  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46180  __ret; \
46181})
46182#else
46183#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46184  uint64x1_t __s0 = __p0; \
46185  uint64_t __ret; \
46186  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46187  __ret; \
46188})
46189#endif
46190
46191#ifdef __LITTLE_ENDIAN__
46192#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46193  uint16x4_t __s0 = __p0; \
46194  uint16_t __ret; \
46195  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46196  __ret; \
46197})
46198#else
46199#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46200  uint16x4_t __s0 = __p0; \
46201  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46202  uint16_t __ret; \
46203  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46204  __ret; \
46205})
46206#endif
46207
46208#ifdef __LITTLE_ENDIAN__
46209#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46210  int8x8_t __s0 = __p0; \
46211  int8_t __ret; \
46212  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46213  __ret; \
46214})
46215#else
46216#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46217  int8x8_t __s0 = __p0; \
46218  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46219  int8_t __ret; \
46220  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46221  __ret; \
46222})
46223#endif
46224
46225#ifdef __LITTLE_ENDIAN__
46226#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46227  float64x1_t __s0 = __p0; \
46228  float64_t __ret; \
46229  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46230  __ret; \
46231})
46232#else
46233#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46234  float64x1_t __s0 = __p0; \
46235  float64_t __ret; \
46236  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46237  __ret; \
46238})
46239#endif
46240
46241#ifdef __LITTLE_ENDIAN__
46242#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46243  float32x2_t __s0 = __p0; \
46244  float32_t __ret; \
46245  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__s0, __p1); \
46246  __ret; \
46247})
46248#else
46249#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46250  float32x2_t __s0 = __p0; \
46251  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46252  float32_t __ret; \
46253  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__rev0, __p1); \
46254  __ret; \
46255})
46256#endif
46257
46258#ifdef __LITTLE_ENDIAN__
46259#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46260  int32x2_t __s0 = __p0; \
46261  int32_t __ret; \
46262  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46263  __ret; \
46264})
46265#else
46266#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46267  int32x2_t __s0 = __p0; \
46268  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46269  int32_t __ret; \
46270  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46271  __ret; \
46272})
46273#endif
46274
46275#ifdef __LITTLE_ENDIAN__
46276#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46277  int64x1_t __s0 = __p0; \
46278  int64_t __ret; \
46279  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46280  __ret; \
46281})
46282#else
46283#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46284  int64x1_t __s0 = __p0; \
46285  int64_t __ret; \
46286  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46287  __ret; \
46288})
46289#endif
46290
46291#ifdef __LITTLE_ENDIAN__
46292#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46293  int16x4_t __s0 = __p0; \
46294  int16_t __ret; \
46295  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46296  __ret; \
46297})
46298#else
46299#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46300  int16x4_t __s0 = __p0; \
46301  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46302  int16_t __ret; \
46303  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46304  __ret; \
46305})
46306#endif
46307
46308#ifdef __LITTLE_ENDIAN__
46309#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46310  poly64x1_t __s0 = __p0; \
46311  poly64x1_t __ret; \
46312  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46313  __ret; \
46314})
46315#else
46316#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46317  poly64x1_t __s0 = __p0; \
46318  poly64x1_t __ret; \
46319  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46320  __ret; \
46321})
46322#endif
46323
46324#ifdef __LITTLE_ENDIAN__
46325#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46326  poly64x1_t __s0 = __p0; \
46327  poly64x2_t __ret; \
46328  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46329  __ret; \
46330})
46331#else
46332#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46333  poly64x1_t __s0 = __p0; \
46334  poly64x2_t __ret; \
46335  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46337  __ret; \
46338})
46339#endif
46340
46341#ifdef __LITTLE_ENDIAN__
46342#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46343  float64x1_t __s0 = __p0; \
46344  float64x2_t __ret; \
46345  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46346  __ret; \
46347})
46348#else
46349#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46350  float64x1_t __s0 = __p0; \
46351  float64x2_t __ret; \
46352  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46353  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46354  __ret; \
46355})
46356#endif
46357
46358#ifdef __LITTLE_ENDIAN__
46359#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46360  float16x4_t __s0 = __p0; \
46361  float16x8_t __ret; \
46362  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46363  __ret; \
46364})
46365#else
46366#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46367  float16x4_t __s0 = __p0; \
46368  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46369  float16x8_t __ret; \
46370  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46371  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46372  __ret; \
46373})
46374#endif
46375
46376#ifdef __LITTLE_ENDIAN__
46377#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46378  float64x1_t __s0 = __p0; \
46379  float64x1_t __ret; \
46380  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46381  __ret; \
46382})
46383#else
46384#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46385  float64x1_t __s0 = __p0; \
46386  float64x1_t __ret; \
46387  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46388  __ret; \
46389})
46390#endif
46391
46392#ifdef __LITTLE_ENDIAN__
46393#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46394  float16x4_t __s0 = __p0; \
46395  float16x4_t __ret; \
46396  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46397  __ret; \
46398})
46399#else
46400#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46401  float16x4_t __s0 = __p0; \
46402  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46403  float16x4_t __ret; \
46404  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46405  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46406  __ret; \
46407})
46408#endif
46409
46410#ifdef __LITTLE_ENDIAN__
46411#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46412  poly8x16_t __s0 = __p0; \
46413  poly8_t __ret; \
46414  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46415  __ret; \
46416})
46417#else
46418#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46419  poly8x16_t __s0 = __p0; \
46420  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46421  poly8_t __ret; \
46422  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46423  __ret; \
46424})
46425#endif
46426
46427#ifdef __LITTLE_ENDIAN__
46428#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46429  poly16x8_t __s0 = __p0; \
46430  poly16_t __ret; \
46431  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46432  __ret; \
46433})
46434#else
46435#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46436  poly16x8_t __s0 = __p0; \
46437  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46438  poly16_t __ret; \
46439  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46440  __ret; \
46441})
46442#endif
46443
46444#ifdef __LITTLE_ENDIAN__
46445#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46446  uint8x16_t __s0 = __p0; \
46447  uint8_t __ret; \
46448  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46449  __ret; \
46450})
46451#else
46452#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46453  uint8x16_t __s0 = __p0; \
46454  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46455  uint8_t __ret; \
46456  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46457  __ret; \
46458})
46459#endif
46460
46461#ifdef __LITTLE_ENDIAN__
46462#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46463  uint32x4_t __s0 = __p0; \
46464  uint32_t __ret; \
46465  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46466  __ret; \
46467})
46468#else
46469#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46470  uint32x4_t __s0 = __p0; \
46471  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46472  uint32_t __ret; \
46473  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46474  __ret; \
46475})
46476#endif
46477
46478#ifdef __LITTLE_ENDIAN__
46479#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46480  uint64x2_t __s0 = __p0; \
46481  uint64_t __ret; \
46482  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46483  __ret; \
46484})
46485#else
46486#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46487  uint64x2_t __s0 = __p0; \
46488  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46489  uint64_t __ret; \
46490  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46491  __ret; \
46492})
46493#endif
46494
46495#ifdef __LITTLE_ENDIAN__
46496#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46497  uint16x8_t __s0 = __p0; \
46498  uint16_t __ret; \
46499  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46500  __ret; \
46501})
46502#else
46503#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46504  uint16x8_t __s0 = __p0; \
46505  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46506  uint16_t __ret; \
46507  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46508  __ret; \
46509})
46510#endif
46511
46512#ifdef __LITTLE_ENDIAN__
46513#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46514  int8x16_t __s0 = __p0; \
46515  int8_t __ret; \
46516  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46517  __ret; \
46518})
46519#else
46520#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46521  int8x16_t __s0 = __p0; \
46522  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46523  int8_t __ret; \
46524  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46525  __ret; \
46526})
46527#endif
46528
46529#ifdef __LITTLE_ENDIAN__
46530#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46531  float64x2_t __s0 = __p0; \
46532  float64_t __ret; \
46533  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__s0, __p1); \
46534  __ret; \
46535})
46536#else
46537#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46538  float64x2_t __s0 = __p0; \
46539  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46540  float64_t __ret; \
46541  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__rev0, __p1); \
46542  __ret; \
46543})
46544#endif
46545
46546#ifdef __LITTLE_ENDIAN__
46547#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46548  float32x4_t __s0 = __p0; \
46549  float32_t __ret; \
46550  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__s0, __p1); \
46551  __ret; \
46552})
46553#else
46554#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46555  float32x4_t __s0 = __p0; \
46556  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46557  float32_t __ret; \
46558  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__rev0, __p1); \
46559  __ret; \
46560})
46561#endif
46562
46563#ifdef __LITTLE_ENDIAN__
46564#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46565  int32x4_t __s0 = __p0; \
46566  int32_t __ret; \
46567  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46568  __ret; \
46569})
46570#else
46571#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46572  int32x4_t __s0 = __p0; \
46573  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46574  int32_t __ret; \
46575  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46576  __ret; \
46577})
46578#endif
46579
46580#ifdef __LITTLE_ENDIAN__
46581#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46582  int64x2_t __s0 = __p0; \
46583  int64_t __ret; \
46584  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46585  __ret; \
46586})
46587#else
46588#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46589  int64x2_t __s0 = __p0; \
46590  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46591  int64_t __ret; \
46592  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46593  __ret; \
46594})
46595#endif
46596
46597#ifdef __LITTLE_ENDIAN__
46598#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46599  int16x8_t __s0 = __p0; \
46600  int16_t __ret; \
46601  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46602  __ret; \
46603})
46604#else
46605#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46606  int16x8_t __s0 = __p0; \
46607  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46608  int16_t __ret; \
46609  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46610  __ret; \
46611})
46612#endif
46613
46614#ifdef __LITTLE_ENDIAN__
46615#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46616  poly8x16_t __s0 = __p0; \
46617  poly8x8_t __ret; \
46618  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46619  __ret; \
46620})
46621#else
46622#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46623  poly8x16_t __s0 = __p0; \
46624  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46625  poly8x8_t __ret; \
46626  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46627  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46628  __ret; \
46629})
46630#endif
46631
46632#ifdef __LITTLE_ENDIAN__
46633#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46634  poly64x2_t __s0 = __p0; \
46635  poly64x1_t __ret; \
46636  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46637  __ret; \
46638})
46639#else
46640#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46641  poly64x2_t __s0 = __p0; \
46642  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46643  poly64x1_t __ret; \
46644  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46645  __ret; \
46646})
46647#endif
46648
46649#ifdef __LITTLE_ENDIAN__
46650#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46651  poly16x8_t __s0 = __p0; \
46652  poly16x4_t __ret; \
46653  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46654  __ret; \
46655})
46656#else
46657#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46658  poly16x8_t __s0 = __p0; \
46659  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46660  poly16x4_t __ret; \
46661  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46662  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46663  __ret; \
46664})
46665#endif
46666
46667#ifdef __LITTLE_ENDIAN__
46668#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46669  poly8x16_t __s0 = __p0; \
46670  poly8x16_t __ret; \
46671  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46672  __ret; \
46673})
46674#else
46675#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46676  poly8x16_t __s0 = __p0; \
46677  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46678  poly8x16_t __ret; \
46679  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46680  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46681  __ret; \
46682})
46683#endif
46684
46685#ifdef __LITTLE_ENDIAN__
46686#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46687  poly64x2_t __s0 = __p0; \
46688  poly64x2_t __ret; \
46689  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46690  __ret; \
46691})
46692#else
46693#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46694  poly64x2_t __s0 = __p0; \
46695  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46696  poly64x2_t __ret; \
46697  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46699  __ret; \
46700})
46701#endif
46702
46703#ifdef __LITTLE_ENDIAN__
46704#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46705  poly16x8_t __s0 = __p0; \
46706  poly16x8_t __ret; \
46707  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46708  __ret; \
46709})
46710#else
46711#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46712  poly16x8_t __s0 = __p0; \
46713  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46714  poly16x8_t __ret; \
46715  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46716  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46717  __ret; \
46718})
46719#endif
46720
46721#ifdef __LITTLE_ENDIAN__
46722#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46723  uint8x16_t __s0 = __p0; \
46724  uint8x16_t __ret; \
46725  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46726  __ret; \
46727})
46728#else
46729#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46730  uint8x16_t __s0 = __p0; \
46731  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46732  uint8x16_t __ret; \
46733  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46734  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46735  __ret; \
46736})
46737#endif
46738
46739#ifdef __LITTLE_ENDIAN__
46740#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46741  uint32x4_t __s0 = __p0; \
46742  uint32x4_t __ret; \
46743  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46744  __ret; \
46745})
46746#else
46747#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46748  uint32x4_t __s0 = __p0; \
46749  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46750  uint32x4_t __ret; \
46751  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46753  __ret; \
46754})
46755#endif
46756
46757#ifdef __LITTLE_ENDIAN__
46758#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46759  uint64x2_t __s0 = __p0; \
46760  uint64x2_t __ret; \
46761  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46762  __ret; \
46763})
46764#else
46765#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46766  uint64x2_t __s0 = __p0; \
46767  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46768  uint64x2_t __ret; \
46769  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46770  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46771  __ret; \
46772})
46773#endif
46774
46775#ifdef __LITTLE_ENDIAN__
46776#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46777  uint16x8_t __s0 = __p0; \
46778  uint16x8_t __ret; \
46779  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46780  __ret; \
46781})
46782#else
46783#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46784  uint16x8_t __s0 = __p0; \
46785  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46786  uint16x8_t __ret; \
46787  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46788  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46789  __ret; \
46790})
46791#endif
46792
46793#ifdef __LITTLE_ENDIAN__
46794#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46795  int8x16_t __s0 = __p0; \
46796  int8x16_t __ret; \
46797  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46798  __ret; \
46799})
46800#else
46801#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46802  int8x16_t __s0 = __p0; \
46803  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46804  int8x16_t __ret; \
46805  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46806  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46807  __ret; \
46808})
46809#endif
46810
46811#ifdef __LITTLE_ENDIAN__
46812#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46813  float64x2_t __s0 = __p0; \
46814  float64x2_t __ret; \
46815  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46816  __ret; \
46817})
46818#else
46819#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46820  float64x2_t __s0 = __p0; \
46821  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46822  float64x2_t __ret; \
46823  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46824  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46825  __ret; \
46826})
46827#endif
46828
46829#ifdef __LITTLE_ENDIAN__
46830#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46831  float32x4_t __s0 = __p0; \
46832  float32x4_t __ret; \
46833  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46834  __ret; \
46835})
46836#else
46837#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46838  float32x4_t __s0 = __p0; \
46839  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46840  float32x4_t __ret; \
46841  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46843  __ret; \
46844})
46845#endif
46846
46847#ifdef __LITTLE_ENDIAN__
46848#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46849  float16x8_t __s0 = __p0; \
46850  float16x8_t __ret; \
46851  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46852  __ret; \
46853})
46854#else
46855#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46856  float16x8_t __s0 = __p0; \
46857  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46858  float16x8_t __ret; \
46859  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46860  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46861  __ret; \
46862})
46863#endif
46864
46865#ifdef __LITTLE_ENDIAN__
46866#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46867  int32x4_t __s0 = __p0; \
46868  int32x4_t __ret; \
46869  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46870  __ret; \
46871})
46872#else
46873#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46874  int32x4_t __s0 = __p0; \
46875  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46876  int32x4_t __ret; \
46877  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46878  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46879  __ret; \
46880})
46881#endif
46882
46883#ifdef __LITTLE_ENDIAN__
46884#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46885  int64x2_t __s0 = __p0; \
46886  int64x2_t __ret; \
46887  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46888  __ret; \
46889})
46890#else
46891#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46892  int64x2_t __s0 = __p0; \
46893  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46894  int64x2_t __ret; \
46895  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46896  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46897  __ret; \
46898})
46899#endif
46900
46901#ifdef __LITTLE_ENDIAN__
46902#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46903  int16x8_t __s0 = __p0; \
46904  int16x8_t __ret; \
46905  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46906  __ret; \
46907})
46908#else
46909#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46910  int16x8_t __s0 = __p0; \
46911  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46912  int16x8_t __ret; \
46913  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46914  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46915  __ret; \
46916})
46917#endif
46918
46919#ifdef __LITTLE_ENDIAN__
46920#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46921  uint8x16_t __s0 = __p0; \
46922  uint8x8_t __ret; \
46923  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46924  __ret; \
46925})
46926#else
46927#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46928  uint8x16_t __s0 = __p0; \
46929  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46930  uint8x8_t __ret; \
46931  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46933  __ret; \
46934})
46935#endif
46936
46937#ifdef __LITTLE_ENDIAN__
46938#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46939  uint32x4_t __s0 = __p0; \
46940  uint32x2_t __ret; \
46941  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46942  __ret; \
46943})
46944#else
46945#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46946  uint32x4_t __s0 = __p0; \
46947  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46948  uint32x2_t __ret; \
46949  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46950  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46951  __ret; \
46952})
46953#endif
46954
46955#ifdef __LITTLE_ENDIAN__
46956#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46957  uint64x2_t __s0 = __p0; \
46958  uint64x1_t __ret; \
46959  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46960  __ret; \
46961})
46962#else
46963#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46964  uint64x2_t __s0 = __p0; \
46965  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46966  uint64x1_t __ret; \
46967  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46968  __ret; \
46969})
46970#endif
46971
46972#ifdef __LITTLE_ENDIAN__
46973#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46974  uint16x8_t __s0 = __p0; \
46975  uint16x4_t __ret; \
46976  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46977  __ret; \
46978})
46979#else
46980#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46981  uint16x8_t __s0 = __p0; \
46982  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46983  uint16x4_t __ret; \
46984  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46985  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46986  __ret; \
46987})
46988#endif
46989
46990#ifdef __LITTLE_ENDIAN__
46991#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46992  int8x16_t __s0 = __p0; \
46993  int8x8_t __ret; \
46994  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46995  __ret; \
46996})
46997#else
46998#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46999  int8x16_t __s0 = __p0; \
47000  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47001  int8x8_t __ret; \
47002  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
47003  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
47004  __ret; \
47005})
47006#endif
47007
47008#ifdef __LITTLE_ENDIAN__
47009#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
47010  float64x2_t __s0 = __p0; \
47011  float64x1_t __ret; \
47012  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
47013  __ret; \
47014})
47015#else
47016#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
47017  float64x2_t __s0 = __p0; \
47018  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47019  float64x1_t __ret; \
47020  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
47021  __ret; \
47022})
47023#endif
47024
47025#ifdef __LITTLE_ENDIAN__
47026#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
47027  float32x4_t __s0 = __p0; \
47028  float32x2_t __ret; \
47029  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
47030  __ret; \
47031})
47032#else
47033#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
47034  float32x4_t __s0 = __p0; \
47035  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47036  float32x2_t __ret; \
47037  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
47038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47039  __ret; \
47040})
47041#endif
47042
47043#ifdef __LITTLE_ENDIAN__
47044#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47045  float16x8_t __s0 = __p0; \
47046  float16x4_t __ret; \
47047  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47048  __ret; \
47049})
47050#else
47051#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47052  float16x8_t __s0 = __p0; \
47053  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47054  float16x4_t __ret; \
47055  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47056  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47057  __ret; \
47058})
47059#endif
47060
47061#ifdef __LITTLE_ENDIAN__
47062#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47063  int32x4_t __s0 = __p0; \
47064  int32x2_t __ret; \
47065  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
47066  __ret; \
47067})
47068#else
47069#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47070  int32x4_t __s0 = __p0; \
47071  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47072  int32x2_t __ret; \
47073  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
47074  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47075  __ret; \
47076})
47077#endif
47078
47079#ifdef __LITTLE_ENDIAN__
47080#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47081  int64x2_t __s0 = __p0; \
47082  int64x1_t __ret; \
47083  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
47084  __ret; \
47085})
47086#else
47087#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47088  int64x2_t __s0 = __p0; \
47089  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47090  int64x1_t __ret; \
47091  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
47092  __ret; \
47093})
47094#endif
47095
47096#ifdef __LITTLE_ENDIAN__
47097#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47098  int16x8_t __s0 = __p0; \
47099  int16x4_t __ret; \
47100  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47101  __ret; \
47102})
47103#else
47104#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47105  int16x8_t __s0 = __p0; \
47106  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47107  int16x4_t __ret; \
47108  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47109  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47110  __ret; \
47111})
47112#endif
47113
47114#ifdef __LITTLE_ENDIAN__
47115__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47116  poly64x1_t __ret;
47117  __ret = (poly64x1_t) {__p0};
47118  return __ret;
47119}
47120#else
47121__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47122  poly64x1_t __ret;
47123  __ret = (poly64x1_t) {__p0};
47124  return __ret;
47125}
47126#endif
47127
47128#ifdef __LITTLE_ENDIAN__
47129__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47130  poly64x2_t __ret;
47131  __ret = (poly64x2_t) {__p0, __p0};
47132  return __ret;
47133}
47134#else
47135__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47136  poly64x2_t __ret;
47137  __ret = (poly64x2_t) {__p0, __p0};
47138  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47139  return __ret;
47140}
47141#endif
47142
47143#ifdef __LITTLE_ENDIAN__
47144__ai float64x2_t vdupq_n_f64(float64_t __p0) {
47145  float64x2_t __ret;
47146  __ret = (float64x2_t) {__p0, __p0};
47147  return __ret;
47148}
47149#else
47150__ai float64x2_t vdupq_n_f64(float64_t __p0) {
47151  float64x2_t __ret;
47152  __ret = (float64x2_t) {__p0, __p0};
47153  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47154  return __ret;
47155}
47156#endif
47157
47158#ifdef __LITTLE_ENDIAN__
47159__ai float64x1_t vdup_n_f64(float64_t __p0) {
47160  float64x1_t __ret;
47161  __ret = (float64x1_t) {__p0};
47162  return __ret;
47163}
47164#else
47165__ai float64x1_t vdup_n_f64(float64_t __p0) {
47166  float64x1_t __ret;
47167  __ret = (float64x1_t) {__p0};
47168  return __ret;
47169}
47170#endif
47171
47172#ifdef __LITTLE_ENDIAN__
47173#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47174  poly64x1_t __s0 = __p0; \
47175  poly64x1_t __s1 = __p1; \
47176  poly64x1_t __ret; \
47177  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47178  __ret; \
47179})
47180#else
47181#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47182  poly64x1_t __s0 = __p0; \
47183  poly64x1_t __s1 = __p1; \
47184  poly64x1_t __ret; \
47185  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47186  __ret; \
47187})
47188#endif
47189
47190#ifdef __LITTLE_ENDIAN__
47191#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47192  poly64x2_t __s0 = __p0; \
47193  poly64x2_t __s1 = __p1; \
47194  poly64x2_t __ret; \
47195  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
47196  __ret; \
47197})
47198#else
47199#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47200  poly64x2_t __s0 = __p0; \
47201  poly64x2_t __s1 = __p1; \
47202  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47203  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47204  poly64x2_t __ret; \
47205  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
47206  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47207  __ret; \
47208})
47209#endif
47210
47211#ifdef __LITTLE_ENDIAN__
47212#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47213  float64x2_t __s0 = __p0; \
47214  float64x2_t __s1 = __p1; \
47215  float64x2_t __ret; \
47216  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
47217  __ret; \
47218})
47219#else
47220#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47221  float64x2_t __s0 = __p0; \
47222  float64x2_t __s1 = __p1; \
47223  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47224  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47225  float64x2_t __ret; \
47226  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
47227  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47228  __ret; \
47229})
47230#endif
47231
47232#ifdef __LITTLE_ENDIAN__
47233#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47234  float64x1_t __s0 = __p0; \
47235  float64x1_t __s1 = __p1; \
47236  float64x1_t __ret; \
47237  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47238  __ret; \
47239})
47240#else
47241#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47242  float64x1_t __s0 = __p0; \
47243  float64x1_t __s1 = __p1; \
47244  float64x1_t __ret; \
47245  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47246  __ret; \
47247})
47248#endif
47249
47250#ifdef __LITTLE_ENDIAN__
47251__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47252  float64x2_t __ret;
47253  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47254  return __ret;
47255}
47256#else
47257__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47258  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47259  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47260  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47261  float64x2_t __ret;
47262  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
47263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47264  return __ret;
47265}
47266__ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47267  float64x2_t __ret;
47268  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47269  return __ret;
47270}
47271#endif
47272
47273#ifdef __LITTLE_ENDIAN__
47274__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47275  float64x1_t __ret;
47276  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47277  return __ret;
47278}
47279#else
47280__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47281  float64x1_t __ret;
47282  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47283  return __ret;
47284}
47285__ai float64x1_t __noswap_vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47286  float64x1_t __ret;
47287  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47288  return __ret;
47289}
47290#endif
47291
47292#ifdef __LITTLE_ENDIAN__
47293#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47294  float64_t __s0 = __p0; \
47295  float64_t __s1 = __p1; \
47296  float64x1_t __s2 = __p2; \
47297  float64_t __ret; \
47298  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47299  __ret; \
47300})
47301#else
47302#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47303  float64_t __s0 = __p0; \
47304  float64_t __s1 = __p1; \
47305  float64x1_t __s2 = __p2; \
47306  float64_t __ret; \
47307  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47308  __ret; \
47309})
47310#define __noswap_vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47311  float64_t __s0 = __p0; \
47312  float64_t __s1 = __p1; \
47313  float64x1_t __s2 = __p2; \
47314  float64_t __ret; \
47315  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47316  __ret; \
47317})
47318#endif
47319
47320#ifdef __LITTLE_ENDIAN__
47321#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47322  float32_t __s0 = __p0; \
47323  float32_t __s1 = __p1; \
47324  float32x2_t __s2 = __p2; \
47325  float32_t __ret; \
47326  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47327  __ret; \
47328})
47329#else
47330#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47331  float32_t __s0 = __p0; \
47332  float32_t __s1 = __p1; \
47333  float32x2_t __s2 = __p2; \
47334  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47335  float32_t __ret; \
47336  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__rev2, __p3); \
47337  __ret; \
47338})
47339#define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47340  float32_t __s0 = __p0; \
47341  float32_t __s1 = __p1; \
47342  float32x2_t __s2 = __p2; \
47343  float32_t __ret; \
47344  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47345  __ret; \
47346})
47347#endif
47348
47349#ifdef __LITTLE_ENDIAN__
47350#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47351  float64x2_t __s0 = __p0; \
47352  float64x2_t __s1 = __p1; \
47353  float64x1_t __s2 = __p2; \
47354  float64x2_t __ret; \
47355  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47356  __ret; \
47357})
47358#else
47359#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47360  float64x2_t __s0 = __p0; \
47361  float64x2_t __s1 = __p1; \
47362  float64x1_t __s2 = __p2; \
47363  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47364  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47365  float64x2_t __ret; \
47366  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
47367  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47368  __ret; \
47369})
47370#define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47371  float64x2_t __s0 = __p0; \
47372  float64x2_t __s1 = __p1; \
47373  float64x1_t __s2 = __p2; \
47374  float64x2_t __ret; \
47375  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47376  __ret; \
47377})
47378#endif
47379
47380#ifdef __LITTLE_ENDIAN__
47381#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47382  float32x4_t __s0 = __p0; \
47383  float32x4_t __s1 = __p1; \
47384  float32x2_t __s2 = __p2; \
47385  float32x4_t __ret; \
47386  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47387  __ret; \
47388})
47389#else
47390#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47391  float32x4_t __s0 = __p0; \
47392  float32x4_t __s1 = __p1; \
47393  float32x2_t __s2 = __p2; \
47394  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47395  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47396  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47397  float32x4_t __ret; \
47398  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
47399  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47400  __ret; \
47401})
47402#define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47403  float32x4_t __s0 = __p0; \
47404  float32x4_t __s1 = __p1; \
47405  float32x2_t __s2 = __p2; \
47406  float32x4_t __ret; \
47407  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47408  __ret; \
47409})
47410#endif
47411
47412#ifdef __LITTLE_ENDIAN__
47413#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47414  float64x1_t __s0 = __p0; \
47415  float64x1_t __s1 = __p1; \
47416  float64x1_t __s2 = __p2; \
47417  float64x1_t __ret; \
47418  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47419  __ret; \
47420})
47421#else
47422#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47423  float64x1_t __s0 = __p0; \
47424  float64x1_t __s1 = __p1; \
47425  float64x1_t __s2 = __p2; \
47426  float64x1_t __ret; \
47427  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47428  __ret; \
47429})
47430#define __noswap_vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47431  float64x1_t __s0 = __p0; \
47432  float64x1_t __s1 = __p1; \
47433  float64x1_t __s2 = __p2; \
47434  float64x1_t __ret; \
47435  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47436  __ret; \
47437})
47438#endif
47439
47440#ifdef __LITTLE_ENDIAN__
47441#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47442  float32x2_t __s0 = __p0; \
47443  float32x2_t __s1 = __p1; \
47444  float32x2_t __s2 = __p2; \
47445  float32x2_t __ret; \
47446  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47447  __ret; \
47448})
47449#else
47450#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47451  float32x2_t __s0 = __p0; \
47452  float32x2_t __s1 = __p1; \
47453  float32x2_t __s2 = __p2; \
47454  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47455  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47456  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47457  float32x2_t __ret; \
47458  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
47459  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47460  __ret; \
47461})
47462#define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47463  float32x2_t __s0 = __p0; \
47464  float32x2_t __s1 = __p1; \
47465  float32x2_t __s2 = __p2; \
47466  float32x2_t __ret; \
47467  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47468  __ret; \
47469})
47470#endif
47471
47472#ifdef __LITTLE_ENDIAN__
47473#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47474  float64_t __s0 = __p0; \
47475  float64_t __s1 = __p1; \
47476  float64x2_t __s2 = __p2; \
47477  float64_t __ret; \
47478  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47479  __ret; \
47480})
47481#else
47482#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47483  float64_t __s0 = __p0; \
47484  float64_t __s1 = __p1; \
47485  float64x2_t __s2 = __p2; \
47486  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47487  float64_t __ret; \
47488  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__rev2, __p3); \
47489  __ret; \
47490})
47491#define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47492  float64_t __s0 = __p0; \
47493  float64_t __s1 = __p1; \
47494  float64x2_t __s2 = __p2; \
47495  float64_t __ret; \
47496  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47497  __ret; \
47498})
47499#endif
47500
47501#ifdef __LITTLE_ENDIAN__
47502#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47503  float32_t __s0 = __p0; \
47504  float32_t __s1 = __p1; \
47505  float32x4_t __s2 = __p2; \
47506  float32_t __ret; \
47507  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47508  __ret; \
47509})
47510#else
47511#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47512  float32_t __s0 = __p0; \
47513  float32_t __s1 = __p1; \
47514  float32x4_t __s2 = __p2; \
47515  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47516  float32_t __ret; \
47517  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__rev2, __p3); \
47518  __ret; \
47519})
47520#define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47521  float32_t __s0 = __p0; \
47522  float32_t __s1 = __p1; \
47523  float32x4_t __s2 = __p2; \
47524  float32_t __ret; \
47525  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47526  __ret; \
47527})
47528#endif
47529
47530#ifdef __LITTLE_ENDIAN__
47531#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47532  float64x2_t __s0 = __p0; \
47533  float64x2_t __s1 = __p1; \
47534  float64x2_t __s2 = __p2; \
47535  float64x2_t __ret; \
47536  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47537  __ret; \
47538})
47539#else
47540#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47541  float64x2_t __s0 = __p0; \
47542  float64x2_t __s1 = __p1; \
47543  float64x2_t __s2 = __p2; \
47544  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47545  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47546  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47547  float64x2_t __ret; \
47548  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
47549  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47550  __ret; \
47551})
47552#define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47553  float64x2_t __s0 = __p0; \
47554  float64x2_t __s1 = __p1; \
47555  float64x2_t __s2 = __p2; \
47556  float64x2_t __ret; \
47557  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47558  __ret; \
47559})
47560#endif
47561
47562#ifdef __LITTLE_ENDIAN__
47563#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47564  float32x4_t __s0 = __p0; \
47565  float32x4_t __s1 = __p1; \
47566  float32x4_t __s2 = __p2; \
47567  float32x4_t __ret; \
47568  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47569  __ret; \
47570})
47571#else
47572#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47573  float32x4_t __s0 = __p0; \
47574  float32x4_t __s1 = __p1; \
47575  float32x4_t __s2 = __p2; \
47576  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47577  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47578  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47579  float32x4_t __ret; \
47580  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
47581  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47582  __ret; \
47583})
47584#define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47585  float32x4_t __s0 = __p0; \
47586  float32x4_t __s1 = __p1; \
47587  float32x4_t __s2 = __p2; \
47588  float32x4_t __ret; \
47589  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47590  __ret; \
47591})
47592#endif
47593
47594#ifdef __LITTLE_ENDIAN__
47595#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47596  float64x1_t __s0 = __p0; \
47597  float64x1_t __s1 = __p1; \
47598  float64x2_t __s2 = __p2; \
47599  float64x1_t __ret; \
47600  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47601  __ret; \
47602})
47603#else
47604#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47605  float64x1_t __s0 = __p0; \
47606  float64x1_t __s1 = __p1; \
47607  float64x2_t __s2 = __p2; \
47608  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47609  float64x1_t __ret; \
47610  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
47611  __ret; \
47612})
47613#define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47614  float64x1_t __s0 = __p0; \
47615  float64x1_t __s1 = __p1; \
47616  float64x2_t __s2 = __p2; \
47617  float64x1_t __ret; \
47618  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47619  __ret; \
47620})
47621#endif
47622
47623#ifdef __LITTLE_ENDIAN__
47624#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47625  float32x2_t __s0 = __p0; \
47626  float32x2_t __s1 = __p1; \
47627  float32x4_t __s2 = __p2; \
47628  float32x2_t __ret; \
47629  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47630  __ret; \
47631})
47632#else
47633#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47634  float32x2_t __s0 = __p0; \
47635  float32x2_t __s1 = __p1; \
47636  float32x4_t __s2 = __p2; \
47637  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47638  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47639  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47640  float32x2_t __ret; \
47641  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
47642  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47643  __ret; \
47644})
47645#define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47646  float32x2_t __s0 = __p0; \
47647  float32x2_t __s1 = __p1; \
47648  float32x4_t __s2 = __p2; \
47649  float32x2_t __ret; \
47650  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47651  __ret; \
47652})
47653#endif
47654
47655#ifdef __LITTLE_ENDIAN__
47656__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47657  float64x2_t __ret;
47658  __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
47659  return __ret;
47660}
47661#else
47662__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47663  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47664  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47665  float64x2_t __ret;
47666  __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
47667  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47668  return __ret;
47669}
47670#endif
47671
47672#ifdef __LITTLE_ENDIAN__
47673__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47674  float32x4_t __ret;
47675  __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
47676  return __ret;
47677}
47678#else
47679__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47680  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47681  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47682  float32x4_t __ret;
47683  __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
47684  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47685  return __ret;
47686}
47687#endif
47688
47689#ifdef __LITTLE_ENDIAN__
47690__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47691  float32x2_t __ret;
47692  __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
47693  return __ret;
47694}
47695#else
47696__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47697  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47698  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47699  float32x2_t __ret;
47700  __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
47701  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47702  return __ret;
47703}
47704#endif
47705
47706#ifdef __LITTLE_ENDIAN__
47707__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47708  float64x2_t __ret;
47709  __ret = vfmaq_f64(__p0, -__p1, __p2);
47710  return __ret;
47711}
47712#else
47713__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47714  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47715  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47716  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47717  float64x2_t __ret;
47718  __ret = __noswap_vfmaq_f64(__rev0, -__rev1, __rev2);
47719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47720  return __ret;
47721}
47722#endif
47723
47724#ifdef __LITTLE_ENDIAN__
47725__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47726  float64x1_t __ret;
47727  __ret = vfma_f64(__p0, -__p1, __p2);
47728  return __ret;
47729}
47730#else
47731__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47732  float64x1_t __ret;
47733  __ret = __noswap_vfma_f64(__p0, -__p1, __p2);
47734  return __ret;
47735}
47736#endif
47737
47738#ifdef __LITTLE_ENDIAN__
47739#define vfmsd_lane_f64(__p0_88, __p1_88, __p2_88, __p3_88) __extension__ ({ \
47740  float64_t __s0_88 = __p0_88; \
47741  float64_t __s1_88 = __p1_88; \
47742  float64x1_t __s2_88 = __p2_88; \
47743  float64_t __ret_88; \
47744  __ret_88 = vfmad_lane_f64(__s0_88, -__s1_88, __s2_88, __p3_88); \
47745  __ret_88; \
47746})
47747#else
47748#define vfmsd_lane_f64(__p0_89, __p1_89, __p2_89, __p3_89) __extension__ ({ \
47749  float64_t __s0_89 = __p0_89; \
47750  float64_t __s1_89 = __p1_89; \
47751  float64x1_t __s2_89 = __p2_89; \
47752  float64_t __ret_89; \
47753  __ret_89 = __noswap_vfmad_lane_f64(__s0_89, -__s1_89, __s2_89, __p3_89); \
47754  __ret_89; \
47755})
47756#endif
47757
47758#ifdef __LITTLE_ENDIAN__
47759#define vfmss_lane_f32(__p0_90, __p1_90, __p2_90, __p3_90) __extension__ ({ \
47760  float32_t __s0_90 = __p0_90; \
47761  float32_t __s1_90 = __p1_90; \
47762  float32x2_t __s2_90 = __p2_90; \
47763  float32_t __ret_90; \
47764  __ret_90 = vfmas_lane_f32(__s0_90, -__s1_90, __s2_90, __p3_90); \
47765  __ret_90; \
47766})
47767#else
47768#define vfmss_lane_f32(__p0_91, __p1_91, __p2_91, __p3_91) __extension__ ({ \
47769  float32_t __s0_91 = __p0_91; \
47770  float32_t __s1_91 = __p1_91; \
47771  float32x2_t __s2_91 = __p2_91; \
47772  float32x2_t __rev2_91;  __rev2_91 = __builtin_shufflevector(__s2_91, __s2_91, 1, 0); \
47773  float32_t __ret_91; \
47774  __ret_91 = __noswap_vfmas_lane_f32(__s0_91, -__s1_91, __rev2_91, __p3_91); \
47775  __ret_91; \
47776})
47777#endif
47778
47779#ifdef __LITTLE_ENDIAN__
47780#define vfmsq_lane_f64(__p0_92, __p1_92, __p2_92, __p3_92) __extension__ ({ \
47781  float64x2_t __s0_92 = __p0_92; \
47782  float64x2_t __s1_92 = __p1_92; \
47783  float64x1_t __s2_92 = __p2_92; \
47784  float64x2_t __ret_92; \
47785  __ret_92 = vfmaq_lane_f64(__s0_92, -__s1_92, __s2_92, __p3_92); \
47786  __ret_92; \
47787})
47788#else
47789#define vfmsq_lane_f64(__p0_93, __p1_93, __p2_93, __p3_93) __extension__ ({ \
47790  float64x2_t __s0_93 = __p0_93; \
47791  float64x2_t __s1_93 = __p1_93; \
47792  float64x1_t __s2_93 = __p2_93; \
47793  float64x2_t __rev0_93;  __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
47794  float64x2_t __rev1_93;  __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
47795  float64x2_t __ret_93; \
47796  __ret_93 = __noswap_vfmaq_lane_f64(__rev0_93, -__rev1_93, __s2_93, __p3_93); \
47797  __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
47798  __ret_93; \
47799})
47800#endif
47801
47802#ifdef __LITTLE_ENDIAN__
47803#define vfmsq_lane_f32(__p0_94, __p1_94, __p2_94, __p3_94) __extension__ ({ \
47804  float32x4_t __s0_94 = __p0_94; \
47805  float32x4_t __s1_94 = __p1_94; \
47806  float32x2_t __s2_94 = __p2_94; \
47807  float32x4_t __ret_94; \
47808  __ret_94 = vfmaq_lane_f32(__s0_94, -__s1_94, __s2_94, __p3_94); \
47809  __ret_94; \
47810})
47811#else
47812#define vfmsq_lane_f32(__p0_95, __p1_95, __p2_95, __p3_95) __extension__ ({ \
47813  float32x4_t __s0_95 = __p0_95; \
47814  float32x4_t __s1_95 = __p1_95; \
47815  float32x2_t __s2_95 = __p2_95; \
47816  float32x4_t __rev0_95;  __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
47817  float32x4_t __rev1_95;  __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
47818  float32x2_t __rev2_95;  __rev2_95 = __builtin_shufflevector(__s2_95, __s2_95, 1, 0); \
47819  float32x4_t __ret_95; \
47820  __ret_95 = __noswap_vfmaq_lane_f32(__rev0_95, -__rev1_95, __rev2_95, __p3_95); \
47821  __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
47822  __ret_95; \
47823})
47824#endif
47825
47826#ifdef __LITTLE_ENDIAN__
47827#define vfms_lane_f64(__p0_96, __p1_96, __p2_96, __p3_96) __extension__ ({ \
47828  float64x1_t __s0_96 = __p0_96; \
47829  float64x1_t __s1_96 = __p1_96; \
47830  float64x1_t __s2_96 = __p2_96; \
47831  float64x1_t __ret_96; \
47832  __ret_96 = vfma_lane_f64(__s0_96, -__s1_96, __s2_96, __p3_96); \
47833  __ret_96; \
47834})
47835#else
47836#define vfms_lane_f64(__p0_97, __p1_97, __p2_97, __p3_97) __extension__ ({ \
47837  float64x1_t __s0_97 = __p0_97; \
47838  float64x1_t __s1_97 = __p1_97; \
47839  float64x1_t __s2_97 = __p2_97; \
47840  float64x1_t __ret_97; \
47841  __ret_97 = __noswap_vfma_lane_f64(__s0_97, -__s1_97, __s2_97, __p3_97); \
47842  __ret_97; \
47843})
47844#endif
47845
47846#ifdef __LITTLE_ENDIAN__
47847#define vfms_lane_f32(__p0_98, __p1_98, __p2_98, __p3_98) __extension__ ({ \
47848  float32x2_t __s0_98 = __p0_98; \
47849  float32x2_t __s1_98 = __p1_98; \
47850  float32x2_t __s2_98 = __p2_98; \
47851  float32x2_t __ret_98; \
47852  __ret_98 = vfma_lane_f32(__s0_98, -__s1_98, __s2_98, __p3_98); \
47853  __ret_98; \
47854})
47855#else
47856#define vfms_lane_f32(__p0_99, __p1_99, __p2_99, __p3_99) __extension__ ({ \
47857  float32x2_t __s0_99 = __p0_99; \
47858  float32x2_t __s1_99 = __p1_99; \
47859  float32x2_t __s2_99 = __p2_99; \
47860  float32x2_t __rev0_99;  __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
47861  float32x2_t __rev1_99;  __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
47862  float32x2_t __rev2_99;  __rev2_99 = __builtin_shufflevector(__s2_99, __s2_99, 1, 0); \
47863  float32x2_t __ret_99; \
47864  __ret_99 = __noswap_vfma_lane_f32(__rev0_99, -__rev1_99, __rev2_99, __p3_99); \
47865  __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
47866  __ret_99; \
47867})
47868#endif
47869
47870#ifdef __LITTLE_ENDIAN__
47871#define vfmsd_laneq_f64(__p0_100, __p1_100, __p2_100, __p3_100) __extension__ ({ \
47872  float64_t __s0_100 = __p0_100; \
47873  float64_t __s1_100 = __p1_100; \
47874  float64x2_t __s2_100 = __p2_100; \
47875  float64_t __ret_100; \
47876  __ret_100 = vfmad_laneq_f64(__s0_100, -__s1_100, __s2_100, __p3_100); \
47877  __ret_100; \
47878})
47879#else
47880#define vfmsd_laneq_f64(__p0_101, __p1_101, __p2_101, __p3_101) __extension__ ({ \
47881  float64_t __s0_101 = __p0_101; \
47882  float64_t __s1_101 = __p1_101; \
47883  float64x2_t __s2_101 = __p2_101; \
47884  float64x2_t __rev2_101;  __rev2_101 = __builtin_shufflevector(__s2_101, __s2_101, 1, 0); \
47885  float64_t __ret_101; \
47886  __ret_101 = __noswap_vfmad_laneq_f64(__s0_101, -__s1_101, __rev2_101, __p3_101); \
47887  __ret_101; \
47888})
47889#endif
47890
47891#ifdef __LITTLE_ENDIAN__
47892#define vfmss_laneq_f32(__p0_102, __p1_102, __p2_102, __p3_102) __extension__ ({ \
47893  float32_t __s0_102 = __p0_102; \
47894  float32_t __s1_102 = __p1_102; \
47895  float32x4_t __s2_102 = __p2_102; \
47896  float32_t __ret_102; \
47897  __ret_102 = vfmas_laneq_f32(__s0_102, -__s1_102, __s2_102, __p3_102); \
47898  __ret_102; \
47899})
47900#else
47901#define vfmss_laneq_f32(__p0_103, __p1_103, __p2_103, __p3_103) __extension__ ({ \
47902  float32_t __s0_103 = __p0_103; \
47903  float32_t __s1_103 = __p1_103; \
47904  float32x4_t __s2_103 = __p2_103; \
47905  float32x4_t __rev2_103;  __rev2_103 = __builtin_shufflevector(__s2_103, __s2_103, 3, 2, 1, 0); \
47906  float32_t __ret_103; \
47907  __ret_103 = __noswap_vfmas_laneq_f32(__s0_103, -__s1_103, __rev2_103, __p3_103); \
47908  __ret_103; \
47909})
47910#endif
47911
47912#ifdef __LITTLE_ENDIAN__
47913#define vfmsq_laneq_f64(__p0_104, __p1_104, __p2_104, __p3_104) __extension__ ({ \
47914  float64x2_t __s0_104 = __p0_104; \
47915  float64x2_t __s1_104 = __p1_104; \
47916  float64x2_t __s2_104 = __p2_104; \
47917  float64x2_t __ret_104; \
47918  __ret_104 = vfmaq_laneq_f64(__s0_104, -__s1_104, __s2_104, __p3_104); \
47919  __ret_104; \
47920})
47921#else
47922#define vfmsq_laneq_f64(__p0_105, __p1_105, __p2_105, __p3_105) __extension__ ({ \
47923  float64x2_t __s0_105 = __p0_105; \
47924  float64x2_t __s1_105 = __p1_105; \
47925  float64x2_t __s2_105 = __p2_105; \
47926  float64x2_t __rev0_105;  __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 1, 0); \
47927  float64x2_t __rev1_105;  __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 1, 0); \
47928  float64x2_t __rev2_105;  __rev2_105 = __builtin_shufflevector(__s2_105, __s2_105, 1, 0); \
47929  float64x2_t __ret_105; \
47930  __ret_105 = __noswap_vfmaq_laneq_f64(__rev0_105, -__rev1_105, __rev2_105, __p3_105); \
47931  __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 1, 0); \
47932  __ret_105; \
47933})
47934#endif
47935
47936#ifdef __LITTLE_ENDIAN__
47937#define vfmsq_laneq_f32(__p0_106, __p1_106, __p2_106, __p3_106) __extension__ ({ \
47938  float32x4_t __s0_106 = __p0_106; \
47939  float32x4_t __s1_106 = __p1_106; \
47940  float32x4_t __s2_106 = __p2_106; \
47941  float32x4_t __ret_106; \
47942  __ret_106 = vfmaq_laneq_f32(__s0_106, -__s1_106, __s2_106, __p3_106); \
47943  __ret_106; \
47944})
47945#else
47946#define vfmsq_laneq_f32(__p0_107, __p1_107, __p2_107, __p3_107) __extension__ ({ \
47947  float32x4_t __s0_107 = __p0_107; \
47948  float32x4_t __s1_107 = __p1_107; \
47949  float32x4_t __s2_107 = __p2_107; \
47950  float32x4_t __rev0_107;  __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 3, 2, 1, 0); \
47951  float32x4_t __rev1_107;  __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 3, 2, 1, 0); \
47952  float32x4_t __rev2_107;  __rev2_107 = __builtin_shufflevector(__s2_107, __s2_107, 3, 2, 1, 0); \
47953  float32x4_t __ret_107; \
47954  __ret_107 = __noswap_vfmaq_laneq_f32(__rev0_107, -__rev1_107, __rev2_107, __p3_107); \
47955  __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 3, 2, 1, 0); \
47956  __ret_107; \
47957})
47958#endif
47959
47960#ifdef __LITTLE_ENDIAN__
47961#define vfms_laneq_f64(__p0_108, __p1_108, __p2_108, __p3_108) __extension__ ({ \
47962  float64x1_t __s0_108 = __p0_108; \
47963  float64x1_t __s1_108 = __p1_108; \
47964  float64x2_t __s2_108 = __p2_108; \
47965  float64x1_t __ret_108; \
47966  __ret_108 = vfma_laneq_f64(__s0_108, -__s1_108, __s2_108, __p3_108); \
47967  __ret_108; \
47968})
47969#else
47970#define vfms_laneq_f64(__p0_109, __p1_109, __p2_109, __p3_109) __extension__ ({ \
47971  float64x1_t __s0_109 = __p0_109; \
47972  float64x1_t __s1_109 = __p1_109; \
47973  float64x2_t __s2_109 = __p2_109; \
47974  float64x2_t __rev2_109;  __rev2_109 = __builtin_shufflevector(__s2_109, __s2_109, 1, 0); \
47975  float64x1_t __ret_109; \
47976  __ret_109 = __noswap_vfma_laneq_f64(__s0_109, -__s1_109, __rev2_109, __p3_109); \
47977  __ret_109; \
47978})
47979#endif
47980
47981#ifdef __LITTLE_ENDIAN__
47982#define vfms_laneq_f32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
47983  float32x2_t __s0_110 = __p0_110; \
47984  float32x2_t __s1_110 = __p1_110; \
47985  float32x4_t __s2_110 = __p2_110; \
47986  float32x2_t __ret_110; \
47987  __ret_110 = vfma_laneq_f32(__s0_110, -__s1_110, __s2_110, __p3_110); \
47988  __ret_110; \
47989})
47990#else
47991#define vfms_laneq_f32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
47992  float32x2_t __s0_111 = __p0_111; \
47993  float32x2_t __s1_111 = __p1_111; \
47994  float32x4_t __s2_111 = __p2_111; \
47995  float32x2_t __rev0_111;  __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
47996  float32x2_t __rev1_111;  __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
47997  float32x4_t __rev2_111;  __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 3, 2, 1, 0); \
47998  float32x2_t __ret_111; \
47999  __ret_111 = __noswap_vfma_laneq_f32(__rev0_111, -__rev1_111, __rev2_111, __p3_111); \
48000  __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
48001  __ret_111; \
48002})
48003#endif
48004
48005#ifdef __LITTLE_ENDIAN__
48006__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48007  float64x2_t __ret;
48008  __ret = vfmaq_f64(__p0, -__p1, (float64x2_t) {__p2, __p2});
48009  return __ret;
48010}
48011#else
48012__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48013  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48014  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48015  float64x2_t __ret;
48016  __ret = __noswap_vfmaq_f64(__rev0, -__rev1, (float64x2_t) {__p2, __p2});
48017  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48018  return __ret;
48019}
48020#endif
48021
48022#ifdef __LITTLE_ENDIAN__
48023__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48024  float32x4_t __ret;
48025  __ret = vfmaq_f32(__p0, -__p1, (float32x4_t) {__p2, __p2, __p2, __p2});
48026  return __ret;
48027}
48028#else
48029__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48030  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48031  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
48032  float32x4_t __ret;
48033  __ret = __noswap_vfmaq_f32(__rev0, -__rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
48034  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48035  return __ret;
48036}
48037#endif
48038
48039#ifdef __LITTLE_ENDIAN__
48040__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48041  float32x2_t __ret;
48042  __ret = vfma_f32(__p0, -__p1, (float32x2_t) {__p2, __p2});
48043  return __ret;
48044}
48045#else
48046__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48047  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48048  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48049  float32x2_t __ret;
48050  __ret = __noswap_vfma_f32(__rev0, -__rev1, (float32x2_t) {__p2, __p2});
48051  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48052  return __ret;
48053}
48054#endif
48055
48056#ifdef __LITTLE_ENDIAN__
48057__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48058  poly64x1_t __ret;
48059  __ret = __builtin_shufflevector(__p0, __p0, 1);
48060  return __ret;
48061}
48062#else
48063__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48064  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48065  poly64x1_t __ret;
48066  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48067  return __ret;
48068}
48069__ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
48070  poly64x1_t __ret;
48071  __ret = __builtin_shufflevector(__p0, __p0, 1);
48072  return __ret;
48073}
48074#endif
48075
48076#ifdef __LITTLE_ENDIAN__
48077__ai float64x1_t vget_high_f64(float64x2_t __p0) {
48078  float64x1_t __ret;
48079  __ret = __builtin_shufflevector(__p0, __p0, 1);
48080  return __ret;
48081}
48082#else
48083__ai float64x1_t vget_high_f64(float64x2_t __p0) {
48084  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48085  float64x1_t __ret;
48086  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48087  return __ret;
48088}
48089#endif
48090
48091#ifdef __LITTLE_ENDIAN__
48092#define vget_lane_p64(__p0, __p1) __extension__ ({ \
48093  poly64x1_t __s0 = __p0; \
48094  poly64_t __ret; \
48095  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48096  __ret; \
48097})
48098#else
48099#define vget_lane_p64(__p0, __p1) __extension__ ({ \
48100  poly64x1_t __s0 = __p0; \
48101  poly64_t __ret; \
48102  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48103  __ret; \
48104})
48105#define __noswap_vget_lane_p64(__p0, __p1) __extension__ ({ \
48106  poly64x1_t __s0 = __p0; \
48107  poly64_t __ret; \
48108  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48109  __ret; \
48110})
48111#endif
48112
48113#ifdef __LITTLE_ENDIAN__
48114#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48115  poly64x2_t __s0 = __p0; \
48116  poly64_t __ret; \
48117  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48118  __ret; \
48119})
48120#else
48121#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48122  poly64x2_t __s0 = __p0; \
48123  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48124  poly64_t __ret; \
48125  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
48126  __ret; \
48127})
48128#define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48129  poly64x2_t __s0 = __p0; \
48130  poly64_t __ret; \
48131  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48132  __ret; \
48133})
48134#endif
48135
48136#ifdef __LITTLE_ENDIAN__
48137#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48138  float64x2_t __s0 = __p0; \
48139  float64_t __ret; \
48140  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48141  __ret; \
48142})
48143#else
48144#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48145  float64x2_t __s0 = __p0; \
48146  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48147  float64_t __ret; \
48148  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__rev0, __p1); \
48149  __ret; \
48150})
48151#define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48152  float64x2_t __s0 = __p0; \
48153  float64_t __ret; \
48154  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48155  __ret; \
48156})
48157#endif
48158
48159#ifdef __LITTLE_ENDIAN__
48160#define vget_lane_f64(__p0, __p1) __extension__ ({ \
48161  float64x1_t __s0 = __p0; \
48162  float64_t __ret; \
48163  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48164  __ret; \
48165})
48166#else
48167#define vget_lane_f64(__p0, __p1) __extension__ ({ \
48168  float64x1_t __s0 = __p0; \
48169  float64_t __ret; \
48170  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48171  __ret; \
48172})
48173#define __noswap_vget_lane_f64(__p0, __p1) __extension__ ({ \
48174  float64x1_t __s0 = __p0; \
48175  float64_t __ret; \
48176  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48177  __ret; \
48178})
48179#endif
48180
48181#ifdef __LITTLE_ENDIAN__
48182__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48183  poly64x1_t __ret;
48184  __ret = __builtin_shufflevector(__p0, __p0, 0);
48185  return __ret;
48186}
48187#else
48188__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48189  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48190  poly64x1_t __ret;
48191  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48192  return __ret;
48193}
48194#endif
48195
48196#ifdef __LITTLE_ENDIAN__
48197__ai float64x1_t vget_low_f64(float64x2_t __p0) {
48198  float64x1_t __ret;
48199  __ret = __builtin_shufflevector(__p0, __p0, 0);
48200  return __ret;
48201}
48202#else
48203__ai float64x1_t vget_low_f64(float64x2_t __p0) {
48204  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48205  float64x1_t __ret;
48206  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48207  return __ret;
48208}
48209#endif
48210
48211#ifdef __LITTLE_ENDIAN__
48212#define vld1_p64(__p0) __extension__ ({ \
48213  poly64x1_t __ret; \
48214  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48215  __ret; \
48216})
48217#else
48218#define vld1_p64(__p0) __extension__ ({ \
48219  poly64x1_t __ret; \
48220  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48221  __ret; \
48222})
48223#endif
48224
48225#ifdef __LITTLE_ENDIAN__
48226#define vld1q_p64(__p0) __extension__ ({ \
48227  poly64x2_t __ret; \
48228  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48229  __ret; \
48230})
48231#else
48232#define vld1q_p64(__p0) __extension__ ({ \
48233  poly64x2_t __ret; \
48234  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48235  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48236  __ret; \
48237})
48238#endif
48239
48240#ifdef __LITTLE_ENDIAN__
48241#define vld1q_f64(__p0) __extension__ ({ \
48242  float64x2_t __ret; \
48243  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48244  __ret; \
48245})
48246#else
48247#define vld1q_f64(__p0) __extension__ ({ \
48248  float64x2_t __ret; \
48249  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48250  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48251  __ret; \
48252})
48253#endif
48254
48255#ifdef __LITTLE_ENDIAN__
48256#define vld1_f64(__p0) __extension__ ({ \
48257  float64x1_t __ret; \
48258  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48259  __ret; \
48260})
48261#else
48262#define vld1_f64(__p0) __extension__ ({ \
48263  float64x1_t __ret; \
48264  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48265  __ret; \
48266})
48267#endif
48268
48269#ifdef __LITTLE_ENDIAN__
48270#define vld1_dup_p64(__p0) __extension__ ({ \
48271  poly64x1_t __ret; \
48272  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48273  __ret; \
48274})
48275#else
48276#define vld1_dup_p64(__p0) __extension__ ({ \
48277  poly64x1_t __ret; \
48278  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48279  __ret; \
48280})
48281#endif
48282
48283#ifdef __LITTLE_ENDIAN__
48284#define vld1q_dup_p64(__p0) __extension__ ({ \
48285  poly64x2_t __ret; \
48286  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48287  __ret; \
48288})
48289#else
48290#define vld1q_dup_p64(__p0) __extension__ ({ \
48291  poly64x2_t __ret; \
48292  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48294  __ret; \
48295})
48296#endif
48297
48298#ifdef __LITTLE_ENDIAN__
48299#define vld1q_dup_f64(__p0) __extension__ ({ \
48300  float64x2_t __ret; \
48301  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48302  __ret; \
48303})
48304#else
48305#define vld1q_dup_f64(__p0) __extension__ ({ \
48306  float64x2_t __ret; \
48307  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48308  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48309  __ret; \
48310})
48311#endif
48312
48313#ifdef __LITTLE_ENDIAN__
48314#define vld1_dup_f64(__p0) __extension__ ({ \
48315  float64x1_t __ret; \
48316  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48317  __ret; \
48318})
48319#else
48320#define vld1_dup_f64(__p0) __extension__ ({ \
48321  float64x1_t __ret; \
48322  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48323  __ret; \
48324})
48325#endif
48326
48327#ifdef __LITTLE_ENDIAN__
48328#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48329  poly64x1_t __s1 = __p1; \
48330  poly64x1_t __ret; \
48331  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48332  __ret; \
48333})
48334#else
48335#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48336  poly64x1_t __s1 = __p1; \
48337  poly64x1_t __ret; \
48338  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48339  __ret; \
48340})
48341#endif
48342
48343#ifdef __LITTLE_ENDIAN__
48344#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48345  poly64x2_t __s1 = __p1; \
48346  poly64x2_t __ret; \
48347  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
48348  __ret; \
48349})
48350#else
48351#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48352  poly64x2_t __s1 = __p1; \
48353  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48354  poly64x2_t __ret; \
48355  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
48356  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48357  __ret; \
48358})
48359#endif
48360
48361#ifdef __LITTLE_ENDIAN__
48362#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48363  float64x2_t __s1 = __p1; \
48364  float64x2_t __ret; \
48365  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
48366  __ret; \
48367})
48368#else
48369#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48370  float64x2_t __s1 = __p1; \
48371  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48372  float64x2_t __ret; \
48373  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
48374  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48375  __ret; \
48376})
48377#endif
48378
48379#ifdef __LITTLE_ENDIAN__
48380#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48381  float64x1_t __s1 = __p1; \
48382  float64x1_t __ret; \
48383  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48384  __ret; \
48385})
48386#else
48387#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48388  float64x1_t __s1 = __p1; \
48389  float64x1_t __ret; \
48390  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48391  __ret; \
48392})
48393#endif
48394
48395#ifdef __LITTLE_ENDIAN__
48396#define vld1_p8_x2(__p0) __extension__ ({ \
48397  poly8x8x2_t __ret; \
48398  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48399  __ret; \
48400})
48401#else
48402#define vld1_p8_x2(__p0) __extension__ ({ \
48403  poly8x8x2_t __ret; \
48404  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48405 \
48406  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48407  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48408  __ret; \
48409})
48410#endif
48411
48412#ifdef __LITTLE_ENDIAN__
48413#define vld1_p64_x2(__p0) __extension__ ({ \
48414  poly64x1x2_t __ret; \
48415  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48416  __ret; \
48417})
48418#else
48419#define vld1_p64_x2(__p0) __extension__ ({ \
48420  poly64x1x2_t __ret; \
48421  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48422  __ret; \
48423})
48424#endif
48425
48426#ifdef __LITTLE_ENDIAN__
48427#define vld1_p16_x2(__p0) __extension__ ({ \
48428  poly16x4x2_t __ret; \
48429  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48430  __ret; \
48431})
48432#else
48433#define vld1_p16_x2(__p0) __extension__ ({ \
48434  poly16x4x2_t __ret; \
48435  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48436 \
48437  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48438  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48439  __ret; \
48440})
48441#endif
48442
48443#ifdef __LITTLE_ENDIAN__
48444#define vld1q_p8_x2(__p0) __extension__ ({ \
48445  poly8x16x2_t __ret; \
48446  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48447  __ret; \
48448})
48449#else
48450#define vld1q_p8_x2(__p0) __extension__ ({ \
48451  poly8x16x2_t __ret; \
48452  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48453 \
48454  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48455  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48456  __ret; \
48457})
48458#endif
48459
48460#ifdef __LITTLE_ENDIAN__
48461#define vld1q_p64_x2(__p0) __extension__ ({ \
48462  poly64x2x2_t __ret; \
48463  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48464  __ret; \
48465})
48466#else
48467#define vld1q_p64_x2(__p0) __extension__ ({ \
48468  poly64x2x2_t __ret; \
48469  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48470 \
48471  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48472  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48473  __ret; \
48474})
48475#endif
48476
48477#ifdef __LITTLE_ENDIAN__
48478#define vld1q_p16_x2(__p0) __extension__ ({ \
48479  poly16x8x2_t __ret; \
48480  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48481  __ret; \
48482})
48483#else
48484#define vld1q_p16_x2(__p0) __extension__ ({ \
48485  poly16x8x2_t __ret; \
48486  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48487 \
48488  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48489  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48490  __ret; \
48491})
48492#endif
48493
48494#ifdef __LITTLE_ENDIAN__
48495#define vld1q_u8_x2(__p0) __extension__ ({ \
48496  uint8x16x2_t __ret; \
48497  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48498  __ret; \
48499})
48500#else
48501#define vld1q_u8_x2(__p0) __extension__ ({ \
48502  uint8x16x2_t __ret; \
48503  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48504 \
48505  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48506  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48507  __ret; \
48508})
48509#endif
48510
48511#ifdef __LITTLE_ENDIAN__
48512#define vld1q_u32_x2(__p0) __extension__ ({ \
48513  uint32x4x2_t __ret; \
48514  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48515  __ret; \
48516})
48517#else
48518#define vld1q_u32_x2(__p0) __extension__ ({ \
48519  uint32x4x2_t __ret; \
48520  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48521 \
48522  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48523  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48524  __ret; \
48525})
48526#endif
48527
48528#ifdef __LITTLE_ENDIAN__
48529#define vld1q_u64_x2(__p0) __extension__ ({ \
48530  uint64x2x2_t __ret; \
48531  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48532  __ret; \
48533})
48534#else
48535#define vld1q_u64_x2(__p0) __extension__ ({ \
48536  uint64x2x2_t __ret; \
48537  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48538 \
48539  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48540  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48541  __ret; \
48542})
48543#endif
48544
48545#ifdef __LITTLE_ENDIAN__
48546#define vld1q_u16_x2(__p0) __extension__ ({ \
48547  uint16x8x2_t __ret; \
48548  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48549  __ret; \
48550})
48551#else
48552#define vld1q_u16_x2(__p0) __extension__ ({ \
48553  uint16x8x2_t __ret; \
48554  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48555 \
48556  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48557  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48558  __ret; \
48559})
48560#endif
48561
48562#ifdef __LITTLE_ENDIAN__
48563#define vld1q_s8_x2(__p0) __extension__ ({ \
48564  int8x16x2_t __ret; \
48565  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48566  __ret; \
48567})
48568#else
48569#define vld1q_s8_x2(__p0) __extension__ ({ \
48570  int8x16x2_t __ret; \
48571  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48572 \
48573  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48574  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48575  __ret; \
48576})
48577#endif
48578
48579#ifdef __LITTLE_ENDIAN__
48580#define vld1q_f64_x2(__p0) __extension__ ({ \
48581  float64x2x2_t __ret; \
48582  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48583  __ret; \
48584})
48585#else
48586#define vld1q_f64_x2(__p0) __extension__ ({ \
48587  float64x2x2_t __ret; \
48588  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48589 \
48590  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48591  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48592  __ret; \
48593})
48594#endif
48595
48596#ifdef __LITTLE_ENDIAN__
48597#define vld1q_f32_x2(__p0) __extension__ ({ \
48598  float32x4x2_t __ret; \
48599  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48600  __ret; \
48601})
48602#else
48603#define vld1q_f32_x2(__p0) __extension__ ({ \
48604  float32x4x2_t __ret; \
48605  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48606 \
48607  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48608  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48609  __ret; \
48610})
48611#endif
48612
48613#ifdef __LITTLE_ENDIAN__
48614#define vld1q_f16_x2(__p0) __extension__ ({ \
48615  float16x8x2_t __ret; \
48616  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48617  __ret; \
48618})
48619#else
48620#define vld1q_f16_x2(__p0) __extension__ ({ \
48621  float16x8x2_t __ret; \
48622  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48623 \
48624  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48625  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48626  __ret; \
48627})
48628#endif
48629
48630#ifdef __LITTLE_ENDIAN__
48631#define vld1q_s32_x2(__p0) __extension__ ({ \
48632  int32x4x2_t __ret; \
48633  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48634  __ret; \
48635})
48636#else
48637#define vld1q_s32_x2(__p0) __extension__ ({ \
48638  int32x4x2_t __ret; \
48639  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48640 \
48641  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48642  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48643  __ret; \
48644})
48645#endif
48646
48647#ifdef __LITTLE_ENDIAN__
48648#define vld1q_s64_x2(__p0) __extension__ ({ \
48649  int64x2x2_t __ret; \
48650  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48651  __ret; \
48652})
48653#else
48654#define vld1q_s64_x2(__p0) __extension__ ({ \
48655  int64x2x2_t __ret; \
48656  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48657 \
48658  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48659  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48660  __ret; \
48661})
48662#endif
48663
48664#ifdef __LITTLE_ENDIAN__
48665#define vld1q_s16_x2(__p0) __extension__ ({ \
48666  int16x8x2_t __ret; \
48667  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48668  __ret; \
48669})
48670#else
48671#define vld1q_s16_x2(__p0) __extension__ ({ \
48672  int16x8x2_t __ret; \
48673  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48674 \
48675  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48676  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48677  __ret; \
48678})
48679#endif
48680
48681#ifdef __LITTLE_ENDIAN__
48682#define vld1_u8_x2(__p0) __extension__ ({ \
48683  uint8x8x2_t __ret; \
48684  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48685  __ret; \
48686})
48687#else
48688#define vld1_u8_x2(__p0) __extension__ ({ \
48689  uint8x8x2_t __ret; \
48690  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48691 \
48692  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48693  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48694  __ret; \
48695})
48696#endif
48697
48698#ifdef __LITTLE_ENDIAN__
48699#define vld1_u32_x2(__p0) __extension__ ({ \
48700  uint32x2x2_t __ret; \
48701  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48702  __ret; \
48703})
48704#else
48705#define vld1_u32_x2(__p0) __extension__ ({ \
48706  uint32x2x2_t __ret; \
48707  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48708 \
48709  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48710  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48711  __ret; \
48712})
48713#endif
48714
48715#ifdef __LITTLE_ENDIAN__
48716#define vld1_u64_x2(__p0) __extension__ ({ \
48717  uint64x1x2_t __ret; \
48718  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48719  __ret; \
48720})
48721#else
48722#define vld1_u64_x2(__p0) __extension__ ({ \
48723  uint64x1x2_t __ret; \
48724  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48725  __ret; \
48726})
48727#endif
48728
48729#ifdef __LITTLE_ENDIAN__
48730#define vld1_u16_x2(__p0) __extension__ ({ \
48731  uint16x4x2_t __ret; \
48732  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48733  __ret; \
48734})
48735#else
48736#define vld1_u16_x2(__p0) __extension__ ({ \
48737  uint16x4x2_t __ret; \
48738  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48739 \
48740  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48741  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48742  __ret; \
48743})
48744#endif
48745
48746#ifdef __LITTLE_ENDIAN__
48747#define vld1_s8_x2(__p0) __extension__ ({ \
48748  int8x8x2_t __ret; \
48749  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48750  __ret; \
48751})
48752#else
48753#define vld1_s8_x2(__p0) __extension__ ({ \
48754  int8x8x2_t __ret; \
48755  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48756 \
48757  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48758  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48759  __ret; \
48760})
48761#endif
48762
48763#ifdef __LITTLE_ENDIAN__
48764#define vld1_f64_x2(__p0) __extension__ ({ \
48765  float64x1x2_t __ret; \
48766  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48767  __ret; \
48768})
48769#else
48770#define vld1_f64_x2(__p0) __extension__ ({ \
48771  float64x1x2_t __ret; \
48772  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48773  __ret; \
48774})
48775#endif
48776
48777#ifdef __LITTLE_ENDIAN__
48778#define vld1_f32_x2(__p0) __extension__ ({ \
48779  float32x2x2_t __ret; \
48780  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48781  __ret; \
48782})
48783#else
48784#define vld1_f32_x2(__p0) __extension__ ({ \
48785  float32x2x2_t __ret; \
48786  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48787 \
48788  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48789  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48790  __ret; \
48791})
48792#endif
48793
48794#ifdef __LITTLE_ENDIAN__
48795#define vld1_f16_x2(__p0) __extension__ ({ \
48796  float16x4x2_t __ret; \
48797  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48798  __ret; \
48799})
48800#else
48801#define vld1_f16_x2(__p0) __extension__ ({ \
48802  float16x4x2_t __ret; \
48803  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48804 \
48805  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48806  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48807  __ret; \
48808})
48809#endif
48810
48811#ifdef __LITTLE_ENDIAN__
48812#define vld1_s32_x2(__p0) __extension__ ({ \
48813  int32x2x2_t __ret; \
48814  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48815  __ret; \
48816})
48817#else
48818#define vld1_s32_x2(__p0) __extension__ ({ \
48819  int32x2x2_t __ret; \
48820  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48821 \
48822  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48823  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48824  __ret; \
48825})
48826#endif
48827
48828#ifdef __LITTLE_ENDIAN__
48829#define vld1_s64_x2(__p0) __extension__ ({ \
48830  int64x1x2_t __ret; \
48831  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48832  __ret; \
48833})
48834#else
48835#define vld1_s64_x2(__p0) __extension__ ({ \
48836  int64x1x2_t __ret; \
48837  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48838  __ret; \
48839})
48840#endif
48841
48842#ifdef __LITTLE_ENDIAN__
48843#define vld1_s16_x2(__p0) __extension__ ({ \
48844  int16x4x2_t __ret; \
48845  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48846  __ret; \
48847})
48848#else
48849#define vld1_s16_x2(__p0) __extension__ ({ \
48850  int16x4x2_t __ret; \
48851  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48852 \
48853  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48854  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48855  __ret; \
48856})
48857#endif
48858
48859#ifdef __LITTLE_ENDIAN__
48860#define vld1_p8_x3(__p0) __extension__ ({ \
48861  poly8x8x3_t __ret; \
48862  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48863  __ret; \
48864})
48865#else
48866#define vld1_p8_x3(__p0) __extension__ ({ \
48867  poly8x8x3_t __ret; \
48868  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48869 \
48870  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48871  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48872  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48873  __ret; \
48874})
48875#endif
48876
48877#ifdef __LITTLE_ENDIAN__
48878#define vld1_p64_x3(__p0) __extension__ ({ \
48879  poly64x1x3_t __ret; \
48880  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48881  __ret; \
48882})
48883#else
48884#define vld1_p64_x3(__p0) __extension__ ({ \
48885  poly64x1x3_t __ret; \
48886  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48887  __ret; \
48888})
48889#endif
48890
48891#ifdef __LITTLE_ENDIAN__
48892#define vld1_p16_x3(__p0) __extension__ ({ \
48893  poly16x4x3_t __ret; \
48894  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48895  __ret; \
48896})
48897#else
48898#define vld1_p16_x3(__p0) __extension__ ({ \
48899  poly16x4x3_t __ret; \
48900  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48901 \
48902  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48903  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48904  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48905  __ret; \
48906})
48907#endif
48908
48909#ifdef __LITTLE_ENDIAN__
48910#define vld1q_p8_x3(__p0) __extension__ ({ \
48911  poly8x16x3_t __ret; \
48912  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48913  __ret; \
48914})
48915#else
48916#define vld1q_p8_x3(__p0) __extension__ ({ \
48917  poly8x16x3_t __ret; \
48918  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48919 \
48920  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48921  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48922  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48923  __ret; \
48924})
48925#endif
48926
48927#ifdef __LITTLE_ENDIAN__
48928#define vld1q_p64_x3(__p0) __extension__ ({ \
48929  poly64x2x3_t __ret; \
48930  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48931  __ret; \
48932})
48933#else
48934#define vld1q_p64_x3(__p0) __extension__ ({ \
48935  poly64x2x3_t __ret; \
48936  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48937 \
48938  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48939  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48940  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48941  __ret; \
48942})
48943#endif
48944
48945#ifdef __LITTLE_ENDIAN__
48946#define vld1q_p16_x3(__p0) __extension__ ({ \
48947  poly16x8x3_t __ret; \
48948  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48949  __ret; \
48950})
48951#else
48952#define vld1q_p16_x3(__p0) __extension__ ({ \
48953  poly16x8x3_t __ret; \
48954  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48955 \
48956  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48957  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48958  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48959  __ret; \
48960})
48961#endif
48962
48963#ifdef __LITTLE_ENDIAN__
48964#define vld1q_u8_x3(__p0) __extension__ ({ \
48965  uint8x16x3_t __ret; \
48966  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48967  __ret; \
48968})
48969#else
48970#define vld1q_u8_x3(__p0) __extension__ ({ \
48971  uint8x16x3_t __ret; \
48972  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48973 \
48974  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48975  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48976  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48977  __ret; \
48978})
48979#endif
48980
48981#ifdef __LITTLE_ENDIAN__
48982#define vld1q_u32_x3(__p0) __extension__ ({ \
48983  uint32x4x3_t __ret; \
48984  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48985  __ret; \
48986})
48987#else
48988#define vld1q_u32_x3(__p0) __extension__ ({ \
48989  uint32x4x3_t __ret; \
48990  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48991 \
48992  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48993  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48994  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48995  __ret; \
48996})
48997#endif
48998
48999#ifdef __LITTLE_ENDIAN__
49000#define vld1q_u64_x3(__p0) __extension__ ({ \
49001  uint64x2x3_t __ret; \
49002  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49003  __ret; \
49004})
49005#else
49006#define vld1q_u64_x3(__p0) __extension__ ({ \
49007  uint64x2x3_t __ret; \
49008  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49009 \
49010  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49011  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49012  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49013  __ret; \
49014})
49015#endif
49016
49017#ifdef __LITTLE_ENDIAN__
49018#define vld1q_u16_x3(__p0) __extension__ ({ \
49019  uint16x8x3_t __ret; \
49020  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49021  __ret; \
49022})
49023#else
49024#define vld1q_u16_x3(__p0) __extension__ ({ \
49025  uint16x8x3_t __ret; \
49026  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49027 \
49028  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49029  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49030  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49031  __ret; \
49032})
49033#endif
49034
49035#ifdef __LITTLE_ENDIAN__
49036#define vld1q_s8_x3(__p0) __extension__ ({ \
49037  int8x16x3_t __ret; \
49038  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49039  __ret; \
49040})
49041#else
49042#define vld1q_s8_x3(__p0) __extension__ ({ \
49043  int8x16x3_t __ret; \
49044  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49045 \
49046  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49047  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49048  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49049  __ret; \
49050})
49051#endif
49052
49053#ifdef __LITTLE_ENDIAN__
49054#define vld1q_f64_x3(__p0) __extension__ ({ \
49055  float64x2x3_t __ret; \
49056  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49057  __ret; \
49058})
49059#else
49060#define vld1q_f64_x3(__p0) __extension__ ({ \
49061  float64x2x3_t __ret; \
49062  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49063 \
49064  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49065  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49066  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49067  __ret; \
49068})
49069#endif
49070
49071#ifdef __LITTLE_ENDIAN__
49072#define vld1q_f32_x3(__p0) __extension__ ({ \
49073  float32x4x3_t __ret; \
49074  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49075  __ret; \
49076})
49077#else
49078#define vld1q_f32_x3(__p0) __extension__ ({ \
49079  float32x4x3_t __ret; \
49080  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49081 \
49082  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49083  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49084  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49085  __ret; \
49086})
49087#endif
49088
49089#ifdef __LITTLE_ENDIAN__
49090#define vld1q_f16_x3(__p0) __extension__ ({ \
49091  float16x8x3_t __ret; \
49092  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49093  __ret; \
49094})
49095#else
49096#define vld1q_f16_x3(__p0) __extension__ ({ \
49097  float16x8x3_t __ret; \
49098  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49099 \
49100  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49101  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49102  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49103  __ret; \
49104})
49105#endif
49106
49107#ifdef __LITTLE_ENDIAN__
49108#define vld1q_s32_x3(__p0) __extension__ ({ \
49109  int32x4x3_t __ret; \
49110  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49111  __ret; \
49112})
49113#else
49114#define vld1q_s32_x3(__p0) __extension__ ({ \
49115  int32x4x3_t __ret; \
49116  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49117 \
49118  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49119  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49120  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49121  __ret; \
49122})
49123#endif
49124
49125#ifdef __LITTLE_ENDIAN__
49126#define vld1q_s64_x3(__p0) __extension__ ({ \
49127  int64x2x3_t __ret; \
49128  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49129  __ret; \
49130})
49131#else
49132#define vld1q_s64_x3(__p0) __extension__ ({ \
49133  int64x2x3_t __ret; \
49134  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49135 \
49136  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49137  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49138  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49139  __ret; \
49140})
49141#endif
49142
49143#ifdef __LITTLE_ENDIAN__
49144#define vld1q_s16_x3(__p0) __extension__ ({ \
49145  int16x8x3_t __ret; \
49146  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49147  __ret; \
49148})
49149#else
49150#define vld1q_s16_x3(__p0) __extension__ ({ \
49151  int16x8x3_t __ret; \
49152  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49153 \
49154  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49155  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49156  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49157  __ret; \
49158})
49159#endif
49160
49161#ifdef __LITTLE_ENDIAN__
49162#define vld1_u8_x3(__p0) __extension__ ({ \
49163  uint8x8x3_t __ret; \
49164  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49165  __ret; \
49166})
49167#else
49168#define vld1_u8_x3(__p0) __extension__ ({ \
49169  uint8x8x3_t __ret; \
49170  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49171 \
49172  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49173  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49174  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49175  __ret; \
49176})
49177#endif
49178
49179#ifdef __LITTLE_ENDIAN__
49180#define vld1_u32_x3(__p0) __extension__ ({ \
49181  uint32x2x3_t __ret; \
49182  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49183  __ret; \
49184})
49185#else
49186#define vld1_u32_x3(__p0) __extension__ ({ \
49187  uint32x2x3_t __ret; \
49188  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49189 \
49190  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49191  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49192  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49193  __ret; \
49194})
49195#endif
49196
49197#ifdef __LITTLE_ENDIAN__
49198#define vld1_u64_x3(__p0) __extension__ ({ \
49199  uint64x1x3_t __ret; \
49200  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49201  __ret; \
49202})
49203#else
49204#define vld1_u64_x3(__p0) __extension__ ({ \
49205  uint64x1x3_t __ret; \
49206  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49207  __ret; \
49208})
49209#endif
49210
49211#ifdef __LITTLE_ENDIAN__
49212#define vld1_u16_x3(__p0) __extension__ ({ \
49213  uint16x4x3_t __ret; \
49214  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49215  __ret; \
49216})
49217#else
49218#define vld1_u16_x3(__p0) __extension__ ({ \
49219  uint16x4x3_t __ret; \
49220  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49221 \
49222  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49223  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49224  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49225  __ret; \
49226})
49227#endif
49228
49229#ifdef __LITTLE_ENDIAN__
49230#define vld1_s8_x3(__p0) __extension__ ({ \
49231  int8x8x3_t __ret; \
49232  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49233  __ret; \
49234})
49235#else
49236#define vld1_s8_x3(__p0) __extension__ ({ \
49237  int8x8x3_t __ret; \
49238  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49239 \
49240  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49241  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49242  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49243  __ret; \
49244})
49245#endif
49246
49247#ifdef __LITTLE_ENDIAN__
49248#define vld1_f64_x3(__p0) __extension__ ({ \
49249  float64x1x3_t __ret; \
49250  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49251  __ret; \
49252})
49253#else
49254#define vld1_f64_x3(__p0) __extension__ ({ \
49255  float64x1x3_t __ret; \
49256  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49257  __ret; \
49258})
49259#endif
49260
49261#ifdef __LITTLE_ENDIAN__
49262#define vld1_f32_x3(__p0) __extension__ ({ \
49263  float32x2x3_t __ret; \
49264  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49265  __ret; \
49266})
49267#else
49268#define vld1_f32_x3(__p0) __extension__ ({ \
49269  float32x2x3_t __ret; \
49270  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49271 \
49272  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49273  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49274  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49275  __ret; \
49276})
49277#endif
49278
49279#ifdef __LITTLE_ENDIAN__
49280#define vld1_f16_x3(__p0) __extension__ ({ \
49281  float16x4x3_t __ret; \
49282  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49283  __ret; \
49284})
49285#else
49286#define vld1_f16_x3(__p0) __extension__ ({ \
49287  float16x4x3_t __ret; \
49288  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49289 \
49290  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49291  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49292  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49293  __ret; \
49294})
49295#endif
49296
49297#ifdef __LITTLE_ENDIAN__
49298#define vld1_s32_x3(__p0) __extension__ ({ \
49299  int32x2x3_t __ret; \
49300  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49301  __ret; \
49302})
49303#else
49304#define vld1_s32_x3(__p0) __extension__ ({ \
49305  int32x2x3_t __ret; \
49306  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49307 \
49308  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49309  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49310  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49311  __ret; \
49312})
49313#endif
49314
49315#ifdef __LITTLE_ENDIAN__
49316#define vld1_s64_x3(__p0) __extension__ ({ \
49317  int64x1x3_t __ret; \
49318  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49319  __ret; \
49320})
49321#else
49322#define vld1_s64_x3(__p0) __extension__ ({ \
49323  int64x1x3_t __ret; \
49324  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49325  __ret; \
49326})
49327#endif
49328
49329#ifdef __LITTLE_ENDIAN__
49330#define vld1_s16_x3(__p0) __extension__ ({ \
49331  int16x4x3_t __ret; \
49332  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49333  __ret; \
49334})
49335#else
49336#define vld1_s16_x3(__p0) __extension__ ({ \
49337  int16x4x3_t __ret; \
49338  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49339 \
49340  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49341  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49342  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49343  __ret; \
49344})
49345#endif
49346
49347#ifdef __LITTLE_ENDIAN__
49348#define vld1_p8_x4(__p0) __extension__ ({ \
49349  poly8x8x4_t __ret; \
49350  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49351  __ret; \
49352})
49353#else
49354#define vld1_p8_x4(__p0) __extension__ ({ \
49355  poly8x8x4_t __ret; \
49356  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49357 \
49358  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49359  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49360  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49361  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49362  __ret; \
49363})
49364#endif
49365
49366#ifdef __LITTLE_ENDIAN__
49367#define vld1_p64_x4(__p0) __extension__ ({ \
49368  poly64x1x4_t __ret; \
49369  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49370  __ret; \
49371})
49372#else
49373#define vld1_p64_x4(__p0) __extension__ ({ \
49374  poly64x1x4_t __ret; \
49375  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49376  __ret; \
49377})
49378#endif
49379
49380#ifdef __LITTLE_ENDIAN__
49381#define vld1_p16_x4(__p0) __extension__ ({ \
49382  poly16x4x4_t __ret; \
49383  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49384  __ret; \
49385})
49386#else
49387#define vld1_p16_x4(__p0) __extension__ ({ \
49388  poly16x4x4_t __ret; \
49389  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49390 \
49391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49393  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49394  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49395  __ret; \
49396})
49397#endif
49398
49399#ifdef __LITTLE_ENDIAN__
49400#define vld1q_p8_x4(__p0) __extension__ ({ \
49401  poly8x16x4_t __ret; \
49402  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49403  __ret; \
49404})
49405#else
49406#define vld1q_p8_x4(__p0) __extension__ ({ \
49407  poly8x16x4_t __ret; \
49408  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49409 \
49410  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49411  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49412  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49413  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49414  __ret; \
49415})
49416#endif
49417
49418#ifdef __LITTLE_ENDIAN__
49419#define vld1q_p64_x4(__p0) __extension__ ({ \
49420  poly64x2x4_t __ret; \
49421  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49422  __ret; \
49423})
49424#else
49425#define vld1q_p64_x4(__p0) __extension__ ({ \
49426  poly64x2x4_t __ret; \
49427  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49428 \
49429  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49430  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49431  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49432  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49433  __ret; \
49434})
49435#endif
49436
49437#ifdef __LITTLE_ENDIAN__
49438#define vld1q_p16_x4(__p0) __extension__ ({ \
49439  poly16x8x4_t __ret; \
49440  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49441  __ret; \
49442})
49443#else
49444#define vld1q_p16_x4(__p0) __extension__ ({ \
49445  poly16x8x4_t __ret; \
49446  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49447 \
49448  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49449  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49450  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49451  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49452  __ret; \
49453})
49454#endif
49455
49456#ifdef __LITTLE_ENDIAN__
49457#define vld1q_u8_x4(__p0) __extension__ ({ \
49458  uint8x16x4_t __ret; \
49459  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49460  __ret; \
49461})
49462#else
49463#define vld1q_u8_x4(__p0) __extension__ ({ \
49464  uint8x16x4_t __ret; \
49465  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49466 \
49467  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49468  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49469  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49470  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49471  __ret; \
49472})
49473#endif
49474
49475#ifdef __LITTLE_ENDIAN__
49476#define vld1q_u32_x4(__p0) __extension__ ({ \
49477  uint32x4x4_t __ret; \
49478  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49479  __ret; \
49480})
49481#else
49482#define vld1q_u32_x4(__p0) __extension__ ({ \
49483  uint32x4x4_t __ret; \
49484  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49485 \
49486  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49487  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49488  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49489  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49490  __ret; \
49491})
49492#endif
49493
49494#ifdef __LITTLE_ENDIAN__
49495#define vld1q_u64_x4(__p0) __extension__ ({ \
49496  uint64x2x4_t __ret; \
49497  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49498  __ret; \
49499})
49500#else
49501#define vld1q_u64_x4(__p0) __extension__ ({ \
49502  uint64x2x4_t __ret; \
49503  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49504 \
49505  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49506  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49507  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49508  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49509  __ret; \
49510})
49511#endif
49512
49513#ifdef __LITTLE_ENDIAN__
49514#define vld1q_u16_x4(__p0) __extension__ ({ \
49515  uint16x8x4_t __ret; \
49516  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49517  __ret; \
49518})
49519#else
49520#define vld1q_u16_x4(__p0) __extension__ ({ \
49521  uint16x8x4_t __ret; \
49522  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49523 \
49524  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49525  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49526  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49527  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49528  __ret; \
49529})
49530#endif
49531
49532#ifdef __LITTLE_ENDIAN__
49533#define vld1q_s8_x4(__p0) __extension__ ({ \
49534  int8x16x4_t __ret; \
49535  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49536  __ret; \
49537})
49538#else
49539#define vld1q_s8_x4(__p0) __extension__ ({ \
49540  int8x16x4_t __ret; \
49541  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49542 \
49543  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49544  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49545  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49546  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49547  __ret; \
49548})
49549#endif
49550
49551#ifdef __LITTLE_ENDIAN__
49552#define vld1q_f64_x4(__p0) __extension__ ({ \
49553  float64x2x4_t __ret; \
49554  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49555  __ret; \
49556})
49557#else
49558#define vld1q_f64_x4(__p0) __extension__ ({ \
49559  float64x2x4_t __ret; \
49560  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49561 \
49562  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49563  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49564  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49565  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49566  __ret; \
49567})
49568#endif
49569
49570#ifdef __LITTLE_ENDIAN__
49571#define vld1q_f32_x4(__p0) __extension__ ({ \
49572  float32x4x4_t __ret; \
49573  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49574  __ret; \
49575})
49576#else
49577#define vld1q_f32_x4(__p0) __extension__ ({ \
49578  float32x4x4_t __ret; \
49579  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49580 \
49581  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49582  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49583  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49584  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49585  __ret; \
49586})
49587#endif
49588
49589#ifdef __LITTLE_ENDIAN__
49590#define vld1q_f16_x4(__p0) __extension__ ({ \
49591  float16x8x4_t __ret; \
49592  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49593  __ret; \
49594})
49595#else
49596#define vld1q_f16_x4(__p0) __extension__ ({ \
49597  float16x8x4_t __ret; \
49598  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49599 \
49600  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49601  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49602  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49603  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49604  __ret; \
49605})
49606#endif
49607
49608#ifdef __LITTLE_ENDIAN__
49609#define vld1q_s32_x4(__p0) __extension__ ({ \
49610  int32x4x4_t __ret; \
49611  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49612  __ret; \
49613})
49614#else
49615#define vld1q_s32_x4(__p0) __extension__ ({ \
49616  int32x4x4_t __ret; \
49617  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49618 \
49619  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49620  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49621  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49622  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49623  __ret; \
49624})
49625#endif
49626
49627#ifdef __LITTLE_ENDIAN__
49628#define vld1q_s64_x4(__p0) __extension__ ({ \
49629  int64x2x4_t __ret; \
49630  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49631  __ret; \
49632})
49633#else
49634#define vld1q_s64_x4(__p0) __extension__ ({ \
49635  int64x2x4_t __ret; \
49636  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49637 \
49638  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49639  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49640  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49641  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49642  __ret; \
49643})
49644#endif
49645
49646#ifdef __LITTLE_ENDIAN__
49647#define vld1q_s16_x4(__p0) __extension__ ({ \
49648  int16x8x4_t __ret; \
49649  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49650  __ret; \
49651})
49652#else
49653#define vld1q_s16_x4(__p0) __extension__ ({ \
49654  int16x8x4_t __ret; \
49655  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49656 \
49657  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49658  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49659  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49660  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49661  __ret; \
49662})
49663#endif
49664
49665#ifdef __LITTLE_ENDIAN__
49666#define vld1_u8_x4(__p0) __extension__ ({ \
49667  uint8x8x4_t __ret; \
49668  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49669  __ret; \
49670})
49671#else
49672#define vld1_u8_x4(__p0) __extension__ ({ \
49673  uint8x8x4_t __ret; \
49674  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49675 \
49676  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49677  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49678  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49679  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49680  __ret; \
49681})
49682#endif
49683
49684#ifdef __LITTLE_ENDIAN__
49685#define vld1_u32_x4(__p0) __extension__ ({ \
49686  uint32x2x4_t __ret; \
49687  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49688  __ret; \
49689})
49690#else
49691#define vld1_u32_x4(__p0) __extension__ ({ \
49692  uint32x2x4_t __ret; \
49693  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49694 \
49695  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49696  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49697  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49698  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49699  __ret; \
49700})
49701#endif
49702
49703#ifdef __LITTLE_ENDIAN__
49704#define vld1_u64_x4(__p0) __extension__ ({ \
49705  uint64x1x4_t __ret; \
49706  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49707  __ret; \
49708})
49709#else
49710#define vld1_u64_x4(__p0) __extension__ ({ \
49711  uint64x1x4_t __ret; \
49712  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49713  __ret; \
49714})
49715#endif
49716
49717#ifdef __LITTLE_ENDIAN__
49718#define vld1_u16_x4(__p0) __extension__ ({ \
49719  uint16x4x4_t __ret; \
49720  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49721  __ret; \
49722})
49723#else
49724#define vld1_u16_x4(__p0) __extension__ ({ \
49725  uint16x4x4_t __ret; \
49726  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49727 \
49728  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49729  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49730  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49731  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49732  __ret; \
49733})
49734#endif
49735
49736#ifdef __LITTLE_ENDIAN__
49737#define vld1_s8_x4(__p0) __extension__ ({ \
49738  int8x8x4_t __ret; \
49739  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49740  __ret; \
49741})
49742#else
49743#define vld1_s8_x4(__p0) __extension__ ({ \
49744  int8x8x4_t __ret; \
49745  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49746 \
49747  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49748  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49749  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49750  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49751  __ret; \
49752})
49753#endif
49754
49755#ifdef __LITTLE_ENDIAN__
49756#define vld1_f64_x4(__p0) __extension__ ({ \
49757  float64x1x4_t __ret; \
49758  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49759  __ret; \
49760})
49761#else
49762#define vld1_f64_x4(__p0) __extension__ ({ \
49763  float64x1x4_t __ret; \
49764  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49765  __ret; \
49766})
49767#endif
49768
49769#ifdef __LITTLE_ENDIAN__
49770#define vld1_f32_x4(__p0) __extension__ ({ \
49771  float32x2x4_t __ret; \
49772  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49773  __ret; \
49774})
49775#else
49776#define vld1_f32_x4(__p0) __extension__ ({ \
49777  float32x2x4_t __ret; \
49778  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49779 \
49780  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49781  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49782  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49783  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49784  __ret; \
49785})
49786#endif
49787
49788#ifdef __LITTLE_ENDIAN__
49789#define vld1_f16_x4(__p0) __extension__ ({ \
49790  float16x4x4_t __ret; \
49791  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49792  __ret; \
49793})
49794#else
49795#define vld1_f16_x4(__p0) __extension__ ({ \
49796  float16x4x4_t __ret; \
49797  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49798 \
49799  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49800  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49801  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49802  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49803  __ret; \
49804})
49805#endif
49806
49807#ifdef __LITTLE_ENDIAN__
49808#define vld1_s32_x4(__p0) __extension__ ({ \
49809  int32x2x4_t __ret; \
49810  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49811  __ret; \
49812})
49813#else
49814#define vld1_s32_x4(__p0) __extension__ ({ \
49815  int32x2x4_t __ret; \
49816  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49817 \
49818  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49819  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49820  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49821  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49822  __ret; \
49823})
49824#endif
49825
49826#ifdef __LITTLE_ENDIAN__
49827#define vld1_s64_x4(__p0) __extension__ ({ \
49828  int64x1x4_t __ret; \
49829  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49830  __ret; \
49831})
49832#else
49833#define vld1_s64_x4(__p0) __extension__ ({ \
49834  int64x1x4_t __ret; \
49835  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49836  __ret; \
49837})
49838#endif
49839
49840#ifdef __LITTLE_ENDIAN__
49841#define vld1_s16_x4(__p0) __extension__ ({ \
49842  int16x4x4_t __ret; \
49843  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49844  __ret; \
49845})
49846#else
49847#define vld1_s16_x4(__p0) __extension__ ({ \
49848  int16x4x4_t __ret; \
49849  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49850 \
49851  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49852  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49853  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49854  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49855  __ret; \
49856})
49857#endif
49858
49859#ifdef __LITTLE_ENDIAN__
49860#define vld2_p64(__p0) __extension__ ({ \
49861  poly64x1x2_t __ret; \
49862  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49863  __ret; \
49864})
49865#else
49866#define vld2_p64(__p0) __extension__ ({ \
49867  poly64x1x2_t __ret; \
49868  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49869  __ret; \
49870})
49871#endif
49872
49873#ifdef __LITTLE_ENDIAN__
49874#define vld2q_p64(__p0) __extension__ ({ \
49875  poly64x2x2_t __ret; \
49876  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49877  __ret; \
49878})
49879#else
49880#define vld2q_p64(__p0) __extension__ ({ \
49881  poly64x2x2_t __ret; \
49882  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49883 \
49884  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49885  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49886  __ret; \
49887})
49888#endif
49889
49890#ifdef __LITTLE_ENDIAN__
49891#define vld2q_u64(__p0) __extension__ ({ \
49892  uint64x2x2_t __ret; \
49893  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49894  __ret; \
49895})
49896#else
49897#define vld2q_u64(__p0) __extension__ ({ \
49898  uint64x2x2_t __ret; \
49899  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49900 \
49901  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49902  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49903  __ret; \
49904})
49905#endif
49906
49907#ifdef __LITTLE_ENDIAN__
49908#define vld2q_f64(__p0) __extension__ ({ \
49909  float64x2x2_t __ret; \
49910  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49911  __ret; \
49912})
49913#else
49914#define vld2q_f64(__p0) __extension__ ({ \
49915  float64x2x2_t __ret; \
49916  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49917 \
49918  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49919  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49920  __ret; \
49921})
49922#endif
49923
49924#ifdef __LITTLE_ENDIAN__
49925#define vld2q_s64(__p0) __extension__ ({ \
49926  int64x2x2_t __ret; \
49927  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49928  __ret; \
49929})
49930#else
49931#define vld2q_s64(__p0) __extension__ ({ \
49932  int64x2x2_t __ret; \
49933  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49934 \
49935  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49936  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49937  __ret; \
49938})
49939#endif
49940
49941#ifdef __LITTLE_ENDIAN__
49942#define vld2_f64(__p0) __extension__ ({ \
49943  float64x1x2_t __ret; \
49944  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49945  __ret; \
49946})
49947#else
49948#define vld2_f64(__p0) __extension__ ({ \
49949  float64x1x2_t __ret; \
49950  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49951  __ret; \
49952})
49953#endif
49954
49955#ifdef __LITTLE_ENDIAN__
49956#define vld2_dup_p64(__p0) __extension__ ({ \
49957  poly64x1x2_t __ret; \
49958  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49959  __ret; \
49960})
49961#else
49962#define vld2_dup_p64(__p0) __extension__ ({ \
49963  poly64x1x2_t __ret; \
49964  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49965  __ret; \
49966})
49967#endif
49968
49969#ifdef __LITTLE_ENDIAN__
49970#define vld2q_dup_p8(__p0) __extension__ ({ \
49971  poly8x16x2_t __ret; \
49972  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49973  __ret; \
49974})
49975#else
49976#define vld2q_dup_p8(__p0) __extension__ ({ \
49977  poly8x16x2_t __ret; \
49978  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49979 \
49980  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49981  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49982  __ret; \
49983})
49984#endif
49985
49986#ifdef __LITTLE_ENDIAN__
49987#define vld2q_dup_p64(__p0) __extension__ ({ \
49988  poly64x2x2_t __ret; \
49989  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
49990  __ret; \
49991})
49992#else
49993#define vld2q_dup_p64(__p0) __extension__ ({ \
49994  poly64x2x2_t __ret; \
49995  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
49996 \
49997  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49998  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49999  __ret; \
50000})
50001#endif
50002
50003#ifdef __LITTLE_ENDIAN__
50004#define vld2q_dup_p16(__p0) __extension__ ({ \
50005  poly16x8x2_t __ret; \
50006  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50007  __ret; \
50008})
50009#else
50010#define vld2q_dup_p16(__p0) __extension__ ({ \
50011  poly16x8x2_t __ret; \
50012  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50013 \
50014  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50015  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50016  __ret; \
50017})
50018#endif
50019
50020#ifdef __LITTLE_ENDIAN__
50021#define vld2q_dup_u8(__p0) __extension__ ({ \
50022  uint8x16x2_t __ret; \
50023  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50024  __ret; \
50025})
50026#else
50027#define vld2q_dup_u8(__p0) __extension__ ({ \
50028  uint8x16x2_t __ret; \
50029  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50030 \
50031  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50032  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50033  __ret; \
50034})
50035#endif
50036
50037#ifdef __LITTLE_ENDIAN__
50038#define vld2q_dup_u32(__p0) __extension__ ({ \
50039  uint32x4x2_t __ret; \
50040  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50041  __ret; \
50042})
50043#else
50044#define vld2q_dup_u32(__p0) __extension__ ({ \
50045  uint32x4x2_t __ret; \
50046  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50047 \
50048  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50049  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50050  __ret; \
50051})
50052#endif
50053
50054#ifdef __LITTLE_ENDIAN__
50055#define vld2q_dup_u64(__p0) __extension__ ({ \
50056  uint64x2x2_t __ret; \
50057  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50058  __ret; \
50059})
50060#else
50061#define vld2q_dup_u64(__p0) __extension__ ({ \
50062  uint64x2x2_t __ret; \
50063  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50064 \
50065  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50066  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50067  __ret; \
50068})
50069#endif
50070
50071#ifdef __LITTLE_ENDIAN__
50072#define vld2q_dup_u16(__p0) __extension__ ({ \
50073  uint16x8x2_t __ret; \
50074  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50075  __ret; \
50076})
50077#else
50078#define vld2q_dup_u16(__p0) __extension__ ({ \
50079  uint16x8x2_t __ret; \
50080  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50081 \
50082  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50083  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50084  __ret; \
50085})
50086#endif
50087
50088#ifdef __LITTLE_ENDIAN__
50089#define vld2q_dup_s8(__p0) __extension__ ({ \
50090  int8x16x2_t __ret; \
50091  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50092  __ret; \
50093})
50094#else
50095#define vld2q_dup_s8(__p0) __extension__ ({ \
50096  int8x16x2_t __ret; \
50097  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50098 \
50099  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50100  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50101  __ret; \
50102})
50103#endif
50104
50105#ifdef __LITTLE_ENDIAN__
50106#define vld2q_dup_f64(__p0) __extension__ ({ \
50107  float64x2x2_t __ret; \
50108  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50109  __ret; \
50110})
50111#else
50112#define vld2q_dup_f64(__p0) __extension__ ({ \
50113  float64x2x2_t __ret; \
50114  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50115 \
50116  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50117  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50118  __ret; \
50119})
50120#endif
50121
50122#ifdef __LITTLE_ENDIAN__
50123#define vld2q_dup_f32(__p0) __extension__ ({ \
50124  float32x4x2_t __ret; \
50125  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50126  __ret; \
50127})
50128#else
50129#define vld2q_dup_f32(__p0) __extension__ ({ \
50130  float32x4x2_t __ret; \
50131  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50132 \
50133  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50134  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50135  __ret; \
50136})
50137#endif
50138
50139#ifdef __LITTLE_ENDIAN__
50140#define vld2q_dup_f16(__p0) __extension__ ({ \
50141  float16x8x2_t __ret; \
50142  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50143  __ret; \
50144})
50145#else
50146#define vld2q_dup_f16(__p0) __extension__ ({ \
50147  float16x8x2_t __ret; \
50148  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50149 \
50150  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50151  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50152  __ret; \
50153})
50154#endif
50155
50156#ifdef __LITTLE_ENDIAN__
50157#define vld2q_dup_s32(__p0) __extension__ ({ \
50158  int32x4x2_t __ret; \
50159  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50160  __ret; \
50161})
50162#else
50163#define vld2q_dup_s32(__p0) __extension__ ({ \
50164  int32x4x2_t __ret; \
50165  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50166 \
50167  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50168  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50169  __ret; \
50170})
50171#endif
50172
50173#ifdef __LITTLE_ENDIAN__
50174#define vld2q_dup_s64(__p0) __extension__ ({ \
50175  int64x2x2_t __ret; \
50176  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50177  __ret; \
50178})
50179#else
50180#define vld2q_dup_s64(__p0) __extension__ ({ \
50181  int64x2x2_t __ret; \
50182  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50183 \
50184  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50185  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50186  __ret; \
50187})
50188#endif
50189
50190#ifdef __LITTLE_ENDIAN__
50191#define vld2q_dup_s16(__p0) __extension__ ({ \
50192  int16x8x2_t __ret; \
50193  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50194  __ret; \
50195})
50196#else
50197#define vld2q_dup_s16(__p0) __extension__ ({ \
50198  int16x8x2_t __ret; \
50199  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50200 \
50201  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50202  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50203  __ret; \
50204})
50205#endif
50206
50207#ifdef __LITTLE_ENDIAN__
50208#define vld2_dup_f64(__p0) __extension__ ({ \
50209  float64x1x2_t __ret; \
50210  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50211  __ret; \
50212})
50213#else
50214#define vld2_dup_f64(__p0) __extension__ ({ \
50215  float64x1x2_t __ret; \
50216  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50217  __ret; \
50218})
50219#endif
50220
50221#ifdef __LITTLE_ENDIAN__
50222#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50223  poly64x1x2_t __s1 = __p1; \
50224  poly64x1x2_t __ret; \
50225  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50226  __ret; \
50227})
50228#else
50229#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50230  poly64x1x2_t __s1 = __p1; \
50231  poly64x1x2_t __ret; \
50232  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50233  __ret; \
50234})
50235#endif
50236
50237#ifdef __LITTLE_ENDIAN__
50238#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50239  poly8x16x2_t __s1 = __p1; \
50240  poly8x16x2_t __ret; \
50241  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
50242  __ret; \
50243})
50244#else
50245#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50246  poly8x16x2_t __s1 = __p1; \
50247  poly8x16x2_t __rev1; \
50248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50250  poly8x16x2_t __ret; \
50251  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
50252 \
50253  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50254  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50255  __ret; \
50256})
50257#endif
50258
50259#ifdef __LITTLE_ENDIAN__
50260#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50261  poly64x2x2_t __s1 = __p1; \
50262  poly64x2x2_t __ret; \
50263  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
50264  __ret; \
50265})
50266#else
50267#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50268  poly64x2x2_t __s1 = __p1; \
50269  poly64x2x2_t __rev1; \
50270  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50271  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50272  poly64x2x2_t __ret; \
50273  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
50274 \
50275  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50276  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50277  __ret; \
50278})
50279#endif
50280
50281#ifdef __LITTLE_ENDIAN__
50282#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50283  uint8x16x2_t __s1 = __p1; \
50284  uint8x16x2_t __ret; \
50285  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
50286  __ret; \
50287})
50288#else
50289#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50290  uint8x16x2_t __s1 = __p1; \
50291  uint8x16x2_t __rev1; \
50292  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50293  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50294  uint8x16x2_t __ret; \
50295  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
50296 \
50297  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50298  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50299  __ret; \
50300})
50301#endif
50302
50303#ifdef __LITTLE_ENDIAN__
50304#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50305  uint64x2x2_t __s1 = __p1; \
50306  uint64x2x2_t __ret; \
50307  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
50308  __ret; \
50309})
50310#else
50311#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50312  uint64x2x2_t __s1 = __p1; \
50313  uint64x2x2_t __rev1; \
50314  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50315  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50316  uint64x2x2_t __ret; \
50317  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
50318 \
50319  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50320  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50321  __ret; \
50322})
50323#endif
50324
50325#ifdef __LITTLE_ENDIAN__
50326#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50327  int8x16x2_t __s1 = __p1; \
50328  int8x16x2_t __ret; \
50329  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
50330  __ret; \
50331})
50332#else
50333#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50334  int8x16x2_t __s1 = __p1; \
50335  int8x16x2_t __rev1; \
50336  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50337  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50338  int8x16x2_t __ret; \
50339  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
50340 \
50341  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50342  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50343  __ret; \
50344})
50345#endif
50346
50347#ifdef __LITTLE_ENDIAN__
50348#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50349  float64x2x2_t __s1 = __p1; \
50350  float64x2x2_t __ret; \
50351  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 42); \
50352  __ret; \
50353})
50354#else
50355#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50356  float64x2x2_t __s1 = __p1; \
50357  float64x2x2_t __rev1; \
50358  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50359  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50360  float64x2x2_t __ret; \
50361  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
50362 \
50363  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50364  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50365  __ret; \
50366})
50367#endif
50368
50369#ifdef __LITTLE_ENDIAN__
50370#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50371  int64x2x2_t __s1 = __p1; \
50372  int64x2x2_t __ret; \
50373  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 35); \
50374  __ret; \
50375})
50376#else
50377#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50378  int64x2x2_t __s1 = __p1; \
50379  int64x2x2_t __rev1; \
50380  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50381  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50382  int64x2x2_t __ret; \
50383  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
50384 \
50385  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50386  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50387  __ret; \
50388})
50389#endif
50390
50391#ifdef __LITTLE_ENDIAN__
50392#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50393  uint64x1x2_t __s1 = __p1; \
50394  uint64x1x2_t __ret; \
50395  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50396  __ret; \
50397})
50398#else
50399#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50400  uint64x1x2_t __s1 = __p1; \
50401  uint64x1x2_t __ret; \
50402  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50403  __ret; \
50404})
50405#endif
50406
50407#ifdef __LITTLE_ENDIAN__
50408#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50409  float64x1x2_t __s1 = __p1; \
50410  float64x1x2_t __ret; \
50411  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50412  __ret; \
50413})
50414#else
50415#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50416  float64x1x2_t __s1 = __p1; \
50417  float64x1x2_t __ret; \
50418  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50419  __ret; \
50420})
50421#endif
50422
50423#ifdef __LITTLE_ENDIAN__
50424#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50425  int64x1x2_t __s1 = __p1; \
50426  int64x1x2_t __ret; \
50427  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50428  __ret; \
50429})
50430#else
50431#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50432  int64x1x2_t __s1 = __p1; \
50433  int64x1x2_t __ret; \
50434  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50435  __ret; \
50436})
50437#endif
50438
50439#ifdef __LITTLE_ENDIAN__
50440#define vld3_p64(__p0) __extension__ ({ \
50441  poly64x1x3_t __ret; \
50442  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50443  __ret; \
50444})
50445#else
50446#define vld3_p64(__p0) __extension__ ({ \
50447  poly64x1x3_t __ret; \
50448  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50449  __ret; \
50450})
50451#endif
50452
50453#ifdef __LITTLE_ENDIAN__
50454#define vld3q_p64(__p0) __extension__ ({ \
50455  poly64x2x3_t __ret; \
50456  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50457  __ret; \
50458})
50459#else
50460#define vld3q_p64(__p0) __extension__ ({ \
50461  poly64x2x3_t __ret; \
50462  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50463 \
50464  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50465  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50466  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50467  __ret; \
50468})
50469#endif
50470
50471#ifdef __LITTLE_ENDIAN__
50472#define vld3q_u64(__p0) __extension__ ({ \
50473  uint64x2x3_t __ret; \
50474  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50475  __ret; \
50476})
50477#else
50478#define vld3q_u64(__p0) __extension__ ({ \
50479  uint64x2x3_t __ret; \
50480  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50481 \
50482  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50483  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50484  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50485  __ret; \
50486})
50487#endif
50488
50489#ifdef __LITTLE_ENDIAN__
50490#define vld3q_f64(__p0) __extension__ ({ \
50491  float64x2x3_t __ret; \
50492  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50493  __ret; \
50494})
50495#else
50496#define vld3q_f64(__p0) __extension__ ({ \
50497  float64x2x3_t __ret; \
50498  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50499 \
50500  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50501  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50502  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50503  __ret; \
50504})
50505#endif
50506
50507#ifdef __LITTLE_ENDIAN__
50508#define vld3q_s64(__p0) __extension__ ({ \
50509  int64x2x3_t __ret; \
50510  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50511  __ret; \
50512})
50513#else
50514#define vld3q_s64(__p0) __extension__ ({ \
50515  int64x2x3_t __ret; \
50516  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50517 \
50518  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50519  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50520  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50521  __ret; \
50522})
50523#endif
50524
50525#ifdef __LITTLE_ENDIAN__
50526#define vld3_f64(__p0) __extension__ ({ \
50527  float64x1x3_t __ret; \
50528  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50529  __ret; \
50530})
50531#else
50532#define vld3_f64(__p0) __extension__ ({ \
50533  float64x1x3_t __ret; \
50534  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50535  __ret; \
50536})
50537#endif
50538
50539#ifdef __LITTLE_ENDIAN__
50540#define vld3_dup_p64(__p0) __extension__ ({ \
50541  poly64x1x3_t __ret; \
50542  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50543  __ret; \
50544})
50545#else
50546#define vld3_dup_p64(__p0) __extension__ ({ \
50547  poly64x1x3_t __ret; \
50548  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50549  __ret; \
50550})
50551#endif
50552
50553#ifdef __LITTLE_ENDIAN__
50554#define vld3q_dup_p8(__p0) __extension__ ({ \
50555  poly8x16x3_t __ret; \
50556  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50557  __ret; \
50558})
50559#else
50560#define vld3q_dup_p8(__p0) __extension__ ({ \
50561  poly8x16x3_t __ret; \
50562  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50563 \
50564  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50565  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50566  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50567  __ret; \
50568})
50569#endif
50570
50571#ifdef __LITTLE_ENDIAN__
50572#define vld3q_dup_p64(__p0) __extension__ ({ \
50573  poly64x2x3_t __ret; \
50574  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50575  __ret; \
50576})
50577#else
50578#define vld3q_dup_p64(__p0) __extension__ ({ \
50579  poly64x2x3_t __ret; \
50580  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50581 \
50582  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50583  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50584  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50585  __ret; \
50586})
50587#endif
50588
50589#ifdef __LITTLE_ENDIAN__
50590#define vld3q_dup_p16(__p0) __extension__ ({ \
50591  poly16x8x3_t __ret; \
50592  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50593  __ret; \
50594})
50595#else
50596#define vld3q_dup_p16(__p0) __extension__ ({ \
50597  poly16x8x3_t __ret; \
50598  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50599 \
50600  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50601  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50602  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50603  __ret; \
50604})
50605#endif
50606
50607#ifdef __LITTLE_ENDIAN__
50608#define vld3q_dup_u8(__p0) __extension__ ({ \
50609  uint8x16x3_t __ret; \
50610  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50611  __ret; \
50612})
50613#else
50614#define vld3q_dup_u8(__p0) __extension__ ({ \
50615  uint8x16x3_t __ret; \
50616  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50617 \
50618  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50619  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50620  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50621  __ret; \
50622})
50623#endif
50624
50625#ifdef __LITTLE_ENDIAN__
50626#define vld3q_dup_u32(__p0) __extension__ ({ \
50627  uint32x4x3_t __ret; \
50628  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50629  __ret; \
50630})
50631#else
50632#define vld3q_dup_u32(__p0) __extension__ ({ \
50633  uint32x4x3_t __ret; \
50634  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50635 \
50636  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50637  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50638  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50639  __ret; \
50640})
50641#endif
50642
50643#ifdef __LITTLE_ENDIAN__
50644#define vld3q_dup_u64(__p0) __extension__ ({ \
50645  uint64x2x3_t __ret; \
50646  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50647  __ret; \
50648})
50649#else
50650#define vld3q_dup_u64(__p0) __extension__ ({ \
50651  uint64x2x3_t __ret; \
50652  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50653 \
50654  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50655  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50656  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50657  __ret; \
50658})
50659#endif
50660
50661#ifdef __LITTLE_ENDIAN__
50662#define vld3q_dup_u16(__p0) __extension__ ({ \
50663  uint16x8x3_t __ret; \
50664  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50665  __ret; \
50666})
50667#else
50668#define vld3q_dup_u16(__p0) __extension__ ({ \
50669  uint16x8x3_t __ret; \
50670  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50671 \
50672  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50673  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50674  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50675  __ret; \
50676})
50677#endif
50678
50679#ifdef __LITTLE_ENDIAN__
50680#define vld3q_dup_s8(__p0) __extension__ ({ \
50681  int8x16x3_t __ret; \
50682  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50683  __ret; \
50684})
50685#else
50686#define vld3q_dup_s8(__p0) __extension__ ({ \
50687  int8x16x3_t __ret; \
50688  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50689 \
50690  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50691  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50692  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50693  __ret; \
50694})
50695#endif
50696
50697#ifdef __LITTLE_ENDIAN__
50698#define vld3q_dup_f64(__p0) __extension__ ({ \
50699  float64x2x3_t __ret; \
50700  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50701  __ret; \
50702})
50703#else
50704#define vld3q_dup_f64(__p0) __extension__ ({ \
50705  float64x2x3_t __ret; \
50706  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50707 \
50708  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50709  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50710  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50711  __ret; \
50712})
50713#endif
50714
50715#ifdef __LITTLE_ENDIAN__
50716#define vld3q_dup_f32(__p0) __extension__ ({ \
50717  float32x4x3_t __ret; \
50718  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50719  __ret; \
50720})
50721#else
50722#define vld3q_dup_f32(__p0) __extension__ ({ \
50723  float32x4x3_t __ret; \
50724  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50725 \
50726  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50727  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50728  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50729  __ret; \
50730})
50731#endif
50732
50733#ifdef __LITTLE_ENDIAN__
50734#define vld3q_dup_f16(__p0) __extension__ ({ \
50735  float16x8x3_t __ret; \
50736  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50737  __ret; \
50738})
50739#else
50740#define vld3q_dup_f16(__p0) __extension__ ({ \
50741  float16x8x3_t __ret; \
50742  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50743 \
50744  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50745  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50746  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50747  __ret; \
50748})
50749#endif
50750
50751#ifdef __LITTLE_ENDIAN__
50752#define vld3q_dup_s32(__p0) __extension__ ({ \
50753  int32x4x3_t __ret; \
50754  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50755  __ret; \
50756})
50757#else
50758#define vld3q_dup_s32(__p0) __extension__ ({ \
50759  int32x4x3_t __ret; \
50760  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50761 \
50762  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50763  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50764  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50765  __ret; \
50766})
50767#endif
50768
50769#ifdef __LITTLE_ENDIAN__
50770#define vld3q_dup_s64(__p0) __extension__ ({ \
50771  int64x2x3_t __ret; \
50772  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50773  __ret; \
50774})
50775#else
50776#define vld3q_dup_s64(__p0) __extension__ ({ \
50777  int64x2x3_t __ret; \
50778  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50779 \
50780  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50781  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50782  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50783  __ret; \
50784})
50785#endif
50786
50787#ifdef __LITTLE_ENDIAN__
50788#define vld3q_dup_s16(__p0) __extension__ ({ \
50789  int16x8x3_t __ret; \
50790  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50791  __ret; \
50792})
50793#else
50794#define vld3q_dup_s16(__p0) __extension__ ({ \
50795  int16x8x3_t __ret; \
50796  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50797 \
50798  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50799  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50800  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50801  __ret; \
50802})
50803#endif
50804
50805#ifdef __LITTLE_ENDIAN__
50806#define vld3_dup_f64(__p0) __extension__ ({ \
50807  float64x1x3_t __ret; \
50808  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50809  __ret; \
50810})
50811#else
50812#define vld3_dup_f64(__p0) __extension__ ({ \
50813  float64x1x3_t __ret; \
50814  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50815  __ret; \
50816})
50817#endif
50818
50819#ifdef __LITTLE_ENDIAN__
50820#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50821  poly64x1x3_t __s1 = __p1; \
50822  poly64x1x3_t __ret; \
50823  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50824  __ret; \
50825})
50826#else
50827#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50828  poly64x1x3_t __s1 = __p1; \
50829  poly64x1x3_t __ret; \
50830  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50831  __ret; \
50832})
50833#endif
50834
50835#ifdef __LITTLE_ENDIAN__
50836#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50837  poly8x16x3_t __s1 = __p1; \
50838  poly8x16x3_t __ret; \
50839  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
50840  __ret; \
50841})
50842#else
50843#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50844  poly8x16x3_t __s1 = __p1; \
50845  poly8x16x3_t __rev1; \
50846  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50847  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50848  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50849  poly8x16x3_t __ret; \
50850  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
50851 \
50852  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50853  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50854  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50855  __ret; \
50856})
50857#endif
50858
50859#ifdef __LITTLE_ENDIAN__
50860#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50861  poly64x2x3_t __s1 = __p1; \
50862  poly64x2x3_t __ret; \
50863  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
50864  __ret; \
50865})
50866#else
50867#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50868  poly64x2x3_t __s1 = __p1; \
50869  poly64x2x3_t __rev1; \
50870  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50871  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50872  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50873  poly64x2x3_t __ret; \
50874  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
50875 \
50876  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50877  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50878  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50879  __ret; \
50880})
50881#endif
50882
50883#ifdef __LITTLE_ENDIAN__
50884#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50885  uint8x16x3_t __s1 = __p1; \
50886  uint8x16x3_t __ret; \
50887  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
50888  __ret; \
50889})
50890#else
50891#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50892  uint8x16x3_t __s1 = __p1; \
50893  uint8x16x3_t __rev1; \
50894  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50895  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50896  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50897  uint8x16x3_t __ret; \
50898  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
50899 \
50900  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50901  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50902  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50903  __ret; \
50904})
50905#endif
50906
50907#ifdef __LITTLE_ENDIAN__
50908#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50909  uint64x2x3_t __s1 = __p1; \
50910  uint64x2x3_t __ret; \
50911  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
50912  __ret; \
50913})
50914#else
50915#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50916  uint64x2x3_t __s1 = __p1; \
50917  uint64x2x3_t __rev1; \
50918  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50919  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50920  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50921  uint64x2x3_t __ret; \
50922  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
50923 \
50924  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50925  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50926  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50927  __ret; \
50928})
50929#endif
50930
50931#ifdef __LITTLE_ENDIAN__
50932#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50933  int8x16x3_t __s1 = __p1; \
50934  int8x16x3_t __ret; \
50935  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
50936  __ret; \
50937})
50938#else
50939#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50940  int8x16x3_t __s1 = __p1; \
50941  int8x16x3_t __rev1; \
50942  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50943  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50944  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50945  int8x16x3_t __ret; \
50946  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
50947 \
50948  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50949  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50950  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50951  __ret; \
50952})
50953#endif
50954
50955#ifdef __LITTLE_ENDIAN__
50956#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50957  float64x2x3_t __s1 = __p1; \
50958  float64x2x3_t __ret; \
50959  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
50960  __ret; \
50961})
50962#else
50963#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50964  float64x2x3_t __s1 = __p1; \
50965  float64x2x3_t __rev1; \
50966  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50967  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50968  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50969  float64x2x3_t __ret; \
50970  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
50971 \
50972  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50973  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50974  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50975  __ret; \
50976})
50977#endif
50978
50979#ifdef __LITTLE_ENDIAN__
50980#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50981  int64x2x3_t __s1 = __p1; \
50982  int64x2x3_t __ret; \
50983  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
50984  __ret; \
50985})
50986#else
50987#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50988  int64x2x3_t __s1 = __p1; \
50989  int64x2x3_t __rev1; \
50990  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50991  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50992  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50993  int64x2x3_t __ret; \
50994  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
50995 \
50996  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50997  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50998  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50999  __ret; \
51000})
51001#endif
51002
51003#ifdef __LITTLE_ENDIAN__
51004#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51005  uint64x1x3_t __s1 = __p1; \
51006  uint64x1x3_t __ret; \
51007  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51008  __ret; \
51009})
51010#else
51011#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51012  uint64x1x3_t __s1 = __p1; \
51013  uint64x1x3_t __ret; \
51014  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51015  __ret; \
51016})
51017#endif
51018
51019#ifdef __LITTLE_ENDIAN__
51020#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51021  float64x1x3_t __s1 = __p1; \
51022  float64x1x3_t __ret; \
51023  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51024  __ret; \
51025})
51026#else
51027#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51028  float64x1x3_t __s1 = __p1; \
51029  float64x1x3_t __ret; \
51030  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51031  __ret; \
51032})
51033#endif
51034
51035#ifdef __LITTLE_ENDIAN__
51036#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51037  int64x1x3_t __s1 = __p1; \
51038  int64x1x3_t __ret; \
51039  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51040  __ret; \
51041})
51042#else
51043#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51044  int64x1x3_t __s1 = __p1; \
51045  int64x1x3_t __ret; \
51046  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51047  __ret; \
51048})
51049#endif
51050
51051#ifdef __LITTLE_ENDIAN__
51052#define vld4_p64(__p0) __extension__ ({ \
51053  poly64x1x4_t __ret; \
51054  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51055  __ret; \
51056})
51057#else
51058#define vld4_p64(__p0) __extension__ ({ \
51059  poly64x1x4_t __ret; \
51060  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51061  __ret; \
51062})
51063#endif
51064
51065#ifdef __LITTLE_ENDIAN__
51066#define vld4q_p64(__p0) __extension__ ({ \
51067  poly64x2x4_t __ret; \
51068  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51069  __ret; \
51070})
51071#else
51072#define vld4q_p64(__p0) __extension__ ({ \
51073  poly64x2x4_t __ret; \
51074  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51075 \
51076  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51077  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51078  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51079  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51080  __ret; \
51081})
51082#endif
51083
51084#ifdef __LITTLE_ENDIAN__
51085#define vld4q_u64(__p0) __extension__ ({ \
51086  uint64x2x4_t __ret; \
51087  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51088  __ret; \
51089})
51090#else
51091#define vld4q_u64(__p0) __extension__ ({ \
51092  uint64x2x4_t __ret; \
51093  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51094 \
51095  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51096  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51097  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51098  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51099  __ret; \
51100})
51101#endif
51102
51103#ifdef __LITTLE_ENDIAN__
51104#define vld4q_f64(__p0) __extension__ ({ \
51105  float64x2x4_t __ret; \
51106  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51107  __ret; \
51108})
51109#else
51110#define vld4q_f64(__p0) __extension__ ({ \
51111  float64x2x4_t __ret; \
51112  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51113 \
51114  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51115  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51116  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51117  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51118  __ret; \
51119})
51120#endif
51121
51122#ifdef __LITTLE_ENDIAN__
51123#define vld4q_s64(__p0) __extension__ ({ \
51124  int64x2x4_t __ret; \
51125  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51126  __ret; \
51127})
51128#else
51129#define vld4q_s64(__p0) __extension__ ({ \
51130  int64x2x4_t __ret; \
51131  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51132 \
51133  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51134  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51135  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51136  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51137  __ret; \
51138})
51139#endif
51140
51141#ifdef __LITTLE_ENDIAN__
51142#define vld4_f64(__p0) __extension__ ({ \
51143  float64x1x4_t __ret; \
51144  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51145  __ret; \
51146})
51147#else
51148#define vld4_f64(__p0) __extension__ ({ \
51149  float64x1x4_t __ret; \
51150  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51151  __ret; \
51152})
51153#endif
51154
51155#ifdef __LITTLE_ENDIAN__
51156#define vld4_dup_p64(__p0) __extension__ ({ \
51157  poly64x1x4_t __ret; \
51158  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51159  __ret; \
51160})
51161#else
51162#define vld4_dup_p64(__p0) __extension__ ({ \
51163  poly64x1x4_t __ret; \
51164  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51165  __ret; \
51166})
51167#endif
51168
51169#ifdef __LITTLE_ENDIAN__
51170#define vld4q_dup_p8(__p0) __extension__ ({ \
51171  poly8x16x4_t __ret; \
51172  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51173  __ret; \
51174})
51175#else
51176#define vld4q_dup_p8(__p0) __extension__ ({ \
51177  poly8x16x4_t __ret; \
51178  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51179 \
51180  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51181  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51182  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51183  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51184  __ret; \
51185})
51186#endif
51187
51188#ifdef __LITTLE_ENDIAN__
51189#define vld4q_dup_p64(__p0) __extension__ ({ \
51190  poly64x2x4_t __ret; \
51191  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51192  __ret; \
51193})
51194#else
51195#define vld4q_dup_p64(__p0) __extension__ ({ \
51196  poly64x2x4_t __ret; \
51197  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51198 \
51199  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51200  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51201  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51202  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51203  __ret; \
51204})
51205#endif
51206
51207#ifdef __LITTLE_ENDIAN__
51208#define vld4q_dup_p16(__p0) __extension__ ({ \
51209  poly16x8x4_t __ret; \
51210  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51211  __ret; \
51212})
51213#else
51214#define vld4q_dup_p16(__p0) __extension__ ({ \
51215  poly16x8x4_t __ret; \
51216  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51217 \
51218  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51219  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51220  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51221  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51222  __ret; \
51223})
51224#endif
51225
51226#ifdef __LITTLE_ENDIAN__
51227#define vld4q_dup_u8(__p0) __extension__ ({ \
51228  uint8x16x4_t __ret; \
51229  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51230  __ret; \
51231})
51232#else
51233#define vld4q_dup_u8(__p0) __extension__ ({ \
51234  uint8x16x4_t __ret; \
51235  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51236 \
51237  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51238  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51239  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51240  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51241  __ret; \
51242})
51243#endif
51244
51245#ifdef __LITTLE_ENDIAN__
51246#define vld4q_dup_u32(__p0) __extension__ ({ \
51247  uint32x4x4_t __ret; \
51248  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51249  __ret; \
51250})
51251#else
51252#define vld4q_dup_u32(__p0) __extension__ ({ \
51253  uint32x4x4_t __ret; \
51254  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51255 \
51256  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51257  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51258  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51259  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51260  __ret; \
51261})
51262#endif
51263
51264#ifdef __LITTLE_ENDIAN__
51265#define vld4q_dup_u64(__p0) __extension__ ({ \
51266  uint64x2x4_t __ret; \
51267  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51268  __ret; \
51269})
51270#else
51271#define vld4q_dup_u64(__p0) __extension__ ({ \
51272  uint64x2x4_t __ret; \
51273  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51274 \
51275  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51276  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51277  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51278  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51279  __ret; \
51280})
51281#endif
51282
51283#ifdef __LITTLE_ENDIAN__
51284#define vld4q_dup_u16(__p0) __extension__ ({ \
51285  uint16x8x4_t __ret; \
51286  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51287  __ret; \
51288})
51289#else
51290#define vld4q_dup_u16(__p0) __extension__ ({ \
51291  uint16x8x4_t __ret; \
51292  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51293 \
51294  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51295  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51296  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51297  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51298  __ret; \
51299})
51300#endif
51301
51302#ifdef __LITTLE_ENDIAN__
51303#define vld4q_dup_s8(__p0) __extension__ ({ \
51304  int8x16x4_t __ret; \
51305  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51306  __ret; \
51307})
51308#else
51309#define vld4q_dup_s8(__p0) __extension__ ({ \
51310  int8x16x4_t __ret; \
51311  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51312 \
51313  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51314  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51315  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51316  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51317  __ret; \
51318})
51319#endif
51320
51321#ifdef __LITTLE_ENDIAN__
51322#define vld4q_dup_f64(__p0) __extension__ ({ \
51323  float64x2x4_t __ret; \
51324  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51325  __ret; \
51326})
51327#else
51328#define vld4q_dup_f64(__p0) __extension__ ({ \
51329  float64x2x4_t __ret; \
51330  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51331 \
51332  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51333  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51334  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51335  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51336  __ret; \
51337})
51338#endif
51339
51340#ifdef __LITTLE_ENDIAN__
51341#define vld4q_dup_f32(__p0) __extension__ ({ \
51342  float32x4x4_t __ret; \
51343  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51344  __ret; \
51345})
51346#else
51347#define vld4q_dup_f32(__p0) __extension__ ({ \
51348  float32x4x4_t __ret; \
51349  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51350 \
51351  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51352  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51353  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51354  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51355  __ret; \
51356})
51357#endif
51358
51359#ifdef __LITTLE_ENDIAN__
51360#define vld4q_dup_f16(__p0) __extension__ ({ \
51361  float16x8x4_t __ret; \
51362  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51363  __ret; \
51364})
51365#else
51366#define vld4q_dup_f16(__p0) __extension__ ({ \
51367  float16x8x4_t __ret; \
51368  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51369 \
51370  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51371  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51372  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51373  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51374  __ret; \
51375})
51376#endif
51377
51378#ifdef __LITTLE_ENDIAN__
51379#define vld4q_dup_s32(__p0) __extension__ ({ \
51380  int32x4x4_t __ret; \
51381  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51382  __ret; \
51383})
51384#else
51385#define vld4q_dup_s32(__p0) __extension__ ({ \
51386  int32x4x4_t __ret; \
51387  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51388 \
51389  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51390  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51391  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51392  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51393  __ret; \
51394})
51395#endif
51396
51397#ifdef __LITTLE_ENDIAN__
51398#define vld4q_dup_s64(__p0) __extension__ ({ \
51399  int64x2x4_t __ret; \
51400  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51401  __ret; \
51402})
51403#else
51404#define vld4q_dup_s64(__p0) __extension__ ({ \
51405  int64x2x4_t __ret; \
51406  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51407 \
51408  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51409  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51410  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51411  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51412  __ret; \
51413})
51414#endif
51415
51416#ifdef __LITTLE_ENDIAN__
51417#define vld4q_dup_s16(__p0) __extension__ ({ \
51418  int16x8x4_t __ret; \
51419  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51420  __ret; \
51421})
51422#else
51423#define vld4q_dup_s16(__p0) __extension__ ({ \
51424  int16x8x4_t __ret; \
51425  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51426 \
51427  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51428  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51429  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51430  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51431  __ret; \
51432})
51433#endif
51434
51435#ifdef __LITTLE_ENDIAN__
51436#define vld4_dup_f64(__p0) __extension__ ({ \
51437  float64x1x4_t __ret; \
51438  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51439  __ret; \
51440})
51441#else
51442#define vld4_dup_f64(__p0) __extension__ ({ \
51443  float64x1x4_t __ret; \
51444  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51445  __ret; \
51446})
51447#endif
51448
51449#ifdef __LITTLE_ENDIAN__
51450#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51451  poly64x1x4_t __s1 = __p1; \
51452  poly64x1x4_t __ret; \
51453  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51454  __ret; \
51455})
51456#else
51457#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51458  poly64x1x4_t __s1 = __p1; \
51459  poly64x1x4_t __ret; \
51460  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51461  __ret; \
51462})
51463#endif
51464
51465#ifdef __LITTLE_ENDIAN__
51466#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51467  poly8x16x4_t __s1 = __p1; \
51468  poly8x16x4_t __ret; \
51469  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
51470  __ret; \
51471})
51472#else
51473#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51474  poly8x16x4_t __s1 = __p1; \
51475  poly8x16x4_t __rev1; \
51476  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51477  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51478  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51479  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51480  poly8x16x4_t __ret; \
51481  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
51482 \
51483  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51484  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51485  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51486  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51487  __ret; \
51488})
51489#endif
51490
51491#ifdef __LITTLE_ENDIAN__
51492#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51493  poly64x2x4_t __s1 = __p1; \
51494  poly64x2x4_t __ret; \
51495  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
51496  __ret; \
51497})
51498#else
51499#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51500  poly64x2x4_t __s1 = __p1; \
51501  poly64x2x4_t __rev1; \
51502  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51503  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51504  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51505  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51506  poly64x2x4_t __ret; \
51507  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
51508 \
51509  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51510  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51511  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51512  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51513  __ret; \
51514})
51515#endif
51516
51517#ifdef __LITTLE_ENDIAN__
51518#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51519  uint8x16x4_t __s1 = __p1; \
51520  uint8x16x4_t __ret; \
51521  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
51522  __ret; \
51523})
51524#else
51525#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51526  uint8x16x4_t __s1 = __p1; \
51527  uint8x16x4_t __rev1; \
51528  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51529  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51530  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51531  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51532  uint8x16x4_t __ret; \
51533  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
51534 \
51535  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51536  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51537  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51538  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51539  __ret; \
51540})
51541#endif
51542
51543#ifdef __LITTLE_ENDIAN__
51544#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51545  uint64x2x4_t __s1 = __p1; \
51546  uint64x2x4_t __ret; \
51547  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
51548  __ret; \
51549})
51550#else
51551#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51552  uint64x2x4_t __s1 = __p1; \
51553  uint64x2x4_t __rev1; \
51554  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51555  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51556  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51557  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51558  uint64x2x4_t __ret; \
51559  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
51560 \
51561  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51562  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51563  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51564  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51565  __ret; \
51566})
51567#endif
51568
51569#ifdef __LITTLE_ENDIAN__
51570#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51571  int8x16x4_t __s1 = __p1; \
51572  int8x16x4_t __ret; \
51573  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
51574  __ret; \
51575})
51576#else
51577#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51578  int8x16x4_t __s1 = __p1; \
51579  int8x16x4_t __rev1; \
51580  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51581  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51582  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51583  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51584  int8x16x4_t __ret; \
51585  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
51586 \
51587  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51588  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51589  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51590  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51591  __ret; \
51592})
51593#endif
51594
51595#ifdef __LITTLE_ENDIAN__
51596#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51597  float64x2x4_t __s1 = __p1; \
51598  float64x2x4_t __ret; \
51599  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
51600  __ret; \
51601})
51602#else
51603#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51604  float64x2x4_t __s1 = __p1; \
51605  float64x2x4_t __rev1; \
51606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51608  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51609  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51610  float64x2x4_t __ret; \
51611  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
51612 \
51613  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51614  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51615  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51616  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51617  __ret; \
51618})
51619#endif
51620
51621#ifdef __LITTLE_ENDIAN__
51622#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51623  int64x2x4_t __s1 = __p1; \
51624  int64x2x4_t __ret; \
51625  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
51626  __ret; \
51627})
51628#else
51629#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51630  int64x2x4_t __s1 = __p1; \
51631  int64x2x4_t __rev1; \
51632  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51633  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51634  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51635  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51636  int64x2x4_t __ret; \
51637  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
51638 \
51639  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51640  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51641  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51642  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51643  __ret; \
51644})
51645#endif
51646
51647#ifdef __LITTLE_ENDIAN__
51648#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51649  uint64x1x4_t __s1 = __p1; \
51650  uint64x1x4_t __ret; \
51651  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51652  __ret; \
51653})
51654#else
51655#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51656  uint64x1x4_t __s1 = __p1; \
51657  uint64x1x4_t __ret; \
51658  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51659  __ret; \
51660})
51661#endif
51662
51663#ifdef __LITTLE_ENDIAN__
51664#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51665  float64x1x4_t __s1 = __p1; \
51666  float64x1x4_t __ret; \
51667  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51668  __ret; \
51669})
51670#else
51671#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51672  float64x1x4_t __s1 = __p1; \
51673  float64x1x4_t __ret; \
51674  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51675  __ret; \
51676})
51677#endif
51678
51679#ifdef __LITTLE_ENDIAN__
51680#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51681  int64x1x4_t __s1 = __p1; \
51682  int64x1x4_t __ret; \
51683  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51684  __ret; \
51685})
51686#else
51687#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51688  int64x1x4_t __s1 = __p1; \
51689  int64x1x4_t __ret; \
51690  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51691  __ret; \
51692})
51693#endif
51694
51695#ifdef __LITTLE_ENDIAN__
51696#define vldrq_p128(__p0) __extension__ ({ \
51697  poly128_t __ret; \
51698  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51699  __ret; \
51700})
51701#else
51702#define vldrq_p128(__p0) __extension__ ({ \
51703  poly128_t __ret; \
51704  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51705  __ret; \
51706})
51707#endif
51708
51709#ifdef __LITTLE_ENDIAN__
51710__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51711  float64x2_t __ret;
51712  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51713  return __ret;
51714}
51715#else
51716__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51717  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51718  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51719  float64x2_t __ret;
51720  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51722  return __ret;
51723}
51724#endif
51725
51726#ifdef __LITTLE_ENDIAN__
51727__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51728  float64x1_t __ret;
51729  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51730  return __ret;
51731}
51732#else
51733__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51734  float64x1_t __ret;
51735  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51736  return __ret;
51737}
51738#endif
51739
51740#ifdef __LITTLE_ENDIAN__
51741__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51742  float64_t __ret;
51743  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__p0);
51744  return __ret;
51745}
51746#else
51747__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51748  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51749  float64_t __ret;
51750  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__rev0);
51751  return __ret;
51752}
51753#endif
51754
51755#ifdef __LITTLE_ENDIAN__
51756__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51757  float32_t __ret;
51758  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__p0);
51759  return __ret;
51760}
51761#else
51762__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51763  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51764  float32_t __ret;
51765  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__rev0);
51766  return __ret;
51767}
51768#endif
51769
51770#ifdef __LITTLE_ENDIAN__
51771__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51772  float32_t __ret;
51773  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__p0);
51774  return __ret;
51775}
51776#else
51777__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51778  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51779  float32_t __ret;
51780  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__rev0);
51781  return __ret;
51782}
51783#endif
51784
51785#ifdef __LITTLE_ENDIAN__
51786__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51787  uint8_t __ret;
51788  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__p0);
51789  return __ret;
51790}
51791#else
51792__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51793  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51794  uint8_t __ret;
51795  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__rev0);
51796  return __ret;
51797}
51798#endif
51799
51800#ifdef __LITTLE_ENDIAN__
51801__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51802  uint32_t __ret;
51803  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__p0);
51804  return __ret;
51805}
51806#else
51807__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51808  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51809  uint32_t __ret;
51810  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__rev0);
51811  return __ret;
51812}
51813#endif
51814
51815#ifdef __LITTLE_ENDIAN__
51816__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51817  uint16_t __ret;
51818  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__p0);
51819  return __ret;
51820}
51821#else
51822__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51823  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51824  uint16_t __ret;
51825  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__rev0);
51826  return __ret;
51827}
51828#endif
51829
51830#ifdef __LITTLE_ENDIAN__
51831__ai int8_t vmaxvq_s8(int8x16_t __p0) {
51832  int8_t __ret;
51833  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__p0);
51834  return __ret;
51835}
51836#else
51837__ai int8_t vmaxvq_s8(int8x16_t __p0) {
51838  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51839  int8_t __ret;
51840  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__rev0);
51841  return __ret;
51842}
51843#endif
51844
51845#ifdef __LITTLE_ENDIAN__
51846__ai float64_t vmaxvq_f64(float64x2_t __p0) {
51847  float64_t __ret;
51848  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__p0);
51849  return __ret;
51850}
51851#else
51852__ai float64_t vmaxvq_f64(float64x2_t __p0) {
51853  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51854  float64_t __ret;
51855  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__rev0);
51856  return __ret;
51857}
51858#endif
51859
51860#ifdef __LITTLE_ENDIAN__
51861__ai float32_t vmaxvq_f32(float32x4_t __p0) {
51862  float32_t __ret;
51863  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__p0);
51864  return __ret;
51865}
51866#else
51867__ai float32_t vmaxvq_f32(float32x4_t __p0) {
51868  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51869  float32_t __ret;
51870  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__rev0);
51871  return __ret;
51872}
51873#endif
51874
51875#ifdef __LITTLE_ENDIAN__
51876__ai int32_t vmaxvq_s32(int32x4_t __p0) {
51877  int32_t __ret;
51878  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__p0);
51879  return __ret;
51880}
51881#else
51882__ai int32_t vmaxvq_s32(int32x4_t __p0) {
51883  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51884  int32_t __ret;
51885  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__rev0);
51886  return __ret;
51887}
51888#endif
51889
51890#ifdef __LITTLE_ENDIAN__
51891__ai int16_t vmaxvq_s16(int16x8_t __p0) {
51892  int16_t __ret;
51893  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__p0);
51894  return __ret;
51895}
51896#else
51897__ai int16_t vmaxvq_s16(int16x8_t __p0) {
51898  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51899  int16_t __ret;
51900  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__rev0);
51901  return __ret;
51902}
51903#endif
51904
51905#ifdef __LITTLE_ENDIAN__
51906__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51907  uint8_t __ret;
51908  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__p0);
51909  return __ret;
51910}
51911#else
51912__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51913  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51914  uint8_t __ret;
51915  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__rev0);
51916  return __ret;
51917}
51918#endif
51919
51920#ifdef __LITTLE_ENDIAN__
51921__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51922  uint32_t __ret;
51923  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__p0);
51924  return __ret;
51925}
51926#else
51927__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51928  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51929  uint32_t __ret;
51930  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__rev0);
51931  return __ret;
51932}
51933#endif
51934
51935#ifdef __LITTLE_ENDIAN__
51936__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51937  uint16_t __ret;
51938  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__p0);
51939  return __ret;
51940}
51941#else
51942__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51943  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51944  uint16_t __ret;
51945  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__rev0);
51946  return __ret;
51947}
51948#endif
51949
51950#ifdef __LITTLE_ENDIAN__
51951__ai int8_t vmaxv_s8(int8x8_t __p0) {
51952  int8_t __ret;
51953  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__p0);
51954  return __ret;
51955}
51956#else
51957__ai int8_t vmaxv_s8(int8x8_t __p0) {
51958  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51959  int8_t __ret;
51960  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__rev0);
51961  return __ret;
51962}
51963#endif
51964
51965#ifdef __LITTLE_ENDIAN__
51966__ai float32_t vmaxv_f32(float32x2_t __p0) {
51967  float32_t __ret;
51968  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__p0);
51969  return __ret;
51970}
51971#else
51972__ai float32_t vmaxv_f32(float32x2_t __p0) {
51973  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51974  float32_t __ret;
51975  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__rev0);
51976  return __ret;
51977}
51978#endif
51979
51980#ifdef __LITTLE_ENDIAN__
51981__ai int32_t vmaxv_s32(int32x2_t __p0) {
51982  int32_t __ret;
51983  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__p0);
51984  return __ret;
51985}
51986#else
51987__ai int32_t vmaxv_s32(int32x2_t __p0) {
51988  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51989  int32_t __ret;
51990  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__rev0);
51991  return __ret;
51992}
51993#endif
51994
51995#ifdef __LITTLE_ENDIAN__
51996__ai int16_t vmaxv_s16(int16x4_t __p0) {
51997  int16_t __ret;
51998  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__p0);
51999  return __ret;
52000}
52001#else
52002__ai int16_t vmaxv_s16(int16x4_t __p0) {
52003  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52004  int16_t __ret;
52005  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__rev0);
52006  return __ret;
52007}
52008#endif
52009
52010#ifdef __LITTLE_ENDIAN__
52011__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52012  float64x2_t __ret;
52013  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52014  return __ret;
52015}
52016#else
52017__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52018  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52019  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52020  float64x2_t __ret;
52021  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52023  return __ret;
52024}
52025#endif
52026
52027#ifdef __LITTLE_ENDIAN__
52028__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52029  float64x1_t __ret;
52030  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52031  return __ret;
52032}
52033#else
52034__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52035  float64x1_t __ret;
52036  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52037  return __ret;
52038}
52039#endif
52040
52041#ifdef __LITTLE_ENDIAN__
52042__ai float64_t vminnmvq_f64(float64x2_t __p0) {
52043  float64_t __ret;
52044  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__p0);
52045  return __ret;
52046}
52047#else
52048__ai float64_t vminnmvq_f64(float64x2_t __p0) {
52049  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52050  float64_t __ret;
52051  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__rev0);
52052  return __ret;
52053}
52054#endif
52055
52056#ifdef __LITTLE_ENDIAN__
52057__ai float32_t vminnmvq_f32(float32x4_t __p0) {
52058  float32_t __ret;
52059  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__p0);
52060  return __ret;
52061}
52062#else
52063__ai float32_t vminnmvq_f32(float32x4_t __p0) {
52064  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52065  float32_t __ret;
52066  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__rev0);
52067  return __ret;
52068}
52069#endif
52070
52071#ifdef __LITTLE_ENDIAN__
52072__ai float32_t vminnmv_f32(float32x2_t __p0) {
52073  float32_t __ret;
52074  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__p0);
52075  return __ret;
52076}
52077#else
52078__ai float32_t vminnmv_f32(float32x2_t __p0) {
52079  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52080  float32_t __ret;
52081  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__rev0);
52082  return __ret;
52083}
52084#endif
52085
52086#ifdef __LITTLE_ENDIAN__
52087__ai uint8_t vminvq_u8(uint8x16_t __p0) {
52088  uint8_t __ret;
52089  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__p0);
52090  return __ret;
52091}
52092#else
52093__ai uint8_t vminvq_u8(uint8x16_t __p0) {
52094  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52095  uint8_t __ret;
52096  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__rev0);
52097  return __ret;
52098}
52099#endif
52100
52101#ifdef __LITTLE_ENDIAN__
52102__ai uint32_t vminvq_u32(uint32x4_t __p0) {
52103  uint32_t __ret;
52104  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__p0);
52105  return __ret;
52106}
52107#else
52108__ai uint32_t vminvq_u32(uint32x4_t __p0) {
52109  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52110  uint32_t __ret;
52111  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__rev0);
52112  return __ret;
52113}
52114#endif
52115
52116#ifdef __LITTLE_ENDIAN__
52117__ai uint16_t vminvq_u16(uint16x8_t __p0) {
52118  uint16_t __ret;
52119  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__p0);
52120  return __ret;
52121}
52122#else
52123__ai uint16_t vminvq_u16(uint16x8_t __p0) {
52124  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52125  uint16_t __ret;
52126  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__rev0);
52127  return __ret;
52128}
52129#endif
52130
52131#ifdef __LITTLE_ENDIAN__
52132__ai int8_t vminvq_s8(int8x16_t __p0) {
52133  int8_t __ret;
52134  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__p0);
52135  return __ret;
52136}
52137#else
52138__ai int8_t vminvq_s8(int8x16_t __p0) {
52139  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52140  int8_t __ret;
52141  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__rev0);
52142  return __ret;
52143}
52144#endif
52145
52146#ifdef __LITTLE_ENDIAN__
52147__ai float64_t vminvq_f64(float64x2_t __p0) {
52148  float64_t __ret;
52149  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__p0);
52150  return __ret;
52151}
52152#else
52153__ai float64_t vminvq_f64(float64x2_t __p0) {
52154  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52155  float64_t __ret;
52156  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__rev0);
52157  return __ret;
52158}
52159#endif
52160
52161#ifdef __LITTLE_ENDIAN__
52162__ai float32_t vminvq_f32(float32x4_t __p0) {
52163  float32_t __ret;
52164  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__p0);
52165  return __ret;
52166}
52167#else
52168__ai float32_t vminvq_f32(float32x4_t __p0) {
52169  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52170  float32_t __ret;
52171  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__rev0);
52172  return __ret;
52173}
52174#endif
52175
52176#ifdef __LITTLE_ENDIAN__
52177__ai int32_t vminvq_s32(int32x4_t __p0) {
52178  int32_t __ret;
52179  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__p0);
52180  return __ret;
52181}
52182#else
52183__ai int32_t vminvq_s32(int32x4_t __p0) {
52184  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52185  int32_t __ret;
52186  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__rev0);
52187  return __ret;
52188}
52189#endif
52190
52191#ifdef __LITTLE_ENDIAN__
52192__ai int16_t vminvq_s16(int16x8_t __p0) {
52193  int16_t __ret;
52194  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__p0);
52195  return __ret;
52196}
52197#else
52198__ai int16_t vminvq_s16(int16x8_t __p0) {
52199  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52200  int16_t __ret;
52201  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__rev0);
52202  return __ret;
52203}
52204#endif
52205
52206#ifdef __LITTLE_ENDIAN__
52207__ai uint8_t vminv_u8(uint8x8_t __p0) {
52208  uint8_t __ret;
52209  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__p0);
52210  return __ret;
52211}
52212#else
52213__ai uint8_t vminv_u8(uint8x8_t __p0) {
52214  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52215  uint8_t __ret;
52216  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__rev0);
52217  return __ret;
52218}
52219#endif
52220
52221#ifdef __LITTLE_ENDIAN__
52222__ai uint32_t vminv_u32(uint32x2_t __p0) {
52223  uint32_t __ret;
52224  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__p0);
52225  return __ret;
52226}
52227#else
52228__ai uint32_t vminv_u32(uint32x2_t __p0) {
52229  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52230  uint32_t __ret;
52231  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__rev0);
52232  return __ret;
52233}
52234#endif
52235
52236#ifdef __LITTLE_ENDIAN__
52237__ai uint16_t vminv_u16(uint16x4_t __p0) {
52238  uint16_t __ret;
52239  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__p0);
52240  return __ret;
52241}
52242#else
52243__ai uint16_t vminv_u16(uint16x4_t __p0) {
52244  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52245  uint16_t __ret;
52246  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__rev0);
52247  return __ret;
52248}
52249#endif
52250
52251#ifdef __LITTLE_ENDIAN__
52252__ai int8_t vminv_s8(int8x8_t __p0) {
52253  int8_t __ret;
52254  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__p0);
52255  return __ret;
52256}
52257#else
52258__ai int8_t vminv_s8(int8x8_t __p0) {
52259  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52260  int8_t __ret;
52261  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__rev0);
52262  return __ret;
52263}
52264#endif
52265
52266#ifdef __LITTLE_ENDIAN__
52267__ai float32_t vminv_f32(float32x2_t __p0) {
52268  float32_t __ret;
52269  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__p0);
52270  return __ret;
52271}
52272#else
52273__ai float32_t vminv_f32(float32x2_t __p0) {
52274  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52275  float32_t __ret;
52276  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__rev0);
52277  return __ret;
52278}
52279#endif
52280
52281#ifdef __LITTLE_ENDIAN__
52282__ai int32_t vminv_s32(int32x2_t __p0) {
52283  int32_t __ret;
52284  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__p0);
52285  return __ret;
52286}
52287#else
52288__ai int32_t vminv_s32(int32x2_t __p0) {
52289  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52290  int32_t __ret;
52291  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__rev0);
52292  return __ret;
52293}
52294#endif
52295
52296#ifdef __LITTLE_ENDIAN__
52297__ai int16_t vminv_s16(int16x4_t __p0) {
52298  int16_t __ret;
52299  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__p0);
52300  return __ret;
52301}
52302#else
52303__ai int16_t vminv_s16(int16x4_t __p0) {
52304  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52305  int16_t __ret;
52306  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__rev0);
52307  return __ret;
52308}
52309#endif
52310
52311#ifdef __LITTLE_ENDIAN__
52312__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52313  float64x2_t __ret;
52314  __ret = __p0 + __p1 * __p2;
52315  return __ret;
52316}
52317#else
52318__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52319  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52320  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52321  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52322  float64x2_t __ret;
52323  __ret = __rev0 + __rev1 * __rev2;
52324  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52325  return __ret;
52326}
52327#endif
52328
52329#ifdef __LITTLE_ENDIAN__
52330__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52331  float64x1_t __ret;
52332  __ret = __p0 + __p1 * __p2;
52333  return __ret;
52334}
52335#else
52336__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52337  float64x1_t __ret;
52338  __ret = __p0 + __p1 * __p2;
52339  return __ret;
52340}
52341#endif
52342
52343#ifdef __LITTLE_ENDIAN__
52344#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52345  uint32x4_t __s0 = __p0; \
52346  uint32x4_t __s1 = __p1; \
52347  uint32x4_t __s2 = __p2; \
52348  uint32x4_t __ret; \
52349  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52350  __ret; \
52351})
52352#else
52353#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52354  uint32x4_t __s0 = __p0; \
52355  uint32x4_t __s1 = __p1; \
52356  uint32x4_t __s2 = __p2; \
52357  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52358  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52359  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52360  uint32x4_t __ret; \
52361  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52362  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52363  __ret; \
52364})
52365#endif
52366
52367#ifdef __LITTLE_ENDIAN__
52368#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52369  uint16x8_t __s0 = __p0; \
52370  uint16x8_t __s1 = __p1; \
52371  uint16x8_t __s2 = __p2; \
52372  uint16x8_t __ret; \
52373  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52374  __ret; \
52375})
52376#else
52377#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52378  uint16x8_t __s0 = __p0; \
52379  uint16x8_t __s1 = __p1; \
52380  uint16x8_t __s2 = __p2; \
52381  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52382  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52383  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52384  uint16x8_t __ret; \
52385  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52386  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52387  __ret; \
52388})
52389#endif
52390
52391#ifdef __LITTLE_ENDIAN__
52392#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52393  float32x4_t __s0 = __p0; \
52394  float32x4_t __s1 = __p1; \
52395  float32x4_t __s2 = __p2; \
52396  float32x4_t __ret; \
52397  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52398  __ret; \
52399})
52400#else
52401#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52402  float32x4_t __s0 = __p0; \
52403  float32x4_t __s1 = __p1; \
52404  float32x4_t __s2 = __p2; \
52405  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52406  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52407  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52408  float32x4_t __ret; \
52409  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52411  __ret; \
52412})
52413#endif
52414
52415#ifdef __LITTLE_ENDIAN__
52416#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52417  int32x4_t __s0 = __p0; \
52418  int32x4_t __s1 = __p1; \
52419  int32x4_t __s2 = __p2; \
52420  int32x4_t __ret; \
52421  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52422  __ret; \
52423})
52424#else
52425#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52426  int32x4_t __s0 = __p0; \
52427  int32x4_t __s1 = __p1; \
52428  int32x4_t __s2 = __p2; \
52429  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52430  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52431  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52432  int32x4_t __ret; \
52433  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52434  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52435  __ret; \
52436})
52437#endif
52438
52439#ifdef __LITTLE_ENDIAN__
52440#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52441  int16x8_t __s0 = __p0; \
52442  int16x8_t __s1 = __p1; \
52443  int16x8_t __s2 = __p2; \
52444  int16x8_t __ret; \
52445  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52446  __ret; \
52447})
52448#else
52449#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52450  int16x8_t __s0 = __p0; \
52451  int16x8_t __s1 = __p1; \
52452  int16x8_t __s2 = __p2; \
52453  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52454  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52455  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52456  int16x8_t __ret; \
52457  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52458  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52459  __ret; \
52460})
52461#endif
52462
52463#ifdef __LITTLE_ENDIAN__
52464#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52465  uint32x2_t __s0 = __p0; \
52466  uint32x2_t __s1 = __p1; \
52467  uint32x4_t __s2 = __p2; \
52468  uint32x2_t __ret; \
52469  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52470  __ret; \
52471})
52472#else
52473#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52474  uint32x2_t __s0 = __p0; \
52475  uint32x2_t __s1 = __p1; \
52476  uint32x4_t __s2 = __p2; \
52477  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52478  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52479  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52480  uint32x2_t __ret; \
52481  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52482  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52483  __ret; \
52484})
52485#endif
52486
52487#ifdef __LITTLE_ENDIAN__
52488#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52489  uint16x4_t __s0 = __p0; \
52490  uint16x4_t __s1 = __p1; \
52491  uint16x8_t __s2 = __p2; \
52492  uint16x4_t __ret; \
52493  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52494  __ret; \
52495})
52496#else
52497#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52498  uint16x4_t __s0 = __p0; \
52499  uint16x4_t __s1 = __p1; \
52500  uint16x8_t __s2 = __p2; \
52501  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52502  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52503  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52504  uint16x4_t __ret; \
52505  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52506  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52507  __ret; \
52508})
52509#endif
52510
52511#ifdef __LITTLE_ENDIAN__
52512#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52513  float32x2_t __s0 = __p0; \
52514  float32x2_t __s1 = __p1; \
52515  float32x4_t __s2 = __p2; \
52516  float32x2_t __ret; \
52517  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52518  __ret; \
52519})
52520#else
52521#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52522  float32x2_t __s0 = __p0; \
52523  float32x2_t __s1 = __p1; \
52524  float32x4_t __s2 = __p2; \
52525  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52526  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52527  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52528  float32x2_t __ret; \
52529  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52530  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52531  __ret; \
52532})
52533#endif
52534
52535#ifdef __LITTLE_ENDIAN__
52536#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52537  int32x2_t __s0 = __p0; \
52538  int32x2_t __s1 = __p1; \
52539  int32x4_t __s2 = __p2; \
52540  int32x2_t __ret; \
52541  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52542  __ret; \
52543})
52544#else
52545#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52546  int32x2_t __s0 = __p0; \
52547  int32x2_t __s1 = __p1; \
52548  int32x4_t __s2 = __p2; \
52549  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52550  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52551  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52552  int32x2_t __ret; \
52553  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52554  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52555  __ret; \
52556})
52557#endif
52558
52559#ifdef __LITTLE_ENDIAN__
52560#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52561  int16x4_t __s0 = __p0; \
52562  int16x4_t __s1 = __p1; \
52563  int16x8_t __s2 = __p2; \
52564  int16x4_t __ret; \
52565  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52566  __ret; \
52567})
52568#else
52569#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52570  int16x4_t __s0 = __p0; \
52571  int16x4_t __s1 = __p1; \
52572  int16x8_t __s2 = __p2; \
52573  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52574  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52575  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52576  int16x4_t __ret; \
52577  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52578  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52579  __ret; \
52580})
52581#endif
52582
52583#ifdef __LITTLE_ENDIAN__
52584__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52585  float64x2_t __ret;
52586  __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
52587  return __ret;
52588}
52589#else
52590__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52591  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52592  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52593  float64x2_t __ret;
52594  __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
52595  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52596  return __ret;
52597}
52598#endif
52599
52600#ifdef __LITTLE_ENDIAN__
52601#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52602  uint64x2_t __s0 = __p0; \
52603  uint32x4_t __s1 = __p1; \
52604  uint32x2_t __s2 = __p2; \
52605  uint64x2_t __ret; \
52606  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52607  __ret; \
52608})
52609#else
52610#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52611  uint64x2_t __s0 = __p0; \
52612  uint32x4_t __s1 = __p1; \
52613  uint32x2_t __s2 = __p2; \
52614  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52615  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52616  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52617  uint64x2_t __ret; \
52618  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52619  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52620  __ret; \
52621})
52622#endif
52623
52624#ifdef __LITTLE_ENDIAN__
52625#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52626  uint32x4_t __s0 = __p0; \
52627  uint16x8_t __s1 = __p1; \
52628  uint16x4_t __s2 = __p2; \
52629  uint32x4_t __ret; \
52630  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52631  __ret; \
52632})
52633#else
52634#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52635  uint32x4_t __s0 = __p0; \
52636  uint16x8_t __s1 = __p1; \
52637  uint16x4_t __s2 = __p2; \
52638  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52639  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52640  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52641  uint32x4_t __ret; \
52642  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52643  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52644  __ret; \
52645})
52646#endif
52647
52648#ifdef __LITTLE_ENDIAN__
52649#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52650  int64x2_t __s0 = __p0; \
52651  int32x4_t __s1 = __p1; \
52652  int32x2_t __s2 = __p2; \
52653  int64x2_t __ret; \
52654  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52655  __ret; \
52656})
52657#else
52658#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52659  int64x2_t __s0 = __p0; \
52660  int32x4_t __s1 = __p1; \
52661  int32x2_t __s2 = __p2; \
52662  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52663  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52664  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52665  int64x2_t __ret; \
52666  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52667  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52668  __ret; \
52669})
52670#endif
52671
52672#ifdef __LITTLE_ENDIAN__
52673#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52674  int32x4_t __s0 = __p0; \
52675  int16x8_t __s1 = __p1; \
52676  int16x4_t __s2 = __p2; \
52677  int32x4_t __ret; \
52678  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52679  __ret; \
52680})
52681#else
52682#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52683  int32x4_t __s0 = __p0; \
52684  int16x8_t __s1 = __p1; \
52685  int16x4_t __s2 = __p2; \
52686  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52687  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52688  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52689  int32x4_t __ret; \
52690  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52691  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52692  __ret; \
52693})
52694#endif
52695
52696#ifdef __LITTLE_ENDIAN__
52697#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52698  uint64x2_t __s0 = __p0; \
52699  uint32x4_t __s1 = __p1; \
52700  uint32x4_t __s2 = __p2; \
52701  uint64x2_t __ret; \
52702  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52703  __ret; \
52704})
52705#else
52706#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52707  uint64x2_t __s0 = __p0; \
52708  uint32x4_t __s1 = __p1; \
52709  uint32x4_t __s2 = __p2; \
52710  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52711  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52712  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52713  uint64x2_t __ret; \
52714  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52715  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52716  __ret; \
52717})
52718#endif
52719
52720#ifdef __LITTLE_ENDIAN__
52721#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52722  uint32x4_t __s0 = __p0; \
52723  uint16x8_t __s1 = __p1; \
52724  uint16x8_t __s2 = __p2; \
52725  uint32x4_t __ret; \
52726  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52727  __ret; \
52728})
52729#else
52730#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52731  uint32x4_t __s0 = __p0; \
52732  uint16x8_t __s1 = __p1; \
52733  uint16x8_t __s2 = __p2; \
52734  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52735  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52736  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52737  uint32x4_t __ret; \
52738  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52739  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52740  __ret; \
52741})
52742#endif
52743
52744#ifdef __LITTLE_ENDIAN__
52745#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52746  int64x2_t __s0 = __p0; \
52747  int32x4_t __s1 = __p1; \
52748  int32x4_t __s2 = __p2; \
52749  int64x2_t __ret; \
52750  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52751  __ret; \
52752})
52753#else
52754#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52755  int64x2_t __s0 = __p0; \
52756  int32x4_t __s1 = __p1; \
52757  int32x4_t __s2 = __p2; \
52758  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52759  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52760  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52761  int64x2_t __ret; \
52762  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52763  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52764  __ret; \
52765})
52766#endif
52767
52768#ifdef __LITTLE_ENDIAN__
52769#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52770  int32x4_t __s0 = __p0; \
52771  int16x8_t __s1 = __p1; \
52772  int16x8_t __s2 = __p2; \
52773  int32x4_t __ret; \
52774  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52775  __ret; \
52776})
52777#else
52778#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52779  int32x4_t __s0 = __p0; \
52780  int16x8_t __s1 = __p1; \
52781  int16x8_t __s2 = __p2; \
52782  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52783  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52784  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52785  int32x4_t __ret; \
52786  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52787  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52788  __ret; \
52789})
52790#endif
52791
52792#ifdef __LITTLE_ENDIAN__
52793#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52794  uint64x2_t __s0 = __p0; \
52795  uint32x2_t __s1 = __p1; \
52796  uint32x4_t __s2 = __p2; \
52797  uint64x2_t __ret; \
52798  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52799  __ret; \
52800})
52801#else
52802#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52803  uint64x2_t __s0 = __p0; \
52804  uint32x2_t __s1 = __p1; \
52805  uint32x4_t __s2 = __p2; \
52806  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52807  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52808  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52809  uint64x2_t __ret; \
52810  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52812  __ret; \
52813})
52814#endif
52815
52816#ifdef __LITTLE_ENDIAN__
52817#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52818  uint32x4_t __s0 = __p0; \
52819  uint16x4_t __s1 = __p1; \
52820  uint16x8_t __s2 = __p2; \
52821  uint32x4_t __ret; \
52822  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52823  __ret; \
52824})
52825#else
52826#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52827  uint32x4_t __s0 = __p0; \
52828  uint16x4_t __s1 = __p1; \
52829  uint16x8_t __s2 = __p2; \
52830  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52831  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52832  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52833  uint32x4_t __ret; \
52834  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52835  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52836  __ret; \
52837})
52838#endif
52839
52840#ifdef __LITTLE_ENDIAN__
52841#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52842  int64x2_t __s0 = __p0; \
52843  int32x2_t __s1 = __p1; \
52844  int32x4_t __s2 = __p2; \
52845  int64x2_t __ret; \
52846  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52847  __ret; \
52848})
52849#else
52850#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52851  int64x2_t __s0 = __p0; \
52852  int32x2_t __s1 = __p1; \
52853  int32x4_t __s2 = __p2; \
52854  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52855  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52856  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52857  int64x2_t __ret; \
52858  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52859  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52860  __ret; \
52861})
52862#endif
52863
52864#ifdef __LITTLE_ENDIAN__
52865#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52866  int32x4_t __s0 = __p0; \
52867  int16x4_t __s1 = __p1; \
52868  int16x8_t __s2 = __p2; \
52869  int32x4_t __ret; \
52870  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52871  __ret; \
52872})
52873#else
52874#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52875  int32x4_t __s0 = __p0; \
52876  int16x4_t __s1 = __p1; \
52877  int16x8_t __s2 = __p2; \
52878  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52879  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52880  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52881  int32x4_t __ret; \
52882  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52883  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52884  __ret; \
52885})
52886#endif
52887
52888#ifdef __LITTLE_ENDIAN__
52889__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52890  float64x2_t __ret;
52891  __ret = __p0 - __p1 * __p2;
52892  return __ret;
52893}
52894#else
52895__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52896  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52897  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52898  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52899  float64x2_t __ret;
52900  __ret = __rev0 - __rev1 * __rev2;
52901  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52902  return __ret;
52903}
52904#endif
52905
52906#ifdef __LITTLE_ENDIAN__
52907__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52908  float64x1_t __ret;
52909  __ret = __p0 - __p1 * __p2;
52910  return __ret;
52911}
52912#else
52913__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52914  float64x1_t __ret;
52915  __ret = __p0 - __p1 * __p2;
52916  return __ret;
52917}
52918#endif
52919
52920#ifdef __LITTLE_ENDIAN__
52921#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52922  uint32x4_t __s0 = __p0; \
52923  uint32x4_t __s1 = __p1; \
52924  uint32x4_t __s2 = __p2; \
52925  uint32x4_t __ret; \
52926  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52927  __ret; \
52928})
52929#else
52930#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52931  uint32x4_t __s0 = __p0; \
52932  uint32x4_t __s1 = __p1; \
52933  uint32x4_t __s2 = __p2; \
52934  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52935  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52936  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52937  uint32x4_t __ret; \
52938  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52939  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52940  __ret; \
52941})
52942#endif
52943
52944#ifdef __LITTLE_ENDIAN__
52945#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52946  uint16x8_t __s0 = __p0; \
52947  uint16x8_t __s1 = __p1; \
52948  uint16x8_t __s2 = __p2; \
52949  uint16x8_t __ret; \
52950  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52951  __ret; \
52952})
52953#else
52954#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52955  uint16x8_t __s0 = __p0; \
52956  uint16x8_t __s1 = __p1; \
52957  uint16x8_t __s2 = __p2; \
52958  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52959  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52960  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52961  uint16x8_t __ret; \
52962  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52964  __ret; \
52965})
52966#endif
52967
52968#ifdef __LITTLE_ENDIAN__
52969#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52970  float32x4_t __s0 = __p0; \
52971  float32x4_t __s1 = __p1; \
52972  float32x4_t __s2 = __p2; \
52973  float32x4_t __ret; \
52974  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52975  __ret; \
52976})
52977#else
52978#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52979  float32x4_t __s0 = __p0; \
52980  float32x4_t __s1 = __p1; \
52981  float32x4_t __s2 = __p2; \
52982  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52983  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52984  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52985  float32x4_t __ret; \
52986  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52987  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52988  __ret; \
52989})
52990#endif
52991
52992#ifdef __LITTLE_ENDIAN__
52993#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52994  int32x4_t __s0 = __p0; \
52995  int32x4_t __s1 = __p1; \
52996  int32x4_t __s2 = __p2; \
52997  int32x4_t __ret; \
52998  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52999  __ret; \
53000})
53001#else
53002#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53003  int32x4_t __s0 = __p0; \
53004  int32x4_t __s1 = __p1; \
53005  int32x4_t __s2 = __p2; \
53006  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53007  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53008  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53009  int32x4_t __ret; \
53010  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53011  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53012  __ret; \
53013})
53014#endif
53015
53016#ifdef __LITTLE_ENDIAN__
53017#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53018  int16x8_t __s0 = __p0; \
53019  int16x8_t __s1 = __p1; \
53020  int16x8_t __s2 = __p2; \
53021  int16x8_t __ret; \
53022  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53023  __ret; \
53024})
53025#else
53026#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53027  int16x8_t __s0 = __p0; \
53028  int16x8_t __s1 = __p1; \
53029  int16x8_t __s2 = __p2; \
53030  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53031  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53032  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53033  int16x8_t __ret; \
53034  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53035  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53036  __ret; \
53037})
53038#endif
53039
53040#ifdef __LITTLE_ENDIAN__
53041#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53042  uint32x2_t __s0 = __p0; \
53043  uint32x2_t __s1 = __p1; \
53044  uint32x4_t __s2 = __p2; \
53045  uint32x2_t __ret; \
53046  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53047  __ret; \
53048})
53049#else
53050#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53051  uint32x2_t __s0 = __p0; \
53052  uint32x2_t __s1 = __p1; \
53053  uint32x4_t __s2 = __p2; \
53054  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53055  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53056  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53057  uint32x2_t __ret; \
53058  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53059  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53060  __ret; \
53061})
53062#endif
53063
53064#ifdef __LITTLE_ENDIAN__
53065#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53066  uint16x4_t __s0 = __p0; \
53067  uint16x4_t __s1 = __p1; \
53068  uint16x8_t __s2 = __p2; \
53069  uint16x4_t __ret; \
53070  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53071  __ret; \
53072})
53073#else
53074#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53075  uint16x4_t __s0 = __p0; \
53076  uint16x4_t __s1 = __p1; \
53077  uint16x8_t __s2 = __p2; \
53078  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53079  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53080  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53081  uint16x4_t __ret; \
53082  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53084  __ret; \
53085})
53086#endif
53087
53088#ifdef __LITTLE_ENDIAN__
53089#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53090  float32x2_t __s0 = __p0; \
53091  float32x2_t __s1 = __p1; \
53092  float32x4_t __s2 = __p2; \
53093  float32x2_t __ret; \
53094  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53095  __ret; \
53096})
53097#else
53098#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53099  float32x2_t __s0 = __p0; \
53100  float32x2_t __s1 = __p1; \
53101  float32x4_t __s2 = __p2; \
53102  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53103  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53104  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53105  float32x2_t __ret; \
53106  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53108  __ret; \
53109})
53110#endif
53111
53112#ifdef __LITTLE_ENDIAN__
53113#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53114  int32x2_t __s0 = __p0; \
53115  int32x2_t __s1 = __p1; \
53116  int32x4_t __s2 = __p2; \
53117  int32x2_t __ret; \
53118  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53119  __ret; \
53120})
53121#else
53122#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53123  int32x2_t __s0 = __p0; \
53124  int32x2_t __s1 = __p1; \
53125  int32x4_t __s2 = __p2; \
53126  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53127  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53128  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53129  int32x2_t __ret; \
53130  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53131  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53132  __ret; \
53133})
53134#endif
53135
53136#ifdef __LITTLE_ENDIAN__
53137#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53138  int16x4_t __s0 = __p0; \
53139  int16x4_t __s1 = __p1; \
53140  int16x8_t __s2 = __p2; \
53141  int16x4_t __ret; \
53142  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53143  __ret; \
53144})
53145#else
53146#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53147  int16x4_t __s0 = __p0; \
53148  int16x4_t __s1 = __p1; \
53149  int16x8_t __s2 = __p2; \
53150  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53151  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53152  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53153  int16x4_t __ret; \
53154  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53156  __ret; \
53157})
53158#endif
53159
53160#ifdef __LITTLE_ENDIAN__
53161__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53162  float64x2_t __ret;
53163  __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
53164  return __ret;
53165}
53166#else
53167__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53168  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53169  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53170  float64x2_t __ret;
53171  __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
53172  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53173  return __ret;
53174}
53175#endif
53176
53177#ifdef __LITTLE_ENDIAN__
53178#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53179  uint64x2_t __s0 = __p0; \
53180  uint32x4_t __s1 = __p1; \
53181  uint32x2_t __s2 = __p2; \
53182  uint64x2_t __ret; \
53183  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53184  __ret; \
53185})
53186#else
53187#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53188  uint64x2_t __s0 = __p0; \
53189  uint32x4_t __s1 = __p1; \
53190  uint32x2_t __s2 = __p2; \
53191  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53192  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53193  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53194  uint64x2_t __ret; \
53195  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53196  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53197  __ret; \
53198})
53199#endif
53200
53201#ifdef __LITTLE_ENDIAN__
53202#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53203  uint32x4_t __s0 = __p0; \
53204  uint16x8_t __s1 = __p1; \
53205  uint16x4_t __s2 = __p2; \
53206  uint32x4_t __ret; \
53207  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53208  __ret; \
53209})
53210#else
53211#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53212  uint32x4_t __s0 = __p0; \
53213  uint16x8_t __s1 = __p1; \
53214  uint16x4_t __s2 = __p2; \
53215  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53216  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53217  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53218  uint32x4_t __ret; \
53219  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53220  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53221  __ret; \
53222})
53223#endif
53224
53225#ifdef __LITTLE_ENDIAN__
53226#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53227  int64x2_t __s0 = __p0; \
53228  int32x4_t __s1 = __p1; \
53229  int32x2_t __s2 = __p2; \
53230  int64x2_t __ret; \
53231  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53232  __ret; \
53233})
53234#else
53235#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53236  int64x2_t __s0 = __p0; \
53237  int32x4_t __s1 = __p1; \
53238  int32x2_t __s2 = __p2; \
53239  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53240  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53241  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53242  int64x2_t __ret; \
53243  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53244  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53245  __ret; \
53246})
53247#endif
53248
53249#ifdef __LITTLE_ENDIAN__
53250#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53251  int32x4_t __s0 = __p0; \
53252  int16x8_t __s1 = __p1; \
53253  int16x4_t __s2 = __p2; \
53254  int32x4_t __ret; \
53255  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53256  __ret; \
53257})
53258#else
53259#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53260  int32x4_t __s0 = __p0; \
53261  int16x8_t __s1 = __p1; \
53262  int16x4_t __s2 = __p2; \
53263  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53264  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53265  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53266  int32x4_t __ret; \
53267  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53269  __ret; \
53270})
53271#endif
53272
53273#ifdef __LITTLE_ENDIAN__
53274#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53275  uint64x2_t __s0 = __p0; \
53276  uint32x4_t __s1 = __p1; \
53277  uint32x4_t __s2 = __p2; \
53278  uint64x2_t __ret; \
53279  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53280  __ret; \
53281})
53282#else
53283#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53284  uint64x2_t __s0 = __p0; \
53285  uint32x4_t __s1 = __p1; \
53286  uint32x4_t __s2 = __p2; \
53287  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53288  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53289  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53290  uint64x2_t __ret; \
53291  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53292  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53293  __ret; \
53294})
53295#endif
53296
53297#ifdef __LITTLE_ENDIAN__
53298#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53299  uint32x4_t __s0 = __p0; \
53300  uint16x8_t __s1 = __p1; \
53301  uint16x8_t __s2 = __p2; \
53302  uint32x4_t __ret; \
53303  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53304  __ret; \
53305})
53306#else
53307#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53308  uint32x4_t __s0 = __p0; \
53309  uint16x8_t __s1 = __p1; \
53310  uint16x8_t __s2 = __p2; \
53311  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53312  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53313  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53314  uint32x4_t __ret; \
53315  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53316  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53317  __ret; \
53318})
53319#endif
53320
53321#ifdef __LITTLE_ENDIAN__
53322#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53323  int64x2_t __s0 = __p0; \
53324  int32x4_t __s1 = __p1; \
53325  int32x4_t __s2 = __p2; \
53326  int64x2_t __ret; \
53327  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53328  __ret; \
53329})
53330#else
53331#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53332  int64x2_t __s0 = __p0; \
53333  int32x4_t __s1 = __p1; \
53334  int32x4_t __s2 = __p2; \
53335  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53336  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53337  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53338  int64x2_t __ret; \
53339  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53340  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53341  __ret; \
53342})
53343#endif
53344
53345#ifdef __LITTLE_ENDIAN__
53346#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53347  int32x4_t __s0 = __p0; \
53348  int16x8_t __s1 = __p1; \
53349  int16x8_t __s2 = __p2; \
53350  int32x4_t __ret; \
53351  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53352  __ret; \
53353})
53354#else
53355#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53356  int32x4_t __s0 = __p0; \
53357  int16x8_t __s1 = __p1; \
53358  int16x8_t __s2 = __p2; \
53359  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53360  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53361  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53362  int32x4_t __ret; \
53363  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53364  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53365  __ret; \
53366})
53367#endif
53368
53369#ifdef __LITTLE_ENDIAN__
53370#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53371  uint64x2_t __s0 = __p0; \
53372  uint32x2_t __s1 = __p1; \
53373  uint32x4_t __s2 = __p2; \
53374  uint64x2_t __ret; \
53375  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53376  __ret; \
53377})
53378#else
53379#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53380  uint64x2_t __s0 = __p0; \
53381  uint32x2_t __s1 = __p1; \
53382  uint32x4_t __s2 = __p2; \
53383  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53384  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53385  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53386  uint64x2_t __ret; \
53387  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53388  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53389  __ret; \
53390})
53391#endif
53392
53393#ifdef __LITTLE_ENDIAN__
53394#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53395  uint32x4_t __s0 = __p0; \
53396  uint16x4_t __s1 = __p1; \
53397  uint16x8_t __s2 = __p2; \
53398  uint32x4_t __ret; \
53399  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53400  __ret; \
53401})
53402#else
53403#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53404  uint32x4_t __s0 = __p0; \
53405  uint16x4_t __s1 = __p1; \
53406  uint16x8_t __s2 = __p2; \
53407  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53408  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53409  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53410  uint32x4_t __ret; \
53411  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53412  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53413  __ret; \
53414})
53415#endif
53416
53417#ifdef __LITTLE_ENDIAN__
53418#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53419  int64x2_t __s0 = __p0; \
53420  int32x2_t __s1 = __p1; \
53421  int32x4_t __s2 = __p2; \
53422  int64x2_t __ret; \
53423  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53424  __ret; \
53425})
53426#else
53427#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53428  int64x2_t __s0 = __p0; \
53429  int32x2_t __s1 = __p1; \
53430  int32x4_t __s2 = __p2; \
53431  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53432  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53433  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53434  int64x2_t __ret; \
53435  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53436  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53437  __ret; \
53438})
53439#endif
53440
53441#ifdef __LITTLE_ENDIAN__
53442#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53443  int32x4_t __s0 = __p0; \
53444  int16x4_t __s1 = __p1; \
53445  int16x8_t __s2 = __p2; \
53446  int32x4_t __ret; \
53447  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53448  __ret; \
53449})
53450#else
53451#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53452  int32x4_t __s0 = __p0; \
53453  int16x4_t __s1 = __p1; \
53454  int16x8_t __s2 = __p2; \
53455  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53456  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53457  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53458  int32x4_t __ret; \
53459  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53461  __ret; \
53462})
53463#endif
53464
53465#ifdef __LITTLE_ENDIAN__
53466__ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53467  poly64x1_t __ret;
53468  __ret = (poly64x1_t) {__p0};
53469  return __ret;
53470}
53471#else
53472__ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53473  poly64x1_t __ret;
53474  __ret = (poly64x1_t) {__p0};
53475  return __ret;
53476}
53477#endif
53478
53479#ifdef __LITTLE_ENDIAN__
53480__ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53481  poly64x2_t __ret;
53482  __ret = (poly64x2_t) {__p0, __p0};
53483  return __ret;
53484}
53485#else
53486__ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53487  poly64x2_t __ret;
53488  __ret = (poly64x2_t) {__p0, __p0};
53489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53490  return __ret;
53491}
53492#endif
53493
53494#ifdef __LITTLE_ENDIAN__
53495__ai float64x2_t vmovq_n_f64(float64_t __p0) {
53496  float64x2_t __ret;
53497  __ret = (float64x2_t) {__p0, __p0};
53498  return __ret;
53499}
53500#else
53501__ai float64x2_t vmovq_n_f64(float64_t __p0) {
53502  float64x2_t __ret;
53503  __ret = (float64x2_t) {__p0, __p0};
53504  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53505  return __ret;
53506}
53507#endif
53508
53509#ifdef __LITTLE_ENDIAN__
53510__ai float64x1_t vmov_n_f64(float64_t __p0) {
53511  float64x1_t __ret;
53512  __ret = (float64x1_t) {__p0};
53513  return __ret;
53514}
53515#else
53516__ai float64x1_t vmov_n_f64(float64_t __p0) {
53517  float64x1_t __ret;
53518  __ret = (float64x1_t) {__p0};
53519  return __ret;
53520}
53521#endif
53522
53523#ifdef __LITTLE_ENDIAN__
53524__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_112) {
53525  uint16x8_t __ret_112;
53526  uint8x8_t __a1_112 = vget_high_u8(__p0_112);
53527  __ret_112 = (uint16x8_t)(vshll_n_u8(__a1_112, 0));
53528  return __ret_112;
53529}
53530#else
53531__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_113) {
53532  uint8x16_t __rev0_113;  __rev0_113 = __builtin_shufflevector(__p0_113, __p0_113, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53533  uint16x8_t __ret_113;
53534  uint8x8_t __a1_113 = __noswap_vget_high_u8(__rev0_113);
53535  __ret_113 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_113, 0));
53536  __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 7, 6, 5, 4, 3, 2, 1, 0);
53537  return __ret_113;
53538}
53539__ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_114) {
53540  uint16x8_t __ret_114;
53541  uint8x8_t __a1_114 = __noswap_vget_high_u8(__p0_114);
53542  __ret_114 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_114, 0));
53543  return __ret_114;
53544}
53545#endif
53546
53547#ifdef __LITTLE_ENDIAN__
53548__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_115) {
53549  uint64x2_t __ret_115;
53550  uint32x2_t __a1_115 = vget_high_u32(__p0_115);
53551  __ret_115 = (uint64x2_t)(vshll_n_u32(__a1_115, 0));
53552  return __ret_115;
53553}
53554#else
53555__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_116) {
53556  uint32x4_t __rev0_116;  __rev0_116 = __builtin_shufflevector(__p0_116, __p0_116, 3, 2, 1, 0);
53557  uint64x2_t __ret_116;
53558  uint32x2_t __a1_116 = __noswap_vget_high_u32(__rev0_116);
53559  __ret_116 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_116, 0));
53560  __ret_116 = __builtin_shufflevector(__ret_116, __ret_116, 1, 0);
53561  return __ret_116;
53562}
53563__ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_117) {
53564  uint64x2_t __ret_117;
53565  uint32x2_t __a1_117 = __noswap_vget_high_u32(__p0_117);
53566  __ret_117 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_117, 0));
53567  return __ret_117;
53568}
53569#endif
53570
53571#ifdef __LITTLE_ENDIAN__
53572__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_118) {
53573  uint32x4_t __ret_118;
53574  uint16x4_t __a1_118 = vget_high_u16(__p0_118);
53575  __ret_118 = (uint32x4_t)(vshll_n_u16(__a1_118, 0));
53576  return __ret_118;
53577}
53578#else
53579__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_119) {
53580  uint16x8_t __rev0_119;  __rev0_119 = __builtin_shufflevector(__p0_119, __p0_119, 7, 6, 5, 4, 3, 2, 1, 0);
53581  uint32x4_t __ret_119;
53582  uint16x4_t __a1_119 = __noswap_vget_high_u16(__rev0_119);
53583  __ret_119 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_119, 0));
53584  __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 3, 2, 1, 0);
53585  return __ret_119;
53586}
53587__ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_120) {
53588  uint32x4_t __ret_120;
53589  uint16x4_t __a1_120 = __noswap_vget_high_u16(__p0_120);
53590  __ret_120 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_120, 0));
53591  return __ret_120;
53592}
53593#endif
53594
53595#ifdef __LITTLE_ENDIAN__
53596__ai int16x8_t vmovl_high_s8(int8x16_t __p0_121) {
53597  int16x8_t __ret_121;
53598  int8x8_t __a1_121 = vget_high_s8(__p0_121);
53599  __ret_121 = (int16x8_t)(vshll_n_s8(__a1_121, 0));
53600  return __ret_121;
53601}
53602#else
53603__ai int16x8_t vmovl_high_s8(int8x16_t __p0_122) {
53604  int8x16_t __rev0_122;  __rev0_122 = __builtin_shufflevector(__p0_122, __p0_122, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53605  int16x8_t __ret_122;
53606  int8x8_t __a1_122 = __noswap_vget_high_s8(__rev0_122);
53607  __ret_122 = (int16x8_t)(__noswap_vshll_n_s8(__a1_122, 0));
53608  __ret_122 = __builtin_shufflevector(__ret_122, __ret_122, 7, 6, 5, 4, 3, 2, 1, 0);
53609  return __ret_122;
53610}
53611__ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_123) {
53612  int16x8_t __ret_123;
53613  int8x8_t __a1_123 = __noswap_vget_high_s8(__p0_123);
53614  __ret_123 = (int16x8_t)(__noswap_vshll_n_s8(__a1_123, 0));
53615  return __ret_123;
53616}
53617#endif
53618
53619#ifdef __LITTLE_ENDIAN__
53620__ai int64x2_t vmovl_high_s32(int32x4_t __p0_124) {
53621  int64x2_t __ret_124;
53622  int32x2_t __a1_124 = vget_high_s32(__p0_124);
53623  __ret_124 = (int64x2_t)(vshll_n_s32(__a1_124, 0));
53624  return __ret_124;
53625}
53626#else
53627__ai int64x2_t vmovl_high_s32(int32x4_t __p0_125) {
53628  int32x4_t __rev0_125;  __rev0_125 = __builtin_shufflevector(__p0_125, __p0_125, 3, 2, 1, 0);
53629  int64x2_t __ret_125;
53630  int32x2_t __a1_125 = __noswap_vget_high_s32(__rev0_125);
53631  __ret_125 = (int64x2_t)(__noswap_vshll_n_s32(__a1_125, 0));
53632  __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 1, 0);
53633  return __ret_125;
53634}
53635__ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_126) {
53636  int64x2_t __ret_126;
53637  int32x2_t __a1_126 = __noswap_vget_high_s32(__p0_126);
53638  __ret_126 = (int64x2_t)(__noswap_vshll_n_s32(__a1_126, 0));
53639  return __ret_126;
53640}
53641#endif
53642
53643#ifdef __LITTLE_ENDIAN__
53644__ai int32x4_t vmovl_high_s16(int16x8_t __p0_127) {
53645  int32x4_t __ret_127;
53646  int16x4_t __a1_127 = vget_high_s16(__p0_127);
53647  __ret_127 = (int32x4_t)(vshll_n_s16(__a1_127, 0));
53648  return __ret_127;
53649}
53650#else
53651__ai int32x4_t vmovl_high_s16(int16x8_t __p0_128) {
53652  int16x8_t __rev0_128;  __rev0_128 = __builtin_shufflevector(__p0_128, __p0_128, 7, 6, 5, 4, 3, 2, 1, 0);
53653  int32x4_t __ret_128;
53654  int16x4_t __a1_128 = __noswap_vget_high_s16(__rev0_128);
53655  __ret_128 = (int32x4_t)(__noswap_vshll_n_s16(__a1_128, 0));
53656  __ret_128 = __builtin_shufflevector(__ret_128, __ret_128, 3, 2, 1, 0);
53657  return __ret_128;
53658}
53659__ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_129) {
53660  int32x4_t __ret_129;
53661  int16x4_t __a1_129 = __noswap_vget_high_s16(__p0_129);
53662  __ret_129 = (int32x4_t)(__noswap_vshll_n_s16(__a1_129, 0));
53663  return __ret_129;
53664}
53665#endif
53666
53667#ifdef __LITTLE_ENDIAN__
53668__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53669  uint16x8_t __ret;
53670  __ret = vcombine_u16(__p0, vmovn_u32(__p1));
53671  return __ret;
53672}
53673#else
53674__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53675  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53676  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53677  uint16x8_t __ret;
53678  __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
53679  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53680  return __ret;
53681}
53682#endif
53683
53684#ifdef __LITTLE_ENDIAN__
53685__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53686  uint32x4_t __ret;
53687  __ret = vcombine_u32(__p0, vmovn_u64(__p1));
53688  return __ret;
53689}
53690#else
53691__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53692  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53693  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53694  uint32x4_t __ret;
53695  __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
53696  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53697  return __ret;
53698}
53699#endif
53700
53701#ifdef __LITTLE_ENDIAN__
53702__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53703  uint8x16_t __ret;
53704  __ret = vcombine_u8(__p0, vmovn_u16(__p1));
53705  return __ret;
53706}
53707#else
53708__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53709  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53710  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53711  uint8x16_t __ret;
53712  __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
53713  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53714  return __ret;
53715}
53716#endif
53717
53718#ifdef __LITTLE_ENDIAN__
53719__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53720  int16x8_t __ret;
53721  __ret = vcombine_s16(__p0, vmovn_s32(__p1));
53722  return __ret;
53723}
53724#else
53725__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53726  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53727  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53728  int16x8_t __ret;
53729  __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
53730  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53731  return __ret;
53732}
53733#endif
53734
53735#ifdef __LITTLE_ENDIAN__
53736__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53737  int32x4_t __ret;
53738  __ret = vcombine_s32(__p0, vmovn_s64(__p1));
53739  return __ret;
53740}
53741#else
53742__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53743  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53744  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53745  int32x4_t __ret;
53746  __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
53747  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53748  return __ret;
53749}
53750#endif
53751
53752#ifdef __LITTLE_ENDIAN__
53753__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53754  int8x16_t __ret;
53755  __ret = vcombine_s8(__p0, vmovn_s16(__p1));
53756  return __ret;
53757}
53758#else
53759__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53760  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53761  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53762  int8x16_t __ret;
53763  __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
53764  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53765  return __ret;
53766}
53767#endif
53768
53769#ifdef __LITTLE_ENDIAN__
53770__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53771  float64x2_t __ret;
53772  __ret = __p0 * __p1;
53773  return __ret;
53774}
53775#else
53776__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53777  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53778  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53779  float64x2_t __ret;
53780  __ret = __rev0 * __rev1;
53781  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53782  return __ret;
53783}
53784#endif
53785
53786#ifdef __LITTLE_ENDIAN__
53787__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53788  float64x1_t __ret;
53789  __ret = __p0 * __p1;
53790  return __ret;
53791}
53792#else
53793__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53794  float64x1_t __ret;
53795  __ret = __p0 * __p1;
53796  return __ret;
53797}
53798#endif
53799
53800#ifdef __LITTLE_ENDIAN__
53801#define vmuld_lane_f64(__p0_130, __p1_130, __p2_130) __extension__ ({ \
53802  float64_t __s0_130 = __p0_130; \
53803  float64x1_t __s1_130 = __p1_130; \
53804  float64_t __ret_130; \
53805  __ret_130 = __s0_130 * vget_lane_f64(__s1_130, __p2_130); \
53806  __ret_130; \
53807})
53808#else
53809#define vmuld_lane_f64(__p0_131, __p1_131, __p2_131) __extension__ ({ \
53810  float64_t __s0_131 = __p0_131; \
53811  float64x1_t __s1_131 = __p1_131; \
53812  float64_t __ret_131; \
53813  __ret_131 = __s0_131 * __noswap_vget_lane_f64(__s1_131, __p2_131); \
53814  __ret_131; \
53815})
53816#endif
53817
53818#ifdef __LITTLE_ENDIAN__
53819#define vmuls_lane_f32(__p0_132, __p1_132, __p2_132) __extension__ ({ \
53820  float32_t __s0_132 = __p0_132; \
53821  float32x2_t __s1_132 = __p1_132; \
53822  float32_t __ret_132; \
53823  __ret_132 = __s0_132 * vget_lane_f32(__s1_132, __p2_132); \
53824  __ret_132; \
53825})
53826#else
53827#define vmuls_lane_f32(__p0_133, __p1_133, __p2_133) __extension__ ({ \
53828  float32_t __s0_133 = __p0_133; \
53829  float32x2_t __s1_133 = __p1_133; \
53830  float32x2_t __rev1_133;  __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 1, 0); \
53831  float32_t __ret_133; \
53832  __ret_133 = __s0_133 * __noswap_vget_lane_f32(__rev1_133, __p2_133); \
53833  __ret_133; \
53834})
53835#endif
53836
53837#ifdef __LITTLE_ENDIAN__
53838#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53839  float64x1_t __s0 = __p0; \
53840  float64x1_t __s1 = __p1; \
53841  float64x1_t __ret; \
53842  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53843  __ret; \
53844})
53845#else
53846#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53847  float64x1_t __s0 = __p0; \
53848  float64x1_t __s1 = __p1; \
53849  float64x1_t __ret; \
53850  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53851  __ret; \
53852})
53853#endif
53854
53855#ifdef __LITTLE_ENDIAN__
53856#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53857  float64x2_t __s0 = __p0; \
53858  float64x1_t __s1 = __p1; \
53859  float64x2_t __ret; \
53860  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53861  __ret; \
53862})
53863#else
53864#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53865  float64x2_t __s0 = __p0; \
53866  float64x1_t __s1 = __p1; \
53867  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53868  float64x2_t __ret; \
53869  __ret = __rev0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53870  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53871  __ret; \
53872})
53873#endif
53874
53875#ifdef __LITTLE_ENDIAN__
53876#define vmuld_laneq_f64(__p0_134, __p1_134, __p2_134) __extension__ ({ \
53877  float64_t __s0_134 = __p0_134; \
53878  float64x2_t __s1_134 = __p1_134; \
53879  float64_t __ret_134; \
53880  __ret_134 = __s0_134 * vgetq_lane_f64(__s1_134, __p2_134); \
53881  __ret_134; \
53882})
53883#else
53884#define vmuld_laneq_f64(__p0_135, __p1_135, __p2_135) __extension__ ({ \
53885  float64_t __s0_135 = __p0_135; \
53886  float64x2_t __s1_135 = __p1_135; \
53887  float64x2_t __rev1_135;  __rev1_135 = __builtin_shufflevector(__s1_135, __s1_135, 1, 0); \
53888  float64_t __ret_135; \
53889  __ret_135 = __s0_135 * __noswap_vgetq_lane_f64(__rev1_135, __p2_135); \
53890  __ret_135; \
53891})
53892#endif
53893
53894#ifdef __LITTLE_ENDIAN__
53895#define vmuls_laneq_f32(__p0_136, __p1_136, __p2_136) __extension__ ({ \
53896  float32_t __s0_136 = __p0_136; \
53897  float32x4_t __s1_136 = __p1_136; \
53898  float32_t __ret_136; \
53899  __ret_136 = __s0_136 * vgetq_lane_f32(__s1_136, __p2_136); \
53900  __ret_136; \
53901})
53902#else
53903#define vmuls_laneq_f32(__p0_137, __p1_137, __p2_137) __extension__ ({ \
53904  float32_t __s0_137 = __p0_137; \
53905  float32x4_t __s1_137 = __p1_137; \
53906  float32x4_t __rev1_137;  __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 3, 2, 1, 0); \
53907  float32_t __ret_137; \
53908  __ret_137 = __s0_137 * __noswap_vgetq_lane_f32(__rev1_137, __p2_137); \
53909  __ret_137; \
53910})
53911#endif
53912
53913#ifdef __LITTLE_ENDIAN__
53914#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53915  float64x1_t __s0 = __p0; \
53916  float64x2_t __s1 = __p1; \
53917  float64x1_t __ret; \
53918  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
53919  __ret; \
53920})
53921#else
53922#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53923  float64x1_t __s0 = __p0; \
53924  float64x2_t __s1 = __p1; \
53925  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53926  float64x1_t __ret; \
53927  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
53928  __ret; \
53929})
53930#endif
53931
53932#ifdef __LITTLE_ENDIAN__
53933#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53934  uint32x4_t __s0 = __p0; \
53935  uint32x4_t __s1 = __p1; \
53936  uint32x4_t __ret; \
53937  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53938  __ret; \
53939})
53940#else
53941#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53942  uint32x4_t __s0 = __p0; \
53943  uint32x4_t __s1 = __p1; \
53944  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53945  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53946  uint32x4_t __ret; \
53947  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53948  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53949  __ret; \
53950})
53951#endif
53952
53953#ifdef __LITTLE_ENDIAN__
53954#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53955  uint16x8_t __s0 = __p0; \
53956  uint16x8_t __s1 = __p1; \
53957  uint16x8_t __ret; \
53958  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53959  __ret; \
53960})
53961#else
53962#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53963  uint16x8_t __s0 = __p0; \
53964  uint16x8_t __s1 = __p1; \
53965  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53966  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53967  uint16x8_t __ret; \
53968  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53969  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53970  __ret; \
53971})
53972#endif
53973
53974#ifdef __LITTLE_ENDIAN__
53975#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53976  float64x2_t __s0 = __p0; \
53977  float64x2_t __s1 = __p1; \
53978  float64x2_t __ret; \
53979  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53980  __ret; \
53981})
53982#else
53983#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53984  float64x2_t __s0 = __p0; \
53985  float64x2_t __s1 = __p1; \
53986  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53987  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53988  float64x2_t __ret; \
53989  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53990  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53991  __ret; \
53992})
53993#endif
53994
53995#ifdef __LITTLE_ENDIAN__
53996#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53997  float32x4_t __s0 = __p0; \
53998  float32x4_t __s1 = __p1; \
53999  float32x4_t __ret; \
54000  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54001  __ret; \
54002})
54003#else
54004#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54005  float32x4_t __s0 = __p0; \
54006  float32x4_t __s1 = __p1; \
54007  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54008  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54009  float32x4_t __ret; \
54010  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54011  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54012  __ret; \
54013})
54014#endif
54015
54016#ifdef __LITTLE_ENDIAN__
54017#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54018  int32x4_t __s0 = __p0; \
54019  int32x4_t __s1 = __p1; \
54020  int32x4_t __ret; \
54021  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54022  __ret; \
54023})
54024#else
54025#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54026  int32x4_t __s0 = __p0; \
54027  int32x4_t __s1 = __p1; \
54028  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54029  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54030  int32x4_t __ret; \
54031  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54032  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54033  __ret; \
54034})
54035#endif
54036
54037#ifdef __LITTLE_ENDIAN__
54038#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54039  int16x8_t __s0 = __p0; \
54040  int16x8_t __s1 = __p1; \
54041  int16x8_t __ret; \
54042  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54043  __ret; \
54044})
54045#else
54046#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54047  int16x8_t __s0 = __p0; \
54048  int16x8_t __s1 = __p1; \
54049  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54050  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54051  int16x8_t __ret; \
54052  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54053  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
54054  __ret; \
54055})
54056#endif
54057
54058#ifdef __LITTLE_ENDIAN__
54059#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54060  uint32x2_t __s0 = __p0; \
54061  uint32x4_t __s1 = __p1; \
54062  uint32x2_t __ret; \
54063  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54064  __ret; \
54065})
54066#else
54067#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54068  uint32x2_t __s0 = __p0; \
54069  uint32x4_t __s1 = __p1; \
54070  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54071  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54072  uint32x2_t __ret; \
54073  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54074  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54075  __ret; \
54076})
54077#endif
54078
54079#ifdef __LITTLE_ENDIAN__
54080#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54081  uint16x4_t __s0 = __p0; \
54082  uint16x8_t __s1 = __p1; \
54083  uint16x4_t __ret; \
54084  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54085  __ret; \
54086})
54087#else
54088#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54089  uint16x4_t __s0 = __p0; \
54090  uint16x8_t __s1 = __p1; \
54091  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54092  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54093  uint16x4_t __ret; \
54094  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54095  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54096  __ret; \
54097})
54098#endif
54099
54100#ifdef __LITTLE_ENDIAN__
54101#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54102  float32x2_t __s0 = __p0; \
54103  float32x4_t __s1 = __p1; \
54104  float32x2_t __ret; \
54105  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54106  __ret; \
54107})
54108#else
54109#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54110  float32x2_t __s0 = __p0; \
54111  float32x4_t __s1 = __p1; \
54112  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54113  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54114  float32x2_t __ret; \
54115  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54116  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54117  __ret; \
54118})
54119#endif
54120
54121#ifdef __LITTLE_ENDIAN__
54122#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54123  int32x2_t __s0 = __p0; \
54124  int32x4_t __s1 = __p1; \
54125  int32x2_t __ret; \
54126  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54127  __ret; \
54128})
54129#else
54130#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54131  int32x2_t __s0 = __p0; \
54132  int32x4_t __s1 = __p1; \
54133  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54134  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54135  int32x2_t __ret; \
54136  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54137  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54138  __ret; \
54139})
54140#endif
54141
54142#ifdef __LITTLE_ENDIAN__
54143#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54144  int16x4_t __s0 = __p0; \
54145  int16x8_t __s1 = __p1; \
54146  int16x4_t __ret; \
54147  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54148  __ret; \
54149})
54150#else
54151#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54152  int16x4_t __s0 = __p0; \
54153  int16x8_t __s1 = __p1; \
54154  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54155  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54156  int16x4_t __ret; \
54157  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54158  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54159  __ret; \
54160})
54161#endif
54162
54163#ifdef __LITTLE_ENDIAN__
54164__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54165  float64x1_t __ret;
54166  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54167  return __ret;
54168}
54169#else
54170__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54171  float64x1_t __ret;
54172  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54173  return __ret;
54174}
54175#endif
54176
54177#ifdef __LITTLE_ENDIAN__
54178__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54179  float64x2_t __ret;
54180  __ret = __p0 * (float64x2_t) {__p1, __p1};
54181  return __ret;
54182}
54183#else
54184__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54185  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54186  float64x2_t __ret;
54187  __ret = __rev0 * (float64x2_t) {__p1, __p1};
54188  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54189  return __ret;
54190}
54191#endif
54192
54193#ifdef __LITTLE_ENDIAN__
54194__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54195  poly128_t __ret;
54196  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54197  return __ret;
54198}
54199#else
54200__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54201  poly128_t __ret;
54202  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54203  return __ret;
54204}
54205__ai poly128_t __noswap_vmull_p64(poly64_t __p0, poly64_t __p1) {
54206  poly128_t __ret;
54207  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54208  return __ret;
54209}
54210#endif
54211
54212#ifdef __LITTLE_ENDIAN__
54213__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54214  poly16x8_t __ret;
54215  __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
54216  return __ret;
54217}
54218#else
54219__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54220  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54221  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54222  poly16x8_t __ret;
54223  __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
54224  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54225  return __ret;
54226}
54227#endif
54228
54229#ifdef __LITTLE_ENDIAN__
54230__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54231  uint16x8_t __ret;
54232  __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
54233  return __ret;
54234}
54235#else
54236__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54237  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54238  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54239  uint16x8_t __ret;
54240  __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
54241  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54242  return __ret;
54243}
54244#endif
54245
54246#ifdef __LITTLE_ENDIAN__
54247__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54248  uint64x2_t __ret;
54249  __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
54250  return __ret;
54251}
54252#else
54253__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54254  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54255  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54256  uint64x2_t __ret;
54257  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
54258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54259  return __ret;
54260}
54261#endif
54262
54263#ifdef __LITTLE_ENDIAN__
54264__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54265  uint32x4_t __ret;
54266  __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
54267  return __ret;
54268}
54269#else
54270__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54271  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54272  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54273  uint32x4_t __ret;
54274  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
54275  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54276  return __ret;
54277}
54278#endif
54279
54280#ifdef __LITTLE_ENDIAN__
54281__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54282  int16x8_t __ret;
54283  __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
54284  return __ret;
54285}
54286#else
54287__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54288  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54289  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54290  int16x8_t __ret;
54291  __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
54292  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54293  return __ret;
54294}
54295#endif
54296
54297#ifdef __LITTLE_ENDIAN__
54298__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54299  int64x2_t __ret;
54300  __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
54301  return __ret;
54302}
54303#else
54304__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54305  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54306  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54307  int64x2_t __ret;
54308  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
54309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54310  return __ret;
54311}
54312#endif
54313
54314#ifdef __LITTLE_ENDIAN__
54315__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54316  int32x4_t __ret;
54317  __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
54318  return __ret;
54319}
54320#else
54321__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54322  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54323  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54324  int32x4_t __ret;
54325  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
54326  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54327  return __ret;
54328}
54329#endif
54330
54331#ifdef __LITTLE_ENDIAN__
54332__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54333  poly128_t __ret;
54334  __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
54335  return __ret;
54336}
54337#else
54338__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54339  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54340  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54341  poly128_t __ret;
54342  __ret = __noswap_vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
54343  return __ret;
54344}
54345#endif
54346
54347#ifdef __LITTLE_ENDIAN__
54348#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54349  uint32x4_t __s0 = __p0; \
54350  uint32x2_t __s1 = __p1; \
54351  uint64x2_t __ret; \
54352  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54353  __ret; \
54354})
54355#else
54356#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54357  uint32x4_t __s0 = __p0; \
54358  uint32x2_t __s1 = __p1; \
54359  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54360  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54361  uint64x2_t __ret; \
54362  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54363  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54364  __ret; \
54365})
54366#endif
54367
54368#ifdef __LITTLE_ENDIAN__
54369#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54370  uint16x8_t __s0 = __p0; \
54371  uint16x4_t __s1 = __p1; \
54372  uint32x4_t __ret; \
54373  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54374  __ret; \
54375})
54376#else
54377#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54378  uint16x8_t __s0 = __p0; \
54379  uint16x4_t __s1 = __p1; \
54380  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54381  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54382  uint32x4_t __ret; \
54383  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54384  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54385  __ret; \
54386})
54387#endif
54388
54389#ifdef __LITTLE_ENDIAN__
54390#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54391  int32x4_t __s0 = __p0; \
54392  int32x2_t __s1 = __p1; \
54393  int64x2_t __ret; \
54394  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54395  __ret; \
54396})
54397#else
54398#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54399  int32x4_t __s0 = __p0; \
54400  int32x2_t __s1 = __p1; \
54401  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54402  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54403  int64x2_t __ret; \
54404  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54406  __ret; \
54407})
54408#endif
54409
54410#ifdef __LITTLE_ENDIAN__
54411#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54412  int16x8_t __s0 = __p0; \
54413  int16x4_t __s1 = __p1; \
54414  int32x4_t __ret; \
54415  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54416  __ret; \
54417})
54418#else
54419#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54420  int16x8_t __s0 = __p0; \
54421  int16x4_t __s1 = __p1; \
54422  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54423  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54424  int32x4_t __ret; \
54425  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54426  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54427  __ret; \
54428})
54429#endif
54430
54431#ifdef __LITTLE_ENDIAN__
54432#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54433  uint32x4_t __s0 = __p0; \
54434  uint32x4_t __s1 = __p1; \
54435  uint64x2_t __ret; \
54436  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54437  __ret; \
54438})
54439#else
54440#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54441  uint32x4_t __s0 = __p0; \
54442  uint32x4_t __s1 = __p1; \
54443  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54444  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54445  uint64x2_t __ret; \
54446  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54447  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54448  __ret; \
54449})
54450#endif
54451
54452#ifdef __LITTLE_ENDIAN__
54453#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54454  uint16x8_t __s0 = __p0; \
54455  uint16x8_t __s1 = __p1; \
54456  uint32x4_t __ret; \
54457  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54458  __ret; \
54459})
54460#else
54461#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54462  uint16x8_t __s0 = __p0; \
54463  uint16x8_t __s1 = __p1; \
54464  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54465  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54466  uint32x4_t __ret; \
54467  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54468  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54469  __ret; \
54470})
54471#endif
54472
54473#ifdef __LITTLE_ENDIAN__
54474#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54475  int32x4_t __s0 = __p0; \
54476  int32x4_t __s1 = __p1; \
54477  int64x2_t __ret; \
54478  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54479  __ret; \
54480})
54481#else
54482#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54483  int32x4_t __s0 = __p0; \
54484  int32x4_t __s1 = __p1; \
54485  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54486  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54487  int64x2_t __ret; \
54488  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54490  __ret; \
54491})
54492#endif
54493
54494#ifdef __LITTLE_ENDIAN__
54495#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54496  int16x8_t __s0 = __p0; \
54497  int16x8_t __s1 = __p1; \
54498  int32x4_t __ret; \
54499  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54500  __ret; \
54501})
54502#else
54503#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54504  int16x8_t __s0 = __p0; \
54505  int16x8_t __s1 = __p1; \
54506  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54507  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54508  int32x4_t __ret; \
54509  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54510  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54511  __ret; \
54512})
54513#endif
54514
54515#ifdef __LITTLE_ENDIAN__
54516__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54517  uint64x2_t __ret;
54518  __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
54519  return __ret;
54520}
54521#else
54522__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54523  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54524  uint64x2_t __ret;
54525  __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
54526  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54527  return __ret;
54528}
54529#endif
54530
54531#ifdef __LITTLE_ENDIAN__
54532__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54533  uint32x4_t __ret;
54534  __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
54535  return __ret;
54536}
54537#else
54538__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54539  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54540  uint32x4_t __ret;
54541  __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
54542  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54543  return __ret;
54544}
54545#endif
54546
54547#ifdef __LITTLE_ENDIAN__
54548__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54549  int64x2_t __ret;
54550  __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
54551  return __ret;
54552}
54553#else
54554__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54555  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54556  int64x2_t __ret;
54557  __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
54558  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54559  return __ret;
54560}
54561#endif
54562
54563#ifdef __LITTLE_ENDIAN__
54564__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54565  int32x4_t __ret;
54566  __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
54567  return __ret;
54568}
54569#else
54570__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54571  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54572  int32x4_t __ret;
54573  __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
54574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54575  return __ret;
54576}
54577#endif
54578
54579#ifdef __LITTLE_ENDIAN__
54580#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54581  uint32x2_t __s0 = __p0; \
54582  uint32x4_t __s1 = __p1; \
54583  uint64x2_t __ret; \
54584  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54585  __ret; \
54586})
54587#else
54588#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54589  uint32x2_t __s0 = __p0; \
54590  uint32x4_t __s1 = __p1; \
54591  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54592  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54593  uint64x2_t __ret; \
54594  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54595  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54596  __ret; \
54597})
54598#endif
54599
54600#ifdef __LITTLE_ENDIAN__
54601#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54602  uint16x4_t __s0 = __p0; \
54603  uint16x8_t __s1 = __p1; \
54604  uint32x4_t __ret; \
54605  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54606  __ret; \
54607})
54608#else
54609#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54610  uint16x4_t __s0 = __p0; \
54611  uint16x8_t __s1 = __p1; \
54612  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54613  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54614  uint32x4_t __ret; \
54615  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54616  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54617  __ret; \
54618})
54619#endif
54620
54621#ifdef __LITTLE_ENDIAN__
54622#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54623  int32x2_t __s0 = __p0; \
54624  int32x4_t __s1 = __p1; \
54625  int64x2_t __ret; \
54626  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54627  __ret; \
54628})
54629#else
54630#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54631  int32x2_t __s0 = __p0; \
54632  int32x4_t __s1 = __p1; \
54633  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54634  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54635  int64x2_t __ret; \
54636  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54637  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54638  __ret; \
54639})
54640#endif
54641
54642#ifdef __LITTLE_ENDIAN__
54643#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54644  int16x4_t __s0 = __p0; \
54645  int16x8_t __s1 = __p1; \
54646  int32x4_t __ret; \
54647  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54648  __ret; \
54649})
54650#else
54651#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54652  int16x4_t __s0 = __p0; \
54653  int16x8_t __s1 = __p1; \
54654  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54655  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54656  int32x4_t __ret; \
54657  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54658  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54659  __ret; \
54660})
54661#endif
54662
54663#ifdef __LITTLE_ENDIAN__
54664__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54665  float64x2_t __ret;
54666  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54667  return __ret;
54668}
54669#else
54670__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54671  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54672  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54673  float64x2_t __ret;
54674  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54675  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54676  return __ret;
54677}
54678__ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54679  float64x2_t __ret;
54680  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54681  return __ret;
54682}
54683#endif
54684
54685#ifdef __LITTLE_ENDIAN__
54686__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54687  float32x4_t __ret;
54688  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54689  return __ret;
54690}
54691#else
54692__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54693  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54694  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54695  float32x4_t __ret;
54696  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54698  return __ret;
54699}
54700__ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54701  float32x4_t __ret;
54702  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54703  return __ret;
54704}
54705#endif
54706
54707#ifdef __LITTLE_ENDIAN__
54708__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54709  float64x1_t __ret;
54710  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54711  return __ret;
54712}
54713#else
54714__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54715  float64x1_t __ret;
54716  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54717  return __ret;
54718}
54719#endif
54720
54721#ifdef __LITTLE_ENDIAN__
54722__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54723  float32x2_t __ret;
54724  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54725  return __ret;
54726}
54727#else
54728__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54729  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54730  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54731  float32x2_t __ret;
54732  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54733  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54734  return __ret;
54735}
54736__ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54737  float32x2_t __ret;
54738  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54739  return __ret;
54740}
54741#endif
54742
54743#ifdef __LITTLE_ENDIAN__
54744__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54745  float64_t __ret;
54746  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54747  return __ret;
54748}
54749#else
54750__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54751  float64_t __ret;
54752  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54753  return __ret;
54754}
54755__ai float64_t __noswap_vmulxd_f64(float64_t __p0, float64_t __p1) {
54756  float64_t __ret;
54757  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54758  return __ret;
54759}
54760#endif
54761
54762#ifdef __LITTLE_ENDIAN__
54763__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54764  float32_t __ret;
54765  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54766  return __ret;
54767}
54768#else
54769__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54770  float32_t __ret;
54771  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54772  return __ret;
54773}
54774__ai float32_t __noswap_vmulxs_f32(float32_t __p0, float32_t __p1) {
54775  float32_t __ret;
54776  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54777  return __ret;
54778}
54779#endif
54780
54781#ifdef __LITTLE_ENDIAN__
54782#define vmulxd_lane_f64(__p0_138, __p1_138, __p2_138) __extension__ ({ \
54783  float64_t __s0_138 = __p0_138; \
54784  float64x1_t __s1_138 = __p1_138; \
54785  float64_t __ret_138; \
54786  __ret_138 = vmulxd_f64(__s0_138, vget_lane_f64(__s1_138, __p2_138)); \
54787  __ret_138; \
54788})
54789#else
54790#define vmulxd_lane_f64(__p0_139, __p1_139, __p2_139) __extension__ ({ \
54791  float64_t __s0_139 = __p0_139; \
54792  float64x1_t __s1_139 = __p1_139; \
54793  float64_t __ret_139; \
54794  __ret_139 = __noswap_vmulxd_f64(__s0_139, __noswap_vget_lane_f64(__s1_139, __p2_139)); \
54795  __ret_139; \
54796})
54797#endif
54798
54799#ifdef __LITTLE_ENDIAN__
54800#define vmulxs_lane_f32(__p0_140, __p1_140, __p2_140) __extension__ ({ \
54801  float32_t __s0_140 = __p0_140; \
54802  float32x2_t __s1_140 = __p1_140; \
54803  float32_t __ret_140; \
54804  __ret_140 = vmulxs_f32(__s0_140, vget_lane_f32(__s1_140, __p2_140)); \
54805  __ret_140; \
54806})
54807#else
54808#define vmulxs_lane_f32(__p0_141, __p1_141, __p2_141) __extension__ ({ \
54809  float32_t __s0_141 = __p0_141; \
54810  float32x2_t __s1_141 = __p1_141; \
54811  float32x2_t __rev1_141;  __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 1, 0); \
54812  float32_t __ret_141; \
54813  __ret_141 = __noswap_vmulxs_f32(__s0_141, __noswap_vget_lane_f32(__rev1_141, __p2_141)); \
54814  __ret_141; \
54815})
54816#endif
54817
54818#ifdef __LITTLE_ENDIAN__
54819#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54820  float64x2_t __s0 = __p0; \
54821  float64x1_t __s1 = __p1; \
54822  float64x2_t __ret; \
54823  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54824  __ret; \
54825})
54826#else
54827#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54828  float64x2_t __s0 = __p0; \
54829  float64x1_t __s1 = __p1; \
54830  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54831  float64x2_t __ret; \
54832  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54833  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54834  __ret; \
54835})
54836#endif
54837
54838#ifdef __LITTLE_ENDIAN__
54839#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54840  float32x4_t __s0 = __p0; \
54841  float32x2_t __s1 = __p1; \
54842  float32x4_t __ret; \
54843  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54844  __ret; \
54845})
54846#else
54847#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54848  float32x4_t __s0 = __p0; \
54849  float32x2_t __s1 = __p1; \
54850  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54851  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54852  float32x4_t __ret; \
54853  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54855  __ret; \
54856})
54857#endif
54858
54859#ifdef __LITTLE_ENDIAN__
54860#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54861  float32x2_t __s0 = __p0; \
54862  float32x2_t __s1 = __p1; \
54863  float32x2_t __ret; \
54864  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54865  __ret; \
54866})
54867#else
54868#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54869  float32x2_t __s0 = __p0; \
54870  float32x2_t __s1 = __p1; \
54871  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54872  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54873  float32x2_t __ret; \
54874  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54876  __ret; \
54877})
54878#endif
54879
54880#ifdef __LITTLE_ENDIAN__
54881#define vmulxd_laneq_f64(__p0_142, __p1_142, __p2_142) __extension__ ({ \
54882  float64_t __s0_142 = __p0_142; \
54883  float64x2_t __s1_142 = __p1_142; \
54884  float64_t __ret_142; \
54885  __ret_142 = vmulxd_f64(__s0_142, vgetq_lane_f64(__s1_142, __p2_142)); \
54886  __ret_142; \
54887})
54888#else
54889#define vmulxd_laneq_f64(__p0_143, __p1_143, __p2_143) __extension__ ({ \
54890  float64_t __s0_143 = __p0_143; \
54891  float64x2_t __s1_143 = __p1_143; \
54892  float64x2_t __rev1_143;  __rev1_143 = __builtin_shufflevector(__s1_143, __s1_143, 1, 0); \
54893  float64_t __ret_143; \
54894  __ret_143 = __noswap_vmulxd_f64(__s0_143, __noswap_vgetq_lane_f64(__rev1_143, __p2_143)); \
54895  __ret_143; \
54896})
54897#endif
54898
54899#ifdef __LITTLE_ENDIAN__
54900#define vmulxs_laneq_f32(__p0_144, __p1_144, __p2_144) __extension__ ({ \
54901  float32_t __s0_144 = __p0_144; \
54902  float32x4_t __s1_144 = __p1_144; \
54903  float32_t __ret_144; \
54904  __ret_144 = vmulxs_f32(__s0_144, vgetq_lane_f32(__s1_144, __p2_144)); \
54905  __ret_144; \
54906})
54907#else
54908#define vmulxs_laneq_f32(__p0_145, __p1_145, __p2_145) __extension__ ({ \
54909  float32_t __s0_145 = __p0_145; \
54910  float32x4_t __s1_145 = __p1_145; \
54911  float32x4_t __rev1_145;  __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 3, 2, 1, 0); \
54912  float32_t __ret_145; \
54913  __ret_145 = __noswap_vmulxs_f32(__s0_145, __noswap_vgetq_lane_f32(__rev1_145, __p2_145)); \
54914  __ret_145; \
54915})
54916#endif
54917
54918#ifdef __LITTLE_ENDIAN__
54919#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54920  float64x2_t __s0 = __p0; \
54921  float64x2_t __s1 = __p1; \
54922  float64x2_t __ret; \
54923  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54924  __ret; \
54925})
54926#else
54927#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54928  float64x2_t __s0 = __p0; \
54929  float64x2_t __s1 = __p1; \
54930  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54931  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54932  float64x2_t __ret; \
54933  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54934  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54935  __ret; \
54936})
54937#endif
54938
54939#ifdef __LITTLE_ENDIAN__
54940#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54941  float32x4_t __s0 = __p0; \
54942  float32x4_t __s1 = __p1; \
54943  float32x4_t __ret; \
54944  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54945  __ret; \
54946})
54947#else
54948#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54949  float32x4_t __s0 = __p0; \
54950  float32x4_t __s1 = __p1; \
54951  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54952  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54953  float32x4_t __ret; \
54954  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54956  __ret; \
54957})
54958#endif
54959
54960#ifdef __LITTLE_ENDIAN__
54961#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54962  float32x2_t __s0 = __p0; \
54963  float32x4_t __s1 = __p1; \
54964  float32x2_t __ret; \
54965  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54966  __ret; \
54967})
54968#else
54969#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54970  float32x2_t __s0 = __p0; \
54971  float32x4_t __s1 = __p1; \
54972  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54973  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54974  float32x2_t __ret; \
54975  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54976  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54977  __ret; \
54978})
54979#endif
54980
54981#ifdef __LITTLE_ENDIAN__
54982__ai float64x2_t vnegq_f64(float64x2_t __p0) {
54983  float64x2_t __ret;
54984  __ret = -__p0;
54985  return __ret;
54986}
54987#else
54988__ai float64x2_t vnegq_f64(float64x2_t __p0) {
54989  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54990  float64x2_t __ret;
54991  __ret = -__rev0;
54992  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54993  return __ret;
54994}
54995#endif
54996
54997#ifdef __LITTLE_ENDIAN__
54998__ai int64x2_t vnegq_s64(int64x2_t __p0) {
54999  int64x2_t __ret;
55000  __ret = -__p0;
55001  return __ret;
55002}
55003#else
55004__ai int64x2_t vnegq_s64(int64x2_t __p0) {
55005  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55006  int64x2_t __ret;
55007  __ret = -__rev0;
55008  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55009  return __ret;
55010}
55011#endif
55012
55013#ifdef __LITTLE_ENDIAN__
55014__ai float64x1_t vneg_f64(float64x1_t __p0) {
55015  float64x1_t __ret;
55016  __ret = -__p0;
55017  return __ret;
55018}
55019#else
55020__ai float64x1_t vneg_f64(float64x1_t __p0) {
55021  float64x1_t __ret;
55022  __ret = -__p0;
55023  return __ret;
55024}
55025#endif
55026
55027#ifdef __LITTLE_ENDIAN__
55028__ai int64x1_t vneg_s64(int64x1_t __p0) {
55029  int64x1_t __ret;
55030  __ret = -__p0;
55031  return __ret;
55032}
55033#else
55034__ai int64x1_t vneg_s64(int64x1_t __p0) {
55035  int64x1_t __ret;
55036  __ret = -__p0;
55037  return __ret;
55038}
55039#endif
55040
55041#ifdef __LITTLE_ENDIAN__
55042__ai int64_t vnegd_s64(int64_t __p0) {
55043  int64_t __ret;
55044  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55045  return __ret;
55046}
55047#else
55048__ai int64_t vnegd_s64(int64_t __p0) {
55049  int64_t __ret;
55050  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55051  return __ret;
55052}
55053#endif
55054
55055#ifdef __LITTLE_ENDIAN__
55056__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55057  uint8x16_t __ret;
55058  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55059  return __ret;
55060}
55061#else
55062__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55063  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55064  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55065  uint8x16_t __ret;
55066  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55067  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55068  return __ret;
55069}
55070#endif
55071
55072#ifdef __LITTLE_ENDIAN__
55073__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55074  uint32x4_t __ret;
55075  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55076  return __ret;
55077}
55078#else
55079__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55080  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55081  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55082  uint32x4_t __ret;
55083  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55084  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55085  return __ret;
55086}
55087#endif
55088
55089#ifdef __LITTLE_ENDIAN__
55090__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55091  uint64x2_t __ret;
55092  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
55093  return __ret;
55094}
55095#else
55096__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55097  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55098  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55099  uint64x2_t __ret;
55100  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
55101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55102  return __ret;
55103}
55104#endif
55105
55106#ifdef __LITTLE_ENDIAN__
55107__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55108  uint16x8_t __ret;
55109  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55110  return __ret;
55111}
55112#else
55113__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55114  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55115  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55116  uint16x8_t __ret;
55117  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55119  return __ret;
55120}
55121#endif
55122
55123#ifdef __LITTLE_ENDIAN__
55124__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55125  int8x16_t __ret;
55126  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55127  return __ret;
55128}
55129#else
55130__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55131  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55132  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55133  int8x16_t __ret;
55134  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55135  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55136  return __ret;
55137}
55138#endif
55139
55140#ifdef __LITTLE_ENDIAN__
55141__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55142  float64x2_t __ret;
55143  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55144  return __ret;
55145}
55146#else
55147__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55148  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55149  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55150  float64x2_t __ret;
55151  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55152  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55153  return __ret;
55154}
55155#endif
55156
55157#ifdef __LITTLE_ENDIAN__
55158__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55159  float32x4_t __ret;
55160  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55161  return __ret;
55162}
55163#else
55164__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55165  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55166  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55167  float32x4_t __ret;
55168  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55169  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55170  return __ret;
55171}
55172#endif
55173
55174#ifdef __LITTLE_ENDIAN__
55175__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55176  int32x4_t __ret;
55177  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55178  return __ret;
55179}
55180#else
55181__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55182  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55183  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55184  int32x4_t __ret;
55185  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55186  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55187  return __ret;
55188}
55189#endif
55190
55191#ifdef __LITTLE_ENDIAN__
55192__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55193  int64x2_t __ret;
55194  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
55195  return __ret;
55196}
55197#else
55198__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55199  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55200  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55201  int64x2_t __ret;
55202  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
55203  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55204  return __ret;
55205}
55206#endif
55207
55208#ifdef __LITTLE_ENDIAN__
55209__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55210  int16x8_t __ret;
55211  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55212  return __ret;
55213}
55214#else
55215__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55216  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55217  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55218  int16x8_t __ret;
55219  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55220  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55221  return __ret;
55222}
55223#endif
55224
55225#ifdef __LITTLE_ENDIAN__
55226__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55227  uint64_t __ret;
55228  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__p0);
55229  return __ret;
55230}
55231#else
55232__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55233  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55234  uint64_t __ret;
55235  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__rev0);
55236  return __ret;
55237}
55238#endif
55239
55240#ifdef __LITTLE_ENDIAN__
55241__ai float64_t vpaddd_f64(float64x2_t __p0) {
55242  float64_t __ret;
55243  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__p0);
55244  return __ret;
55245}
55246#else
55247__ai float64_t vpaddd_f64(float64x2_t __p0) {
55248  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55249  float64_t __ret;
55250  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__rev0);
55251  return __ret;
55252}
55253#endif
55254
55255#ifdef __LITTLE_ENDIAN__
55256__ai int64_t vpaddd_s64(int64x2_t __p0) {
55257  int64_t __ret;
55258  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__p0);
55259  return __ret;
55260}
55261#else
55262__ai int64_t vpaddd_s64(int64x2_t __p0) {
55263  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55264  int64_t __ret;
55265  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__rev0);
55266  return __ret;
55267}
55268#endif
55269
55270#ifdef __LITTLE_ENDIAN__
55271__ai float32_t vpadds_f32(float32x2_t __p0) {
55272  float32_t __ret;
55273  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__p0);
55274  return __ret;
55275}
55276#else
55277__ai float32_t vpadds_f32(float32x2_t __p0) {
55278  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55279  float32_t __ret;
55280  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__rev0);
55281  return __ret;
55282}
55283#endif
55284
55285#ifdef __LITTLE_ENDIAN__
55286__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55287  uint8x16_t __ret;
55288  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55289  return __ret;
55290}
55291#else
55292__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55293  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55294  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55295  uint8x16_t __ret;
55296  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55297  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55298  return __ret;
55299}
55300#endif
55301
55302#ifdef __LITTLE_ENDIAN__
55303__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55304  uint32x4_t __ret;
55305  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55306  return __ret;
55307}
55308#else
55309__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55310  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55311  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55312  uint32x4_t __ret;
55313  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55315  return __ret;
55316}
55317#endif
55318
55319#ifdef __LITTLE_ENDIAN__
55320__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55321  uint16x8_t __ret;
55322  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55323  return __ret;
55324}
55325#else
55326__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55327  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55328  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55329  uint16x8_t __ret;
55330  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55331  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55332  return __ret;
55333}
55334#endif
55335
55336#ifdef __LITTLE_ENDIAN__
55337__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55338  int8x16_t __ret;
55339  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55340  return __ret;
55341}
55342#else
55343__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55344  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55345  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55346  int8x16_t __ret;
55347  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55348  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55349  return __ret;
55350}
55351#endif
55352
55353#ifdef __LITTLE_ENDIAN__
55354__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55355  float64x2_t __ret;
55356  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55357  return __ret;
55358}
55359#else
55360__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55361  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55362  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55363  float64x2_t __ret;
55364  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55365  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55366  return __ret;
55367}
55368#endif
55369
55370#ifdef __LITTLE_ENDIAN__
55371__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55372  float32x4_t __ret;
55373  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55374  return __ret;
55375}
55376#else
55377__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55378  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55379  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55380  float32x4_t __ret;
55381  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55383  return __ret;
55384}
55385#endif
55386
55387#ifdef __LITTLE_ENDIAN__
55388__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55389  int32x4_t __ret;
55390  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55391  return __ret;
55392}
55393#else
55394__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55395  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55396  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55397  int32x4_t __ret;
55398  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55399  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55400  return __ret;
55401}
55402#endif
55403
55404#ifdef __LITTLE_ENDIAN__
55405__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55406  int16x8_t __ret;
55407  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55408  return __ret;
55409}
55410#else
55411__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55412  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55413  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55414  int16x8_t __ret;
55415  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55416  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55417  return __ret;
55418}
55419#endif
55420
55421#ifdef __LITTLE_ENDIAN__
55422__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55423  float64_t __ret;
55424  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__p0);
55425  return __ret;
55426}
55427#else
55428__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55429  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55430  float64_t __ret;
55431  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__rev0);
55432  return __ret;
55433}
55434#endif
55435
55436#ifdef __LITTLE_ENDIAN__
55437__ai float32_t vpmaxs_f32(float32x2_t __p0) {
55438  float32_t __ret;
55439  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__p0);
55440  return __ret;
55441}
55442#else
55443__ai float32_t vpmaxs_f32(float32x2_t __p0) {
55444  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55445  float32_t __ret;
55446  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__rev0);
55447  return __ret;
55448}
55449#endif
55450
55451#ifdef __LITTLE_ENDIAN__
55452__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55453  float64x2_t __ret;
55454  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55455  return __ret;
55456}
55457#else
55458__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55459  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55460  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55461  float64x2_t __ret;
55462  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55463  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55464  return __ret;
55465}
55466#endif
55467
55468#ifdef __LITTLE_ENDIAN__
55469__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55470  float32x4_t __ret;
55471  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55472  return __ret;
55473}
55474#else
55475__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55476  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55477  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55478  float32x4_t __ret;
55479  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55480  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55481  return __ret;
55482}
55483#endif
55484
55485#ifdef __LITTLE_ENDIAN__
55486__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55487  float32x2_t __ret;
55488  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55489  return __ret;
55490}
55491#else
55492__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55493  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55494  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55495  float32x2_t __ret;
55496  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55497  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55498  return __ret;
55499}
55500#endif
55501
55502#ifdef __LITTLE_ENDIAN__
55503__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55504  float64_t __ret;
55505  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__p0);
55506  return __ret;
55507}
55508#else
55509__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55510  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55511  float64_t __ret;
55512  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__rev0);
55513  return __ret;
55514}
55515#endif
55516
55517#ifdef __LITTLE_ENDIAN__
55518__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55519  float32_t __ret;
55520  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__p0);
55521  return __ret;
55522}
55523#else
55524__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55525  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55526  float32_t __ret;
55527  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__rev0);
55528  return __ret;
55529}
55530#endif
55531
55532#ifdef __LITTLE_ENDIAN__
55533__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55534  uint8x16_t __ret;
55535  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55536  return __ret;
55537}
55538#else
55539__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55540  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55541  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55542  uint8x16_t __ret;
55543  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55544  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55545  return __ret;
55546}
55547#endif
55548
55549#ifdef __LITTLE_ENDIAN__
55550__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55551  uint32x4_t __ret;
55552  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55553  return __ret;
55554}
55555#else
55556__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55557  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55558  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55559  uint32x4_t __ret;
55560  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55561  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55562  return __ret;
55563}
55564#endif
55565
55566#ifdef __LITTLE_ENDIAN__
55567__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55568  uint16x8_t __ret;
55569  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55570  return __ret;
55571}
55572#else
55573__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55574  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55575  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55576  uint16x8_t __ret;
55577  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55578  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55579  return __ret;
55580}
55581#endif
55582
55583#ifdef __LITTLE_ENDIAN__
55584__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55585  int8x16_t __ret;
55586  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55587  return __ret;
55588}
55589#else
55590__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55591  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55592  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55593  int8x16_t __ret;
55594  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55595  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55596  return __ret;
55597}
55598#endif
55599
55600#ifdef __LITTLE_ENDIAN__
55601__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55602  float64x2_t __ret;
55603  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55604  return __ret;
55605}
55606#else
55607__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55608  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55609  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55610  float64x2_t __ret;
55611  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55612  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55613  return __ret;
55614}
55615#endif
55616
55617#ifdef __LITTLE_ENDIAN__
55618__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55619  float32x4_t __ret;
55620  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55621  return __ret;
55622}
55623#else
55624__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55625  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55626  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55627  float32x4_t __ret;
55628  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55630  return __ret;
55631}
55632#endif
55633
55634#ifdef __LITTLE_ENDIAN__
55635__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55636  int32x4_t __ret;
55637  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55638  return __ret;
55639}
55640#else
55641__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55642  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55643  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55644  int32x4_t __ret;
55645  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55646  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55647  return __ret;
55648}
55649#endif
55650
55651#ifdef __LITTLE_ENDIAN__
55652__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55653  int16x8_t __ret;
55654  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55655  return __ret;
55656}
55657#else
55658__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55659  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55660  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55661  int16x8_t __ret;
55662  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55663  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55664  return __ret;
55665}
55666#endif
55667
55668#ifdef __LITTLE_ENDIAN__
55669__ai float64_t vpminqd_f64(float64x2_t __p0) {
55670  float64_t __ret;
55671  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__p0);
55672  return __ret;
55673}
55674#else
55675__ai float64_t vpminqd_f64(float64x2_t __p0) {
55676  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55677  float64_t __ret;
55678  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__rev0);
55679  return __ret;
55680}
55681#endif
55682
55683#ifdef __LITTLE_ENDIAN__
55684__ai float32_t vpmins_f32(float32x2_t __p0) {
55685  float32_t __ret;
55686  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__p0);
55687  return __ret;
55688}
55689#else
55690__ai float32_t vpmins_f32(float32x2_t __p0) {
55691  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55692  float32_t __ret;
55693  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__rev0);
55694  return __ret;
55695}
55696#endif
55697
55698#ifdef __LITTLE_ENDIAN__
55699__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55700  float64x2_t __ret;
55701  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55702  return __ret;
55703}
55704#else
55705__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55706  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55707  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55708  float64x2_t __ret;
55709  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55710  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55711  return __ret;
55712}
55713#endif
55714
55715#ifdef __LITTLE_ENDIAN__
55716__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55717  float32x4_t __ret;
55718  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55719  return __ret;
55720}
55721#else
55722__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55723  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55724  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55725  float32x4_t __ret;
55726  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55727  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55728  return __ret;
55729}
55730#endif
55731
55732#ifdef __LITTLE_ENDIAN__
55733__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55734  float32x2_t __ret;
55735  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55736  return __ret;
55737}
55738#else
55739__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55740  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55741  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55742  float32x2_t __ret;
55743  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55744  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55745  return __ret;
55746}
55747#endif
55748
55749#ifdef __LITTLE_ENDIAN__
55750__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55751  float64_t __ret;
55752  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__p0);
55753  return __ret;
55754}
55755#else
55756__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55757  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55758  float64_t __ret;
55759  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__rev0);
55760  return __ret;
55761}
55762#endif
55763
55764#ifdef __LITTLE_ENDIAN__
55765__ai float32_t vpminnms_f32(float32x2_t __p0) {
55766  float32_t __ret;
55767  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__p0);
55768  return __ret;
55769}
55770#else
55771__ai float32_t vpminnms_f32(float32x2_t __p0) {
55772  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55773  float32_t __ret;
55774  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__rev0);
55775  return __ret;
55776}
55777#endif
55778
55779#ifdef __LITTLE_ENDIAN__
55780__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55781  int64x2_t __ret;
55782  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
55783  return __ret;
55784}
55785#else
55786__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55787  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55788  int64x2_t __ret;
55789  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
55790  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55791  return __ret;
55792}
55793#endif
55794
55795#ifdef __LITTLE_ENDIAN__
55796__ai int64x1_t vqabs_s64(int64x1_t __p0) {
55797  int64x1_t __ret;
55798  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55799  return __ret;
55800}
55801#else
55802__ai int64x1_t vqabs_s64(int64x1_t __p0) {
55803  int64x1_t __ret;
55804  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55805  return __ret;
55806}
55807#endif
55808
55809#ifdef __LITTLE_ENDIAN__
55810__ai int8_t vqabsb_s8(int8_t __p0) {
55811  int8_t __ret;
55812  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55813  return __ret;
55814}
55815#else
55816__ai int8_t vqabsb_s8(int8_t __p0) {
55817  int8_t __ret;
55818  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55819  return __ret;
55820}
55821#endif
55822
55823#ifdef __LITTLE_ENDIAN__
55824__ai int32_t vqabss_s32(int32_t __p0) {
55825  int32_t __ret;
55826  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55827  return __ret;
55828}
55829#else
55830__ai int32_t vqabss_s32(int32_t __p0) {
55831  int32_t __ret;
55832  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55833  return __ret;
55834}
55835#endif
55836
55837#ifdef __LITTLE_ENDIAN__
55838__ai int64_t vqabsd_s64(int64_t __p0) {
55839  int64_t __ret;
55840  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55841  return __ret;
55842}
55843#else
55844__ai int64_t vqabsd_s64(int64_t __p0) {
55845  int64_t __ret;
55846  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55847  return __ret;
55848}
55849#endif
55850
55851#ifdef __LITTLE_ENDIAN__
55852__ai int16_t vqabsh_s16(int16_t __p0) {
55853  int16_t __ret;
55854  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55855  return __ret;
55856}
55857#else
55858__ai int16_t vqabsh_s16(int16_t __p0) {
55859  int16_t __ret;
55860  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55861  return __ret;
55862}
55863#endif
55864
55865#ifdef __LITTLE_ENDIAN__
55866__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55867  uint8_t __ret;
55868  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55869  return __ret;
55870}
55871#else
55872__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55873  uint8_t __ret;
55874  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55875  return __ret;
55876}
55877#endif
55878
55879#ifdef __LITTLE_ENDIAN__
55880__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55881  uint32_t __ret;
55882  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55883  return __ret;
55884}
55885#else
55886__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55887  uint32_t __ret;
55888  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55889  return __ret;
55890}
55891#endif
55892
55893#ifdef __LITTLE_ENDIAN__
55894__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55895  uint64_t __ret;
55896  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55897  return __ret;
55898}
55899#else
55900__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55901  uint64_t __ret;
55902  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55903  return __ret;
55904}
55905#endif
55906
55907#ifdef __LITTLE_ENDIAN__
55908__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55909  uint16_t __ret;
55910  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55911  return __ret;
55912}
55913#else
55914__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55915  uint16_t __ret;
55916  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55917  return __ret;
55918}
55919#endif
55920
55921#ifdef __LITTLE_ENDIAN__
55922__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55923  int8_t __ret;
55924  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55925  return __ret;
55926}
55927#else
55928__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55929  int8_t __ret;
55930  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55931  return __ret;
55932}
55933#endif
55934
55935#ifdef __LITTLE_ENDIAN__
55936__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55937  int32_t __ret;
55938  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55939  return __ret;
55940}
55941#else
55942__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55943  int32_t __ret;
55944  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55945  return __ret;
55946}
55947__ai int32_t __noswap_vqadds_s32(int32_t __p0, int32_t __p1) {
55948  int32_t __ret;
55949  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55950  return __ret;
55951}
55952#endif
55953
55954#ifdef __LITTLE_ENDIAN__
55955__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55956  int64_t __ret;
55957  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55958  return __ret;
55959}
55960#else
55961__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55962  int64_t __ret;
55963  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55964  return __ret;
55965}
55966#endif
55967
55968#ifdef __LITTLE_ENDIAN__
55969__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55970  int16_t __ret;
55971  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55972  return __ret;
55973}
55974#else
55975__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55976  int16_t __ret;
55977  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55978  return __ret;
55979}
55980__ai int16_t __noswap_vqaddh_s16(int16_t __p0, int16_t __p1) {
55981  int16_t __ret;
55982  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55983  return __ret;
55984}
55985#endif
55986
55987#ifdef __LITTLE_ENDIAN__
55988__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55989  int64_t __ret;
55990  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
55991  return __ret;
55992}
55993#else
55994__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55995  int64_t __ret;
55996  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
55997  return __ret;
55998}
55999#endif
56000
56001#ifdef __LITTLE_ENDIAN__
56002__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56003  int32_t __ret;
56004  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56005  return __ret;
56006}
56007#else
56008__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56009  int32_t __ret;
56010  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56011  return __ret;
56012}
56013#endif
56014
56015#ifdef __LITTLE_ENDIAN__
56016__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56017  int64x2_t __ret;
56018  __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56019  return __ret;
56020}
56021#else
56022__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56023  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56024  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56025  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56026  int64x2_t __ret;
56027  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56028  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56029  return __ret;
56030}
56031#endif
56032
56033#ifdef __LITTLE_ENDIAN__
56034__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56035  int32x4_t __ret;
56036  __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56037  return __ret;
56038}
56039#else
56040__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56041  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56042  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56043  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56044  int32x4_t __ret;
56045  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56046  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56047  return __ret;
56048}
56049#endif
56050
56051#ifdef __LITTLE_ENDIAN__
56052#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56053  int64x2_t __s0 = __p0; \
56054  int32x4_t __s1 = __p1; \
56055  int32x2_t __s2 = __p2; \
56056  int64x2_t __ret; \
56057  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56058  __ret; \
56059})
56060#else
56061#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56062  int64x2_t __s0 = __p0; \
56063  int32x4_t __s1 = __p1; \
56064  int32x2_t __s2 = __p2; \
56065  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56066  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56067  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56068  int64x2_t __ret; \
56069  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56070  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56071  __ret; \
56072})
56073#endif
56074
56075#ifdef __LITTLE_ENDIAN__
56076#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56077  int32x4_t __s0 = __p0; \
56078  int16x8_t __s1 = __p1; \
56079  int16x4_t __s2 = __p2; \
56080  int32x4_t __ret; \
56081  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56082  __ret; \
56083})
56084#else
56085#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56086  int32x4_t __s0 = __p0; \
56087  int16x8_t __s1 = __p1; \
56088  int16x4_t __s2 = __p2; \
56089  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56090  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56091  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56092  int32x4_t __ret; \
56093  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56094  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56095  __ret; \
56096})
56097#endif
56098
56099#ifdef __LITTLE_ENDIAN__
56100#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56101  int64x2_t __s0 = __p0; \
56102  int32x4_t __s1 = __p1; \
56103  int32x4_t __s2 = __p2; \
56104  int64x2_t __ret; \
56105  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56106  __ret; \
56107})
56108#else
56109#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56110  int64x2_t __s0 = __p0; \
56111  int32x4_t __s1 = __p1; \
56112  int32x4_t __s2 = __p2; \
56113  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56114  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56115  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56116  int64x2_t __ret; \
56117  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56118  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56119  __ret; \
56120})
56121#endif
56122
56123#ifdef __LITTLE_ENDIAN__
56124#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56125  int32x4_t __s0 = __p0; \
56126  int16x8_t __s1 = __p1; \
56127  int16x8_t __s2 = __p2; \
56128  int32x4_t __ret; \
56129  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56130  __ret; \
56131})
56132#else
56133#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56134  int32x4_t __s0 = __p0; \
56135  int16x8_t __s1 = __p1; \
56136  int16x8_t __s2 = __p2; \
56137  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56138  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56139  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56140  int32x4_t __ret; \
56141  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56142  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56143  __ret; \
56144})
56145#endif
56146
56147#ifdef __LITTLE_ENDIAN__
56148__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56149  int64x2_t __ret;
56150  __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
56151  return __ret;
56152}
56153#else
56154__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56155  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56156  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56157  int64x2_t __ret;
56158  __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56159  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56160  return __ret;
56161}
56162#endif
56163
56164#ifdef __LITTLE_ENDIAN__
56165__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56166  int32x4_t __ret;
56167  __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
56168  return __ret;
56169}
56170#else
56171__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56172  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56173  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56174  int32x4_t __ret;
56175  __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56176  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56177  return __ret;
56178}
56179#endif
56180
56181#ifdef __LITTLE_ENDIAN__
56182#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56183  int64_t __s0 = __p0; \
56184  int32_t __s1 = __p1; \
56185  int32x2_t __s2 = __p2; \
56186  int64_t __ret; \
56187  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56188  __ret; \
56189})
56190#else
56191#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56192  int64_t __s0 = __p0; \
56193  int32_t __s1 = __p1; \
56194  int32x2_t __s2 = __p2; \
56195  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56196  int64_t __ret; \
56197  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56198  __ret; \
56199})
56200#endif
56201
56202#ifdef __LITTLE_ENDIAN__
56203#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56204  int32_t __s0 = __p0; \
56205  int16_t __s1 = __p1; \
56206  int16x4_t __s2 = __p2; \
56207  int32_t __ret; \
56208  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56209  __ret; \
56210})
56211#else
56212#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56213  int32_t __s0 = __p0; \
56214  int16_t __s1 = __p1; \
56215  int16x4_t __s2 = __p2; \
56216  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56217  int32_t __ret; \
56218  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56219  __ret; \
56220})
56221#endif
56222
56223#ifdef __LITTLE_ENDIAN__
56224#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56225  int64_t __s0 = __p0; \
56226  int32_t __s1 = __p1; \
56227  int32x4_t __s2 = __p2; \
56228  int64_t __ret; \
56229  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56230  __ret; \
56231})
56232#else
56233#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56234  int64_t __s0 = __p0; \
56235  int32_t __s1 = __p1; \
56236  int32x4_t __s2 = __p2; \
56237  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56238  int64_t __ret; \
56239  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56240  __ret; \
56241})
56242#endif
56243
56244#ifdef __LITTLE_ENDIAN__
56245#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56246  int32_t __s0 = __p0; \
56247  int16_t __s1 = __p1; \
56248  int16x8_t __s2 = __p2; \
56249  int32_t __ret; \
56250  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56251  __ret; \
56252})
56253#else
56254#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56255  int32_t __s0 = __p0; \
56256  int16_t __s1 = __p1; \
56257  int16x8_t __s2 = __p2; \
56258  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56259  int32_t __ret; \
56260  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56261  __ret; \
56262})
56263#endif
56264
56265#ifdef __LITTLE_ENDIAN__
56266#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56267  int64x2_t __s0 = __p0; \
56268  int32x2_t __s1 = __p1; \
56269  int32x4_t __s2 = __p2; \
56270  int64x2_t __ret; \
56271  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56272  __ret; \
56273})
56274#else
56275#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56276  int64x2_t __s0 = __p0; \
56277  int32x2_t __s1 = __p1; \
56278  int32x4_t __s2 = __p2; \
56279  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56280  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56281  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56282  int64x2_t __ret; \
56283  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56284  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56285  __ret; \
56286})
56287#endif
56288
56289#ifdef __LITTLE_ENDIAN__
56290#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56291  int32x4_t __s0 = __p0; \
56292  int16x4_t __s1 = __p1; \
56293  int16x8_t __s2 = __p2; \
56294  int32x4_t __ret; \
56295  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56296  __ret; \
56297})
56298#else
56299#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56300  int32x4_t __s0 = __p0; \
56301  int16x4_t __s1 = __p1; \
56302  int16x8_t __s2 = __p2; \
56303  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56304  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56305  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56306  int32x4_t __ret; \
56307  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56308  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56309  __ret; \
56310})
56311#endif
56312
56313#ifdef __LITTLE_ENDIAN__
56314__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56315  int64_t __ret;
56316  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56317  return __ret;
56318}
56319#else
56320__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56321  int64_t __ret;
56322  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56323  return __ret;
56324}
56325#endif
56326
56327#ifdef __LITTLE_ENDIAN__
56328__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56329  int32_t __ret;
56330  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56331  return __ret;
56332}
56333#else
56334__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56335  int32_t __ret;
56336  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56337  return __ret;
56338}
56339#endif
56340
56341#ifdef __LITTLE_ENDIAN__
56342__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56343  int64x2_t __ret;
56344  __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56345  return __ret;
56346}
56347#else
56348__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56349  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56350  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56351  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56352  int64x2_t __ret;
56353  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56354  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56355  return __ret;
56356}
56357#endif
56358
56359#ifdef __LITTLE_ENDIAN__
56360__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56361  int32x4_t __ret;
56362  __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56363  return __ret;
56364}
56365#else
56366__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56367  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56368  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56369  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56370  int32x4_t __ret;
56371  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56373  return __ret;
56374}
56375#endif
56376
56377#ifdef __LITTLE_ENDIAN__
56378#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56379  int64x2_t __s0 = __p0; \
56380  int32x4_t __s1 = __p1; \
56381  int32x2_t __s2 = __p2; \
56382  int64x2_t __ret; \
56383  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56384  __ret; \
56385})
56386#else
56387#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56388  int64x2_t __s0 = __p0; \
56389  int32x4_t __s1 = __p1; \
56390  int32x2_t __s2 = __p2; \
56391  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56392  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56393  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56394  int64x2_t __ret; \
56395  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56396  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56397  __ret; \
56398})
56399#endif
56400
56401#ifdef __LITTLE_ENDIAN__
56402#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56403  int32x4_t __s0 = __p0; \
56404  int16x8_t __s1 = __p1; \
56405  int16x4_t __s2 = __p2; \
56406  int32x4_t __ret; \
56407  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56408  __ret; \
56409})
56410#else
56411#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56412  int32x4_t __s0 = __p0; \
56413  int16x8_t __s1 = __p1; \
56414  int16x4_t __s2 = __p2; \
56415  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56416  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56417  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56418  int32x4_t __ret; \
56419  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56420  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56421  __ret; \
56422})
56423#endif
56424
56425#ifdef __LITTLE_ENDIAN__
56426#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56427  int64x2_t __s0 = __p0; \
56428  int32x4_t __s1 = __p1; \
56429  int32x4_t __s2 = __p2; \
56430  int64x2_t __ret; \
56431  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56432  __ret; \
56433})
56434#else
56435#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56436  int64x2_t __s0 = __p0; \
56437  int32x4_t __s1 = __p1; \
56438  int32x4_t __s2 = __p2; \
56439  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56440  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56441  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56442  int64x2_t __ret; \
56443  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56444  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56445  __ret; \
56446})
56447#endif
56448
56449#ifdef __LITTLE_ENDIAN__
56450#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56451  int32x4_t __s0 = __p0; \
56452  int16x8_t __s1 = __p1; \
56453  int16x8_t __s2 = __p2; \
56454  int32x4_t __ret; \
56455  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56456  __ret; \
56457})
56458#else
56459#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56460  int32x4_t __s0 = __p0; \
56461  int16x8_t __s1 = __p1; \
56462  int16x8_t __s2 = __p2; \
56463  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56464  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56465  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56466  int32x4_t __ret; \
56467  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56468  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56469  __ret; \
56470})
56471#endif
56472
56473#ifdef __LITTLE_ENDIAN__
56474__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56475  int64x2_t __ret;
56476  __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
56477  return __ret;
56478}
56479#else
56480__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56481  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56482  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56483  int64x2_t __ret;
56484  __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56485  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56486  return __ret;
56487}
56488#endif
56489
56490#ifdef __LITTLE_ENDIAN__
56491__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56492  int32x4_t __ret;
56493  __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
56494  return __ret;
56495}
56496#else
56497__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56498  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56499  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56500  int32x4_t __ret;
56501  __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56502  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56503  return __ret;
56504}
56505#endif
56506
56507#ifdef __LITTLE_ENDIAN__
56508#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56509  int64_t __s0 = __p0; \
56510  int32_t __s1 = __p1; \
56511  int32x2_t __s2 = __p2; \
56512  int64_t __ret; \
56513  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56514  __ret; \
56515})
56516#else
56517#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56518  int64_t __s0 = __p0; \
56519  int32_t __s1 = __p1; \
56520  int32x2_t __s2 = __p2; \
56521  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56522  int64_t __ret; \
56523  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56524  __ret; \
56525})
56526#endif
56527
56528#ifdef __LITTLE_ENDIAN__
56529#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56530  int32_t __s0 = __p0; \
56531  int16_t __s1 = __p1; \
56532  int16x4_t __s2 = __p2; \
56533  int32_t __ret; \
56534  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56535  __ret; \
56536})
56537#else
56538#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56539  int32_t __s0 = __p0; \
56540  int16_t __s1 = __p1; \
56541  int16x4_t __s2 = __p2; \
56542  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56543  int32_t __ret; \
56544  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56545  __ret; \
56546})
56547#endif
56548
56549#ifdef __LITTLE_ENDIAN__
56550#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56551  int64_t __s0 = __p0; \
56552  int32_t __s1 = __p1; \
56553  int32x4_t __s2 = __p2; \
56554  int64_t __ret; \
56555  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56556  __ret; \
56557})
56558#else
56559#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56560  int64_t __s0 = __p0; \
56561  int32_t __s1 = __p1; \
56562  int32x4_t __s2 = __p2; \
56563  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56564  int64_t __ret; \
56565  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56566  __ret; \
56567})
56568#endif
56569
56570#ifdef __LITTLE_ENDIAN__
56571#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56572  int32_t __s0 = __p0; \
56573  int16_t __s1 = __p1; \
56574  int16x8_t __s2 = __p2; \
56575  int32_t __ret; \
56576  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56577  __ret; \
56578})
56579#else
56580#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56581  int32_t __s0 = __p0; \
56582  int16_t __s1 = __p1; \
56583  int16x8_t __s2 = __p2; \
56584  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56585  int32_t __ret; \
56586  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56587  __ret; \
56588})
56589#endif
56590
56591#ifdef __LITTLE_ENDIAN__
56592#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56593  int64x2_t __s0 = __p0; \
56594  int32x2_t __s1 = __p1; \
56595  int32x4_t __s2 = __p2; \
56596  int64x2_t __ret; \
56597  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56598  __ret; \
56599})
56600#else
56601#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56602  int64x2_t __s0 = __p0; \
56603  int32x2_t __s1 = __p1; \
56604  int32x4_t __s2 = __p2; \
56605  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56606  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56607  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56608  int64x2_t __ret; \
56609  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56610  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56611  __ret; \
56612})
56613#endif
56614
56615#ifdef __LITTLE_ENDIAN__
56616#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56617  int32x4_t __s0 = __p0; \
56618  int16x4_t __s1 = __p1; \
56619  int16x8_t __s2 = __p2; \
56620  int32x4_t __ret; \
56621  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56622  __ret; \
56623})
56624#else
56625#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56626  int32x4_t __s0 = __p0; \
56627  int16x4_t __s1 = __p1; \
56628  int16x8_t __s2 = __p2; \
56629  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56630  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56631  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56632  int32x4_t __ret; \
56633  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56635  __ret; \
56636})
56637#endif
56638
56639#ifdef __LITTLE_ENDIAN__
56640__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56641  int32_t __ret;
56642  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56643  return __ret;
56644}
56645#else
56646__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56647  int32_t __ret;
56648  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56649  return __ret;
56650}
56651__ai int32_t __noswap_vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56652  int32_t __ret;
56653  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56654  return __ret;
56655}
56656#endif
56657
56658#ifdef __LITTLE_ENDIAN__
56659__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56660  int16_t __ret;
56661  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56662  return __ret;
56663}
56664#else
56665__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56666  int16_t __ret;
56667  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56668  return __ret;
56669}
56670__ai int16_t __noswap_vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56671  int16_t __ret;
56672  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56673  return __ret;
56674}
56675#endif
56676
56677#ifdef __LITTLE_ENDIAN__
56678#define vqdmulhs_lane_s32(__p0_146, __p1_146, __p2_146) __extension__ ({ \
56679  int32_t __s0_146 = __p0_146; \
56680  int32x2_t __s1_146 = __p1_146; \
56681  int32_t __ret_146; \
56682  __ret_146 = vqdmulhs_s32(__s0_146, vget_lane_s32(__s1_146, __p2_146)); \
56683  __ret_146; \
56684})
56685#else
56686#define vqdmulhs_lane_s32(__p0_147, __p1_147, __p2_147) __extension__ ({ \
56687  int32_t __s0_147 = __p0_147; \
56688  int32x2_t __s1_147 = __p1_147; \
56689  int32x2_t __rev1_147;  __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 1, 0); \
56690  int32_t __ret_147; \
56691  __ret_147 = __noswap_vqdmulhs_s32(__s0_147, __noswap_vget_lane_s32(__rev1_147, __p2_147)); \
56692  __ret_147; \
56693})
56694#endif
56695
56696#ifdef __LITTLE_ENDIAN__
56697#define vqdmulhh_lane_s16(__p0_148, __p1_148, __p2_148) __extension__ ({ \
56698  int16_t __s0_148 = __p0_148; \
56699  int16x4_t __s1_148 = __p1_148; \
56700  int16_t __ret_148; \
56701  __ret_148 = vqdmulhh_s16(__s0_148, vget_lane_s16(__s1_148, __p2_148)); \
56702  __ret_148; \
56703})
56704#else
56705#define vqdmulhh_lane_s16(__p0_149, __p1_149, __p2_149) __extension__ ({ \
56706  int16_t __s0_149 = __p0_149; \
56707  int16x4_t __s1_149 = __p1_149; \
56708  int16x4_t __rev1_149;  __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
56709  int16_t __ret_149; \
56710  __ret_149 = __noswap_vqdmulhh_s16(__s0_149, __noswap_vget_lane_s16(__rev1_149, __p2_149)); \
56711  __ret_149; \
56712})
56713#endif
56714
56715#ifdef __LITTLE_ENDIAN__
56716#define vqdmulhs_laneq_s32(__p0_150, __p1_150, __p2_150) __extension__ ({ \
56717  int32_t __s0_150 = __p0_150; \
56718  int32x4_t __s1_150 = __p1_150; \
56719  int32_t __ret_150; \
56720  __ret_150 = vqdmulhs_s32(__s0_150, vgetq_lane_s32(__s1_150, __p2_150)); \
56721  __ret_150; \
56722})
56723#else
56724#define vqdmulhs_laneq_s32(__p0_151, __p1_151, __p2_151) __extension__ ({ \
56725  int32_t __s0_151 = __p0_151; \
56726  int32x4_t __s1_151 = __p1_151; \
56727  int32x4_t __rev1_151;  __rev1_151 = __builtin_shufflevector(__s1_151, __s1_151, 3, 2, 1, 0); \
56728  int32_t __ret_151; \
56729  __ret_151 = __noswap_vqdmulhs_s32(__s0_151, __noswap_vgetq_lane_s32(__rev1_151, __p2_151)); \
56730  __ret_151; \
56731})
56732#endif
56733
56734#ifdef __LITTLE_ENDIAN__
56735#define vqdmulhh_laneq_s16(__p0_152, __p1_152, __p2_152) __extension__ ({ \
56736  int16_t __s0_152 = __p0_152; \
56737  int16x8_t __s1_152 = __p1_152; \
56738  int16_t __ret_152; \
56739  __ret_152 = vqdmulhh_s16(__s0_152, vgetq_lane_s16(__s1_152, __p2_152)); \
56740  __ret_152; \
56741})
56742#else
56743#define vqdmulhh_laneq_s16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
56744  int16_t __s0_153 = __p0_153; \
56745  int16x8_t __s1_153 = __p1_153; \
56746  int16x8_t __rev1_153;  __rev1_153 = __builtin_shufflevector(__s1_153, __s1_153, 7, 6, 5, 4, 3, 2, 1, 0); \
56747  int16_t __ret_153; \
56748  __ret_153 = __noswap_vqdmulhh_s16(__s0_153, __noswap_vgetq_lane_s16(__rev1_153, __p2_153)); \
56749  __ret_153; \
56750})
56751#endif
56752
56753#ifdef __LITTLE_ENDIAN__
56754#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56755  int32x4_t __s0 = __p0; \
56756  int32x4_t __s1 = __p1; \
56757  int32x4_t __ret; \
56758  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56759  __ret; \
56760})
56761#else
56762#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56763  int32x4_t __s0 = __p0; \
56764  int32x4_t __s1 = __p1; \
56765  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56766  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56767  int32x4_t __ret; \
56768  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56770  __ret; \
56771})
56772#endif
56773
56774#ifdef __LITTLE_ENDIAN__
56775#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56776  int16x8_t __s0 = __p0; \
56777  int16x8_t __s1 = __p1; \
56778  int16x8_t __ret; \
56779  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56780  __ret; \
56781})
56782#else
56783#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56784  int16x8_t __s0 = __p0; \
56785  int16x8_t __s1 = __p1; \
56786  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56787  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56788  int16x8_t __ret; \
56789  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56790  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56791  __ret; \
56792})
56793#endif
56794
56795#ifdef __LITTLE_ENDIAN__
56796#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56797  int32x2_t __s0 = __p0; \
56798  int32x4_t __s1 = __p1; \
56799  int32x2_t __ret; \
56800  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56801  __ret; \
56802})
56803#else
56804#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56805  int32x2_t __s0 = __p0; \
56806  int32x4_t __s1 = __p1; \
56807  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56808  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56809  int32x2_t __ret; \
56810  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56812  __ret; \
56813})
56814#endif
56815
56816#ifdef __LITTLE_ENDIAN__
56817#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56818  int16x4_t __s0 = __p0; \
56819  int16x8_t __s1 = __p1; \
56820  int16x4_t __ret; \
56821  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56822  __ret; \
56823})
56824#else
56825#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56826  int16x4_t __s0 = __p0; \
56827  int16x8_t __s1 = __p1; \
56828  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56829  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56830  int16x4_t __ret; \
56831  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56832  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56833  __ret; \
56834})
56835#endif
56836
56837#ifdef __LITTLE_ENDIAN__
56838__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56839  int64_t __ret;
56840  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56841  return __ret;
56842}
56843#else
56844__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56845  int64_t __ret;
56846  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56847  return __ret;
56848}
56849__ai int64_t __noswap_vqdmulls_s32(int32_t __p0, int32_t __p1) {
56850  int64_t __ret;
56851  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56852  return __ret;
56853}
56854#endif
56855
56856#ifdef __LITTLE_ENDIAN__
56857__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56858  int32_t __ret;
56859  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56860  return __ret;
56861}
56862#else
56863__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56864  int32_t __ret;
56865  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56866  return __ret;
56867}
56868__ai int32_t __noswap_vqdmullh_s16(int16_t __p0, int16_t __p1) {
56869  int32_t __ret;
56870  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56871  return __ret;
56872}
56873#endif
56874
56875#ifdef __LITTLE_ENDIAN__
56876__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56877  int64x2_t __ret;
56878  __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
56879  return __ret;
56880}
56881#else
56882__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56883  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56884  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56885  int64x2_t __ret;
56886  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
56887  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56888  return __ret;
56889}
56890#endif
56891
56892#ifdef __LITTLE_ENDIAN__
56893__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56894  int32x4_t __ret;
56895  __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
56896  return __ret;
56897}
56898#else
56899__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56900  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56901  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56902  int32x4_t __ret;
56903  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
56904  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56905  return __ret;
56906}
56907#endif
56908
56909#ifdef __LITTLE_ENDIAN__
56910#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56911  int32x4_t __s0 = __p0; \
56912  int32x2_t __s1 = __p1; \
56913  int64x2_t __ret; \
56914  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56915  __ret; \
56916})
56917#else
56918#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56919  int32x4_t __s0 = __p0; \
56920  int32x2_t __s1 = __p1; \
56921  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56922  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56923  int64x2_t __ret; \
56924  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56925  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56926  __ret; \
56927})
56928#endif
56929
56930#ifdef __LITTLE_ENDIAN__
56931#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56932  int16x8_t __s0 = __p0; \
56933  int16x4_t __s1 = __p1; \
56934  int32x4_t __ret; \
56935  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56936  __ret; \
56937})
56938#else
56939#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56940  int16x8_t __s0 = __p0; \
56941  int16x4_t __s1 = __p1; \
56942  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56943  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56944  int32x4_t __ret; \
56945  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56946  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56947  __ret; \
56948})
56949#endif
56950
56951#ifdef __LITTLE_ENDIAN__
56952#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56953  int32x4_t __s0 = __p0; \
56954  int32x4_t __s1 = __p1; \
56955  int64x2_t __ret; \
56956  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56957  __ret; \
56958})
56959#else
56960#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56961  int32x4_t __s0 = __p0; \
56962  int32x4_t __s1 = __p1; \
56963  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56964  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56965  int64x2_t __ret; \
56966  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56968  __ret; \
56969})
56970#endif
56971
56972#ifdef __LITTLE_ENDIAN__
56973#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56974  int16x8_t __s0 = __p0; \
56975  int16x8_t __s1 = __p1; \
56976  int32x4_t __ret; \
56977  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56978  __ret; \
56979})
56980#else
56981#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56982  int16x8_t __s0 = __p0; \
56983  int16x8_t __s1 = __p1; \
56984  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56985  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56986  int32x4_t __ret; \
56987  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56989  __ret; \
56990})
56991#endif
56992
56993#ifdef __LITTLE_ENDIAN__
56994__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
56995  int64x2_t __ret;
56996  __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
56997  return __ret;
56998}
56999#else
57000__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
57001  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57002  int64x2_t __ret;
57003  __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
57004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57005  return __ret;
57006}
57007#endif
57008
57009#ifdef __LITTLE_ENDIAN__
57010__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57011  int32x4_t __ret;
57012  __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
57013  return __ret;
57014}
57015#else
57016__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57017  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57018  int32x4_t __ret;
57019  __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
57020  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57021  return __ret;
57022}
57023#endif
57024
57025#ifdef __LITTLE_ENDIAN__
57026#define vqdmulls_lane_s32(__p0_154, __p1_154, __p2_154) __extension__ ({ \
57027  int32_t __s0_154 = __p0_154; \
57028  int32x2_t __s1_154 = __p1_154; \
57029  int64_t __ret_154; \
57030  __ret_154 = vqdmulls_s32(__s0_154, vget_lane_s32(__s1_154, __p2_154)); \
57031  __ret_154; \
57032})
57033#else
57034#define vqdmulls_lane_s32(__p0_155, __p1_155, __p2_155) __extension__ ({ \
57035  int32_t __s0_155 = __p0_155; \
57036  int32x2_t __s1_155 = __p1_155; \
57037  int32x2_t __rev1_155;  __rev1_155 = __builtin_shufflevector(__s1_155, __s1_155, 1, 0); \
57038  int64_t __ret_155; \
57039  __ret_155 = __noswap_vqdmulls_s32(__s0_155, __noswap_vget_lane_s32(__rev1_155, __p2_155)); \
57040  __ret_155; \
57041})
57042#endif
57043
57044#ifdef __LITTLE_ENDIAN__
57045#define vqdmullh_lane_s16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
57046  int16_t __s0_156 = __p0_156; \
57047  int16x4_t __s1_156 = __p1_156; \
57048  int32_t __ret_156; \
57049  __ret_156 = vqdmullh_s16(__s0_156, vget_lane_s16(__s1_156, __p2_156)); \
57050  __ret_156; \
57051})
57052#else
57053#define vqdmullh_lane_s16(__p0_157, __p1_157, __p2_157) __extension__ ({ \
57054  int16_t __s0_157 = __p0_157; \
57055  int16x4_t __s1_157 = __p1_157; \
57056  int16x4_t __rev1_157;  __rev1_157 = __builtin_shufflevector(__s1_157, __s1_157, 3, 2, 1, 0); \
57057  int32_t __ret_157; \
57058  __ret_157 = __noswap_vqdmullh_s16(__s0_157, __noswap_vget_lane_s16(__rev1_157, __p2_157)); \
57059  __ret_157; \
57060})
57061#endif
57062
57063#ifdef __LITTLE_ENDIAN__
57064#define vqdmulls_laneq_s32(__p0_158, __p1_158, __p2_158) __extension__ ({ \
57065  int32_t __s0_158 = __p0_158; \
57066  int32x4_t __s1_158 = __p1_158; \
57067  int64_t __ret_158; \
57068  __ret_158 = vqdmulls_s32(__s0_158, vgetq_lane_s32(__s1_158, __p2_158)); \
57069  __ret_158; \
57070})
57071#else
57072#define vqdmulls_laneq_s32(__p0_159, __p1_159, __p2_159) __extension__ ({ \
57073  int32_t __s0_159 = __p0_159; \
57074  int32x4_t __s1_159 = __p1_159; \
57075  int32x4_t __rev1_159;  __rev1_159 = __builtin_shufflevector(__s1_159, __s1_159, 3, 2, 1, 0); \
57076  int64_t __ret_159; \
57077  __ret_159 = __noswap_vqdmulls_s32(__s0_159, __noswap_vgetq_lane_s32(__rev1_159, __p2_159)); \
57078  __ret_159; \
57079})
57080#endif
57081
57082#ifdef __LITTLE_ENDIAN__
57083#define vqdmullh_laneq_s16(__p0_160, __p1_160, __p2_160) __extension__ ({ \
57084  int16_t __s0_160 = __p0_160; \
57085  int16x8_t __s1_160 = __p1_160; \
57086  int32_t __ret_160; \
57087  __ret_160 = vqdmullh_s16(__s0_160, vgetq_lane_s16(__s1_160, __p2_160)); \
57088  __ret_160; \
57089})
57090#else
57091#define vqdmullh_laneq_s16(__p0_161, __p1_161, __p2_161) __extension__ ({ \
57092  int16_t __s0_161 = __p0_161; \
57093  int16x8_t __s1_161 = __p1_161; \
57094  int16x8_t __rev1_161;  __rev1_161 = __builtin_shufflevector(__s1_161, __s1_161, 7, 6, 5, 4, 3, 2, 1, 0); \
57095  int32_t __ret_161; \
57096  __ret_161 = __noswap_vqdmullh_s16(__s0_161, __noswap_vgetq_lane_s16(__rev1_161, __p2_161)); \
57097  __ret_161; \
57098})
57099#endif
57100
57101#ifdef __LITTLE_ENDIAN__
57102#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57103  int32x2_t __s0 = __p0; \
57104  int32x4_t __s1 = __p1; \
57105  int64x2_t __ret; \
57106  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57107  __ret; \
57108})
57109#else
57110#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57111  int32x2_t __s0 = __p0; \
57112  int32x4_t __s1 = __p1; \
57113  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57114  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57115  int64x2_t __ret; \
57116  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57117  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57118  __ret; \
57119})
57120#endif
57121
57122#ifdef __LITTLE_ENDIAN__
57123#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57124  int16x4_t __s0 = __p0; \
57125  int16x8_t __s1 = __p1; \
57126  int32x4_t __ret; \
57127  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57128  __ret; \
57129})
57130#else
57131#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57132  int16x4_t __s0 = __p0; \
57133  int16x8_t __s1 = __p1; \
57134  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57135  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57136  int32x4_t __ret; \
57137  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57138  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57139  __ret; \
57140})
57141#endif
57142
57143#ifdef __LITTLE_ENDIAN__
57144__ai int16_t vqmovns_s32(int32_t __p0) {
57145  int16_t __ret;
57146  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57147  return __ret;
57148}
57149#else
57150__ai int16_t vqmovns_s32(int32_t __p0) {
57151  int16_t __ret;
57152  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57153  return __ret;
57154}
57155#endif
57156
57157#ifdef __LITTLE_ENDIAN__
57158__ai int32_t vqmovnd_s64(int64_t __p0) {
57159  int32_t __ret;
57160  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57161  return __ret;
57162}
57163#else
57164__ai int32_t vqmovnd_s64(int64_t __p0) {
57165  int32_t __ret;
57166  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57167  return __ret;
57168}
57169#endif
57170
57171#ifdef __LITTLE_ENDIAN__
57172__ai int8_t vqmovnh_s16(int16_t __p0) {
57173  int8_t __ret;
57174  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57175  return __ret;
57176}
57177#else
57178__ai int8_t vqmovnh_s16(int16_t __p0) {
57179  int8_t __ret;
57180  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57181  return __ret;
57182}
57183#endif
57184
57185#ifdef __LITTLE_ENDIAN__
57186__ai uint16_t vqmovns_u32(uint32_t __p0) {
57187  uint16_t __ret;
57188  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57189  return __ret;
57190}
57191#else
57192__ai uint16_t vqmovns_u32(uint32_t __p0) {
57193  uint16_t __ret;
57194  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57195  return __ret;
57196}
57197#endif
57198
57199#ifdef __LITTLE_ENDIAN__
57200__ai uint32_t vqmovnd_u64(uint64_t __p0) {
57201  uint32_t __ret;
57202  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57203  return __ret;
57204}
57205#else
57206__ai uint32_t vqmovnd_u64(uint64_t __p0) {
57207  uint32_t __ret;
57208  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57209  return __ret;
57210}
57211#endif
57212
57213#ifdef __LITTLE_ENDIAN__
57214__ai uint8_t vqmovnh_u16(uint16_t __p0) {
57215  uint8_t __ret;
57216  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57217  return __ret;
57218}
57219#else
57220__ai uint8_t vqmovnh_u16(uint16_t __p0) {
57221  uint8_t __ret;
57222  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57223  return __ret;
57224}
57225#endif
57226
57227#ifdef __LITTLE_ENDIAN__
57228__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57229  uint16x8_t __ret;
57230  __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
57231  return __ret;
57232}
57233#else
57234__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57235  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57236  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57237  uint16x8_t __ret;
57238  __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
57239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57240  return __ret;
57241}
57242#endif
57243
57244#ifdef __LITTLE_ENDIAN__
57245__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57246  uint32x4_t __ret;
57247  __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
57248  return __ret;
57249}
57250#else
57251__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57252  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57253  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57254  uint32x4_t __ret;
57255  __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
57256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57257  return __ret;
57258}
57259#endif
57260
57261#ifdef __LITTLE_ENDIAN__
57262__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57263  uint8x16_t __ret;
57264  __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
57265  return __ret;
57266}
57267#else
57268__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57269  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57270  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57271  uint8x16_t __ret;
57272  __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
57273  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57274  return __ret;
57275}
57276#endif
57277
57278#ifdef __LITTLE_ENDIAN__
57279__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57280  int16x8_t __ret;
57281  __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
57282  return __ret;
57283}
57284#else
57285__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57286  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57287  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57288  int16x8_t __ret;
57289  __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
57290  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57291  return __ret;
57292}
57293#endif
57294
57295#ifdef __LITTLE_ENDIAN__
57296__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57297  int32x4_t __ret;
57298  __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
57299  return __ret;
57300}
57301#else
57302__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57303  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57304  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57305  int32x4_t __ret;
57306  __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
57307  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57308  return __ret;
57309}
57310#endif
57311
57312#ifdef __LITTLE_ENDIAN__
57313__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57314  int8x16_t __ret;
57315  __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
57316  return __ret;
57317}
57318#else
57319__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57320  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57321  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57322  int8x16_t __ret;
57323  __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
57324  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57325  return __ret;
57326}
57327#endif
57328
57329#ifdef __LITTLE_ENDIAN__
57330__ai int16_t vqmovuns_s32(int32_t __p0) {
57331  int16_t __ret;
57332  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57333  return __ret;
57334}
57335#else
57336__ai int16_t vqmovuns_s32(int32_t __p0) {
57337  int16_t __ret;
57338  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57339  return __ret;
57340}
57341#endif
57342
57343#ifdef __LITTLE_ENDIAN__
57344__ai int32_t vqmovund_s64(int64_t __p0) {
57345  int32_t __ret;
57346  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57347  return __ret;
57348}
57349#else
57350__ai int32_t vqmovund_s64(int64_t __p0) {
57351  int32_t __ret;
57352  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57353  return __ret;
57354}
57355#endif
57356
57357#ifdef __LITTLE_ENDIAN__
57358__ai int8_t vqmovunh_s16(int16_t __p0) {
57359  int8_t __ret;
57360  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57361  return __ret;
57362}
57363#else
57364__ai int8_t vqmovunh_s16(int16_t __p0) {
57365  int8_t __ret;
57366  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57367  return __ret;
57368}
57369#endif
57370
57371#ifdef __LITTLE_ENDIAN__
57372__ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57373  uint16x8_t __ret;
57374  __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
57375  return __ret;
57376}
57377#else
57378__ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57379  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57380  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57381  uint16x8_t __ret;
57382  __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
57383  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57384  return __ret;
57385}
57386#endif
57387
57388#ifdef __LITTLE_ENDIAN__
57389__ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57390  uint32x4_t __ret;
57391  __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
57392  return __ret;
57393}
57394#else
57395__ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57396  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57397  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57398  uint32x4_t __ret;
57399  __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
57400  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57401  return __ret;
57402}
57403#endif
57404
57405#ifdef __LITTLE_ENDIAN__
57406__ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57407  uint8x16_t __ret;
57408  __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
57409  return __ret;
57410}
57411#else
57412__ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57413  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57414  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57415  uint8x16_t __ret;
57416  __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
57417  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57418  return __ret;
57419}
57420#endif
57421
57422#ifdef __LITTLE_ENDIAN__
57423__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57424  int64x2_t __ret;
57425  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
57426  return __ret;
57427}
57428#else
57429__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57430  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57431  int64x2_t __ret;
57432  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
57433  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57434  return __ret;
57435}
57436#endif
57437
57438#ifdef __LITTLE_ENDIAN__
57439__ai int64x1_t vqneg_s64(int64x1_t __p0) {
57440  int64x1_t __ret;
57441  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57442  return __ret;
57443}
57444#else
57445__ai int64x1_t vqneg_s64(int64x1_t __p0) {
57446  int64x1_t __ret;
57447  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57448  return __ret;
57449}
57450#endif
57451
57452#ifdef __LITTLE_ENDIAN__
57453__ai int8_t vqnegb_s8(int8_t __p0) {
57454  int8_t __ret;
57455  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57456  return __ret;
57457}
57458#else
57459__ai int8_t vqnegb_s8(int8_t __p0) {
57460  int8_t __ret;
57461  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57462  return __ret;
57463}
57464#endif
57465
57466#ifdef __LITTLE_ENDIAN__
57467__ai int32_t vqnegs_s32(int32_t __p0) {
57468  int32_t __ret;
57469  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57470  return __ret;
57471}
57472#else
57473__ai int32_t vqnegs_s32(int32_t __p0) {
57474  int32_t __ret;
57475  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57476  return __ret;
57477}
57478#endif
57479
57480#ifdef __LITTLE_ENDIAN__
57481__ai int64_t vqnegd_s64(int64_t __p0) {
57482  int64_t __ret;
57483  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57484  return __ret;
57485}
57486#else
57487__ai int64_t vqnegd_s64(int64_t __p0) {
57488  int64_t __ret;
57489  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57490  return __ret;
57491}
57492#endif
57493
57494#ifdef __LITTLE_ENDIAN__
57495__ai int16_t vqnegh_s16(int16_t __p0) {
57496  int16_t __ret;
57497  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57498  return __ret;
57499}
57500#else
57501__ai int16_t vqnegh_s16(int16_t __p0) {
57502  int16_t __ret;
57503  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57504  return __ret;
57505}
57506#endif
57507
57508#ifdef __LITTLE_ENDIAN__
57509__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57510  int32_t __ret;
57511  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57512  return __ret;
57513}
57514#else
57515__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57516  int32_t __ret;
57517  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57518  return __ret;
57519}
57520__ai int32_t __noswap_vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57521  int32_t __ret;
57522  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57523  return __ret;
57524}
57525#endif
57526
57527#ifdef __LITTLE_ENDIAN__
57528__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57529  int16_t __ret;
57530  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57531  return __ret;
57532}
57533#else
57534__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57535  int16_t __ret;
57536  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57537  return __ret;
57538}
57539__ai int16_t __noswap_vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57540  int16_t __ret;
57541  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57542  return __ret;
57543}
57544#endif
57545
57546#ifdef __LITTLE_ENDIAN__
57547#define vqrdmulhs_lane_s32(__p0_162, __p1_162, __p2_162) __extension__ ({ \
57548  int32_t __s0_162 = __p0_162; \
57549  int32x2_t __s1_162 = __p1_162; \
57550  int32_t __ret_162; \
57551  __ret_162 = vqrdmulhs_s32(__s0_162, vget_lane_s32(__s1_162, __p2_162)); \
57552  __ret_162; \
57553})
57554#else
57555#define vqrdmulhs_lane_s32(__p0_163, __p1_163, __p2_163) __extension__ ({ \
57556  int32_t __s0_163 = __p0_163; \
57557  int32x2_t __s1_163 = __p1_163; \
57558  int32x2_t __rev1_163;  __rev1_163 = __builtin_shufflevector(__s1_163, __s1_163, 1, 0); \
57559  int32_t __ret_163; \
57560  __ret_163 = __noswap_vqrdmulhs_s32(__s0_163, __noswap_vget_lane_s32(__rev1_163, __p2_163)); \
57561  __ret_163; \
57562})
57563#endif
57564
57565#ifdef __LITTLE_ENDIAN__
57566#define vqrdmulhh_lane_s16(__p0_164, __p1_164, __p2_164) __extension__ ({ \
57567  int16_t __s0_164 = __p0_164; \
57568  int16x4_t __s1_164 = __p1_164; \
57569  int16_t __ret_164; \
57570  __ret_164 = vqrdmulhh_s16(__s0_164, vget_lane_s16(__s1_164, __p2_164)); \
57571  __ret_164; \
57572})
57573#else
57574#define vqrdmulhh_lane_s16(__p0_165, __p1_165, __p2_165) __extension__ ({ \
57575  int16_t __s0_165 = __p0_165; \
57576  int16x4_t __s1_165 = __p1_165; \
57577  int16x4_t __rev1_165;  __rev1_165 = __builtin_shufflevector(__s1_165, __s1_165, 3, 2, 1, 0); \
57578  int16_t __ret_165; \
57579  __ret_165 = __noswap_vqrdmulhh_s16(__s0_165, __noswap_vget_lane_s16(__rev1_165, __p2_165)); \
57580  __ret_165; \
57581})
57582#endif
57583
57584#ifdef __LITTLE_ENDIAN__
57585#define vqrdmulhs_laneq_s32(__p0_166, __p1_166, __p2_166) __extension__ ({ \
57586  int32_t __s0_166 = __p0_166; \
57587  int32x4_t __s1_166 = __p1_166; \
57588  int32_t __ret_166; \
57589  __ret_166 = vqrdmulhs_s32(__s0_166, vgetq_lane_s32(__s1_166, __p2_166)); \
57590  __ret_166; \
57591})
57592#else
57593#define vqrdmulhs_laneq_s32(__p0_167, __p1_167, __p2_167) __extension__ ({ \
57594  int32_t __s0_167 = __p0_167; \
57595  int32x4_t __s1_167 = __p1_167; \
57596  int32x4_t __rev1_167;  __rev1_167 = __builtin_shufflevector(__s1_167, __s1_167, 3, 2, 1, 0); \
57597  int32_t __ret_167; \
57598  __ret_167 = __noswap_vqrdmulhs_s32(__s0_167, __noswap_vgetq_lane_s32(__rev1_167, __p2_167)); \
57599  __ret_167; \
57600})
57601#endif
57602
57603#ifdef __LITTLE_ENDIAN__
57604#define vqrdmulhh_laneq_s16(__p0_168, __p1_168, __p2_168) __extension__ ({ \
57605  int16_t __s0_168 = __p0_168; \
57606  int16x8_t __s1_168 = __p1_168; \
57607  int16_t __ret_168; \
57608  __ret_168 = vqrdmulhh_s16(__s0_168, vgetq_lane_s16(__s1_168, __p2_168)); \
57609  __ret_168; \
57610})
57611#else
57612#define vqrdmulhh_laneq_s16(__p0_169, __p1_169, __p2_169) __extension__ ({ \
57613  int16_t __s0_169 = __p0_169; \
57614  int16x8_t __s1_169 = __p1_169; \
57615  int16x8_t __rev1_169;  __rev1_169 = __builtin_shufflevector(__s1_169, __s1_169, 7, 6, 5, 4, 3, 2, 1, 0); \
57616  int16_t __ret_169; \
57617  __ret_169 = __noswap_vqrdmulhh_s16(__s0_169, __noswap_vgetq_lane_s16(__rev1_169, __p2_169)); \
57618  __ret_169; \
57619})
57620#endif
57621
57622#ifdef __LITTLE_ENDIAN__
57623#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57624  int32x4_t __s0 = __p0; \
57625  int32x4_t __s1 = __p1; \
57626  int32x4_t __ret; \
57627  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57628  __ret; \
57629})
57630#else
57631#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57632  int32x4_t __s0 = __p0; \
57633  int32x4_t __s1 = __p1; \
57634  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57635  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57636  int32x4_t __ret; \
57637  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57638  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57639  __ret; \
57640})
57641#endif
57642
57643#ifdef __LITTLE_ENDIAN__
57644#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57645  int16x8_t __s0 = __p0; \
57646  int16x8_t __s1 = __p1; \
57647  int16x8_t __ret; \
57648  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57649  __ret; \
57650})
57651#else
57652#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57653  int16x8_t __s0 = __p0; \
57654  int16x8_t __s1 = __p1; \
57655  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
57656  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57657  int16x8_t __ret; \
57658  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57659  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
57660  __ret; \
57661})
57662#endif
57663
57664#ifdef __LITTLE_ENDIAN__
57665#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57666  int32x2_t __s0 = __p0; \
57667  int32x4_t __s1 = __p1; \
57668  int32x2_t __ret; \
57669  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57670  __ret; \
57671})
57672#else
57673#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57674  int32x2_t __s0 = __p0; \
57675  int32x4_t __s1 = __p1; \
57676  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57677  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57678  int32x2_t __ret; \
57679  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57681  __ret; \
57682})
57683#endif
57684
57685#ifdef __LITTLE_ENDIAN__
57686#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57687  int16x4_t __s0 = __p0; \
57688  int16x8_t __s1 = __p1; \
57689  int16x4_t __ret; \
57690  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57691  __ret; \
57692})
57693#else
57694#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57695  int16x4_t __s0 = __p0; \
57696  int16x8_t __s1 = __p1; \
57697  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57698  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57699  int16x4_t __ret; \
57700  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57702  __ret; \
57703})
57704#endif
57705
57706#ifdef __LITTLE_ENDIAN__
57707__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57708  uint8_t __ret;
57709  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57710  return __ret;
57711}
57712#else
57713__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57714  uint8_t __ret;
57715  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57716  return __ret;
57717}
57718#endif
57719
57720#ifdef __LITTLE_ENDIAN__
57721__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57722  uint32_t __ret;
57723  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57724  return __ret;
57725}
57726#else
57727__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57728  uint32_t __ret;
57729  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57730  return __ret;
57731}
57732#endif
57733
57734#ifdef __LITTLE_ENDIAN__
57735__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57736  uint64_t __ret;
57737  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57738  return __ret;
57739}
57740#else
57741__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57742  uint64_t __ret;
57743  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57744  return __ret;
57745}
57746#endif
57747
57748#ifdef __LITTLE_ENDIAN__
57749__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57750  uint16_t __ret;
57751  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57752  return __ret;
57753}
57754#else
57755__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57756  uint16_t __ret;
57757  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57758  return __ret;
57759}
57760#endif
57761
57762#ifdef __LITTLE_ENDIAN__
57763__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57764  int8_t __ret;
57765  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57766  return __ret;
57767}
57768#else
57769__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57770  int8_t __ret;
57771  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57772  return __ret;
57773}
57774#endif
57775
57776#ifdef __LITTLE_ENDIAN__
57777__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57778  int32_t __ret;
57779  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57780  return __ret;
57781}
57782#else
57783__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57784  int32_t __ret;
57785  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57786  return __ret;
57787}
57788#endif
57789
57790#ifdef __LITTLE_ENDIAN__
57791__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57792  int64_t __ret;
57793  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57794  return __ret;
57795}
57796#else
57797__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57798  int64_t __ret;
57799  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57800  return __ret;
57801}
57802#endif
57803
57804#ifdef __LITTLE_ENDIAN__
57805__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57806  int16_t __ret;
57807  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57808  return __ret;
57809}
57810#else
57811__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57812  int16_t __ret;
57813  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57814  return __ret;
57815}
57816#endif
57817
57818#ifdef __LITTLE_ENDIAN__
57819#define vqrshrn_high_n_u32(__p0_170, __p1_170, __p2_170) __extension__ ({ \
57820  uint16x4_t __s0_170 = __p0_170; \
57821  uint32x4_t __s1_170 = __p1_170; \
57822  uint16x8_t __ret_170; \
57823  __ret_170 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_170), (uint16x4_t)(vqrshrn_n_u32(__s1_170, __p2_170)))); \
57824  __ret_170; \
57825})
57826#else
57827#define vqrshrn_high_n_u32(__p0_171, __p1_171, __p2_171) __extension__ ({ \
57828  uint16x4_t __s0_171 = __p0_171; \
57829  uint32x4_t __s1_171 = __p1_171; \
57830  uint16x4_t __rev0_171;  __rev0_171 = __builtin_shufflevector(__s0_171, __s0_171, 3, 2, 1, 0); \
57831  uint32x4_t __rev1_171;  __rev1_171 = __builtin_shufflevector(__s1_171, __s1_171, 3, 2, 1, 0); \
57832  uint16x8_t __ret_171; \
57833  __ret_171 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_171), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_171, __p2_171)))); \
57834  __ret_171 = __builtin_shufflevector(__ret_171, __ret_171, 7, 6, 5, 4, 3, 2, 1, 0); \
57835  __ret_171; \
57836})
57837#endif
57838
57839#ifdef __LITTLE_ENDIAN__
57840#define vqrshrn_high_n_u64(__p0_172, __p1_172, __p2_172) __extension__ ({ \
57841  uint32x2_t __s0_172 = __p0_172; \
57842  uint64x2_t __s1_172 = __p1_172; \
57843  uint32x4_t __ret_172; \
57844  __ret_172 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_172), (uint32x2_t)(vqrshrn_n_u64(__s1_172, __p2_172)))); \
57845  __ret_172; \
57846})
57847#else
57848#define vqrshrn_high_n_u64(__p0_173, __p1_173, __p2_173) __extension__ ({ \
57849  uint32x2_t __s0_173 = __p0_173; \
57850  uint64x2_t __s1_173 = __p1_173; \
57851  uint32x2_t __rev0_173;  __rev0_173 = __builtin_shufflevector(__s0_173, __s0_173, 1, 0); \
57852  uint64x2_t __rev1_173;  __rev1_173 = __builtin_shufflevector(__s1_173, __s1_173, 1, 0); \
57853  uint32x4_t __ret_173; \
57854  __ret_173 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_173), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_173, __p2_173)))); \
57855  __ret_173 = __builtin_shufflevector(__ret_173, __ret_173, 3, 2, 1, 0); \
57856  __ret_173; \
57857})
57858#endif
57859
57860#ifdef __LITTLE_ENDIAN__
57861#define vqrshrn_high_n_u16(__p0_174, __p1_174, __p2_174) __extension__ ({ \
57862  uint8x8_t __s0_174 = __p0_174; \
57863  uint16x8_t __s1_174 = __p1_174; \
57864  uint8x16_t __ret_174; \
57865  __ret_174 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_174), (uint8x8_t)(vqrshrn_n_u16(__s1_174, __p2_174)))); \
57866  __ret_174; \
57867})
57868#else
57869#define vqrshrn_high_n_u16(__p0_175, __p1_175, __p2_175) __extension__ ({ \
57870  uint8x8_t __s0_175 = __p0_175; \
57871  uint16x8_t __s1_175 = __p1_175; \
57872  uint8x8_t __rev0_175;  __rev0_175 = __builtin_shufflevector(__s0_175, __s0_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57873  uint16x8_t __rev1_175;  __rev1_175 = __builtin_shufflevector(__s1_175, __s1_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57874  uint8x16_t __ret_175; \
57875  __ret_175 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_175), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_175, __p2_175)))); \
57876  __ret_175 = __builtin_shufflevector(__ret_175, __ret_175, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57877  __ret_175; \
57878})
57879#endif
57880
57881#ifdef __LITTLE_ENDIAN__
57882#define vqrshrn_high_n_s32(__p0_176, __p1_176, __p2_176) __extension__ ({ \
57883  int16x4_t __s0_176 = __p0_176; \
57884  int32x4_t __s1_176 = __p1_176; \
57885  int16x8_t __ret_176; \
57886  __ret_176 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_176), (int16x4_t)(vqrshrn_n_s32(__s1_176, __p2_176)))); \
57887  __ret_176; \
57888})
57889#else
57890#define vqrshrn_high_n_s32(__p0_177, __p1_177, __p2_177) __extension__ ({ \
57891  int16x4_t __s0_177 = __p0_177; \
57892  int32x4_t __s1_177 = __p1_177; \
57893  int16x4_t __rev0_177;  __rev0_177 = __builtin_shufflevector(__s0_177, __s0_177, 3, 2, 1, 0); \
57894  int32x4_t __rev1_177;  __rev1_177 = __builtin_shufflevector(__s1_177, __s1_177, 3, 2, 1, 0); \
57895  int16x8_t __ret_177; \
57896  __ret_177 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_177), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_177, __p2_177)))); \
57897  __ret_177 = __builtin_shufflevector(__ret_177, __ret_177, 7, 6, 5, 4, 3, 2, 1, 0); \
57898  __ret_177; \
57899})
57900#endif
57901
57902#ifdef __LITTLE_ENDIAN__
57903#define vqrshrn_high_n_s64(__p0_178, __p1_178, __p2_178) __extension__ ({ \
57904  int32x2_t __s0_178 = __p0_178; \
57905  int64x2_t __s1_178 = __p1_178; \
57906  int32x4_t __ret_178; \
57907  __ret_178 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_178), (int32x2_t)(vqrshrn_n_s64(__s1_178, __p2_178)))); \
57908  __ret_178; \
57909})
57910#else
57911#define vqrshrn_high_n_s64(__p0_179, __p1_179, __p2_179) __extension__ ({ \
57912  int32x2_t __s0_179 = __p0_179; \
57913  int64x2_t __s1_179 = __p1_179; \
57914  int32x2_t __rev0_179;  __rev0_179 = __builtin_shufflevector(__s0_179, __s0_179, 1, 0); \
57915  int64x2_t __rev1_179;  __rev1_179 = __builtin_shufflevector(__s1_179, __s1_179, 1, 0); \
57916  int32x4_t __ret_179; \
57917  __ret_179 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_179), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_179, __p2_179)))); \
57918  __ret_179 = __builtin_shufflevector(__ret_179, __ret_179, 3, 2, 1, 0); \
57919  __ret_179; \
57920})
57921#endif
57922
57923#ifdef __LITTLE_ENDIAN__
57924#define vqrshrn_high_n_s16(__p0_180, __p1_180, __p2_180) __extension__ ({ \
57925  int8x8_t __s0_180 = __p0_180; \
57926  int16x8_t __s1_180 = __p1_180; \
57927  int8x16_t __ret_180; \
57928  __ret_180 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_180), (int8x8_t)(vqrshrn_n_s16(__s1_180, __p2_180)))); \
57929  __ret_180; \
57930})
57931#else
57932#define vqrshrn_high_n_s16(__p0_181, __p1_181, __p2_181) __extension__ ({ \
57933  int8x8_t __s0_181 = __p0_181; \
57934  int16x8_t __s1_181 = __p1_181; \
57935  int8x8_t __rev0_181;  __rev0_181 = __builtin_shufflevector(__s0_181, __s0_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57936  int16x8_t __rev1_181;  __rev1_181 = __builtin_shufflevector(__s1_181, __s1_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57937  int8x16_t __ret_181; \
57938  __ret_181 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_181), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_181, __p2_181)))); \
57939  __ret_181 = __builtin_shufflevector(__ret_181, __ret_181, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57940  __ret_181; \
57941})
57942#endif
57943
57944#ifdef __LITTLE_ENDIAN__
57945#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57946  uint32_t __s0 = __p0; \
57947  uint16_t __ret; \
57948  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57949  __ret; \
57950})
57951#else
57952#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57953  uint32_t __s0 = __p0; \
57954  uint16_t __ret; \
57955  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57956  __ret; \
57957})
57958#endif
57959
57960#ifdef __LITTLE_ENDIAN__
57961#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57962  uint64_t __s0 = __p0; \
57963  uint32_t __ret; \
57964  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57965  __ret; \
57966})
57967#else
57968#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57969  uint64_t __s0 = __p0; \
57970  uint32_t __ret; \
57971  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57972  __ret; \
57973})
57974#endif
57975
57976#ifdef __LITTLE_ENDIAN__
57977#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57978  uint16_t __s0 = __p0; \
57979  uint8_t __ret; \
57980  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57981  __ret; \
57982})
57983#else
57984#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57985  uint16_t __s0 = __p0; \
57986  uint8_t __ret; \
57987  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57988  __ret; \
57989})
57990#endif
57991
57992#ifdef __LITTLE_ENDIAN__
57993#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
57994  int32_t __s0 = __p0; \
57995  int16_t __ret; \
57996  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
57997  __ret; \
57998})
57999#else
58000#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
58001  int32_t __s0 = __p0; \
58002  int16_t __ret; \
58003  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
58004  __ret; \
58005})
58006#endif
58007
58008#ifdef __LITTLE_ENDIAN__
58009#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58010  int64_t __s0 = __p0; \
58011  int32_t __ret; \
58012  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58013  __ret; \
58014})
58015#else
58016#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58017  int64_t __s0 = __p0; \
58018  int32_t __ret; \
58019  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58020  __ret; \
58021})
58022#endif
58023
58024#ifdef __LITTLE_ENDIAN__
58025#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58026  int16_t __s0 = __p0; \
58027  int8_t __ret; \
58028  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58029  __ret; \
58030})
58031#else
58032#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58033  int16_t __s0 = __p0; \
58034  int8_t __ret; \
58035  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58036  __ret; \
58037})
58038#endif
58039
58040#ifdef __LITTLE_ENDIAN__
58041#define vqrshrun_high_n_s32(__p0_182, __p1_182, __p2_182) __extension__ ({ \
58042  int16x4_t __s0_182 = __p0_182; \
58043  int32x4_t __s1_182 = __p1_182; \
58044  int16x8_t __ret_182; \
58045  __ret_182 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_182), (int16x4_t)(vqrshrun_n_s32(__s1_182, __p2_182)))); \
58046  __ret_182; \
58047})
58048#else
58049#define vqrshrun_high_n_s32(__p0_183, __p1_183, __p2_183) __extension__ ({ \
58050  int16x4_t __s0_183 = __p0_183; \
58051  int32x4_t __s1_183 = __p1_183; \
58052  int16x4_t __rev0_183;  __rev0_183 = __builtin_shufflevector(__s0_183, __s0_183, 3, 2, 1, 0); \
58053  int32x4_t __rev1_183;  __rev1_183 = __builtin_shufflevector(__s1_183, __s1_183, 3, 2, 1, 0); \
58054  int16x8_t __ret_183; \
58055  __ret_183 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_183), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_183, __p2_183)))); \
58056  __ret_183 = __builtin_shufflevector(__ret_183, __ret_183, 7, 6, 5, 4, 3, 2, 1, 0); \
58057  __ret_183; \
58058})
58059#endif
58060
58061#ifdef __LITTLE_ENDIAN__
58062#define vqrshrun_high_n_s64(__p0_184, __p1_184, __p2_184) __extension__ ({ \
58063  int32x2_t __s0_184 = __p0_184; \
58064  int64x2_t __s1_184 = __p1_184; \
58065  int32x4_t __ret_184; \
58066  __ret_184 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_184), (int32x2_t)(vqrshrun_n_s64(__s1_184, __p2_184)))); \
58067  __ret_184; \
58068})
58069#else
58070#define vqrshrun_high_n_s64(__p0_185, __p1_185, __p2_185) __extension__ ({ \
58071  int32x2_t __s0_185 = __p0_185; \
58072  int64x2_t __s1_185 = __p1_185; \
58073  int32x2_t __rev0_185;  __rev0_185 = __builtin_shufflevector(__s0_185, __s0_185, 1, 0); \
58074  int64x2_t __rev1_185;  __rev1_185 = __builtin_shufflevector(__s1_185, __s1_185, 1, 0); \
58075  int32x4_t __ret_185; \
58076  __ret_185 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_185), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_185, __p2_185)))); \
58077  __ret_185 = __builtin_shufflevector(__ret_185, __ret_185, 3, 2, 1, 0); \
58078  __ret_185; \
58079})
58080#endif
58081
58082#ifdef __LITTLE_ENDIAN__
58083#define vqrshrun_high_n_s16(__p0_186, __p1_186, __p2_186) __extension__ ({ \
58084  int8x8_t __s0_186 = __p0_186; \
58085  int16x8_t __s1_186 = __p1_186; \
58086  int8x16_t __ret_186; \
58087  __ret_186 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_186), (int8x8_t)(vqrshrun_n_s16(__s1_186, __p2_186)))); \
58088  __ret_186; \
58089})
58090#else
58091#define vqrshrun_high_n_s16(__p0_187, __p1_187, __p2_187) __extension__ ({ \
58092  int8x8_t __s0_187 = __p0_187; \
58093  int16x8_t __s1_187 = __p1_187; \
58094  int8x8_t __rev0_187;  __rev0_187 = __builtin_shufflevector(__s0_187, __s0_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58095  int16x8_t __rev1_187;  __rev1_187 = __builtin_shufflevector(__s1_187, __s1_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58096  int8x16_t __ret_187; \
58097  __ret_187 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_187), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_187, __p2_187)))); \
58098  __ret_187 = __builtin_shufflevector(__ret_187, __ret_187, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58099  __ret_187; \
58100})
58101#endif
58102
58103#ifdef __LITTLE_ENDIAN__
58104#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58105  int32_t __s0 = __p0; \
58106  int16_t __ret; \
58107  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58108  __ret; \
58109})
58110#else
58111#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58112  int32_t __s0 = __p0; \
58113  int16_t __ret; \
58114  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58115  __ret; \
58116})
58117#endif
58118
58119#ifdef __LITTLE_ENDIAN__
58120#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58121  int64_t __s0 = __p0; \
58122  int32_t __ret; \
58123  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58124  __ret; \
58125})
58126#else
58127#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58128  int64_t __s0 = __p0; \
58129  int32_t __ret; \
58130  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58131  __ret; \
58132})
58133#endif
58134
58135#ifdef __LITTLE_ENDIAN__
58136#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58137  int16_t __s0 = __p0; \
58138  int8_t __ret; \
58139  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58140  __ret; \
58141})
58142#else
58143#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58144  int16_t __s0 = __p0; \
58145  int8_t __ret; \
58146  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58147  __ret; \
58148})
58149#endif
58150
58151#ifdef __LITTLE_ENDIAN__
58152__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58153  uint8_t __ret;
58154  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58155  return __ret;
58156}
58157#else
58158__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58159  uint8_t __ret;
58160  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58161  return __ret;
58162}
58163#endif
58164
58165#ifdef __LITTLE_ENDIAN__
58166__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58167  uint32_t __ret;
58168  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58169  return __ret;
58170}
58171#else
58172__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58173  uint32_t __ret;
58174  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58175  return __ret;
58176}
58177#endif
58178
58179#ifdef __LITTLE_ENDIAN__
58180__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58181  uint64_t __ret;
58182  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58183  return __ret;
58184}
58185#else
58186__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58187  uint64_t __ret;
58188  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58189  return __ret;
58190}
58191#endif
58192
58193#ifdef __LITTLE_ENDIAN__
58194__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58195  uint16_t __ret;
58196  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58197  return __ret;
58198}
58199#else
58200__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58201  uint16_t __ret;
58202  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58203  return __ret;
58204}
58205#endif
58206
58207#ifdef __LITTLE_ENDIAN__
58208__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58209  int8_t __ret;
58210  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58211  return __ret;
58212}
58213#else
58214__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58215  int8_t __ret;
58216  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58217  return __ret;
58218}
58219#endif
58220
58221#ifdef __LITTLE_ENDIAN__
58222__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58223  int32_t __ret;
58224  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58225  return __ret;
58226}
58227#else
58228__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58229  int32_t __ret;
58230  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58231  return __ret;
58232}
58233#endif
58234
58235#ifdef __LITTLE_ENDIAN__
58236__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58237  int64_t __ret;
58238  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58239  return __ret;
58240}
58241#else
58242__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58243  int64_t __ret;
58244  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58245  return __ret;
58246}
58247#endif
58248
58249#ifdef __LITTLE_ENDIAN__
58250__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58251  int16_t __ret;
58252  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58253  return __ret;
58254}
58255#else
58256__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58257  int16_t __ret;
58258  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58259  return __ret;
58260}
58261#endif
58262
58263#ifdef __LITTLE_ENDIAN__
58264#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58265  uint8_t __s0 = __p0; \
58266  uint8_t __ret; \
58267  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58268  __ret; \
58269})
58270#else
58271#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58272  uint8_t __s0 = __p0; \
58273  uint8_t __ret; \
58274  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58275  __ret; \
58276})
58277#endif
58278
58279#ifdef __LITTLE_ENDIAN__
58280#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58281  uint32_t __s0 = __p0; \
58282  uint32_t __ret; \
58283  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58284  __ret; \
58285})
58286#else
58287#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58288  uint32_t __s0 = __p0; \
58289  uint32_t __ret; \
58290  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58291  __ret; \
58292})
58293#endif
58294
58295#ifdef __LITTLE_ENDIAN__
58296#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58297  uint64_t __s0 = __p0; \
58298  uint64_t __ret; \
58299  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58300  __ret; \
58301})
58302#else
58303#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58304  uint64_t __s0 = __p0; \
58305  uint64_t __ret; \
58306  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58307  __ret; \
58308})
58309#endif
58310
58311#ifdef __LITTLE_ENDIAN__
58312#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58313  uint16_t __s0 = __p0; \
58314  uint16_t __ret; \
58315  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58316  __ret; \
58317})
58318#else
58319#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58320  uint16_t __s0 = __p0; \
58321  uint16_t __ret; \
58322  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58323  __ret; \
58324})
58325#endif
58326
58327#ifdef __LITTLE_ENDIAN__
58328#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58329  int8_t __s0 = __p0; \
58330  int8_t __ret; \
58331  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58332  __ret; \
58333})
58334#else
58335#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58336  int8_t __s0 = __p0; \
58337  int8_t __ret; \
58338  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58339  __ret; \
58340})
58341#endif
58342
58343#ifdef __LITTLE_ENDIAN__
58344#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58345  int32_t __s0 = __p0; \
58346  int32_t __ret; \
58347  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58348  __ret; \
58349})
58350#else
58351#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58352  int32_t __s0 = __p0; \
58353  int32_t __ret; \
58354  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58355  __ret; \
58356})
58357#endif
58358
58359#ifdef __LITTLE_ENDIAN__
58360#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58361  int64_t __s0 = __p0; \
58362  int64_t __ret; \
58363  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58364  __ret; \
58365})
58366#else
58367#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58368  int64_t __s0 = __p0; \
58369  int64_t __ret; \
58370  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58371  __ret; \
58372})
58373#endif
58374
58375#ifdef __LITTLE_ENDIAN__
58376#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58377  int16_t __s0 = __p0; \
58378  int16_t __ret; \
58379  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58380  __ret; \
58381})
58382#else
58383#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58384  int16_t __s0 = __p0; \
58385  int16_t __ret; \
58386  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58387  __ret; \
58388})
58389#endif
58390
58391#ifdef __LITTLE_ENDIAN__
58392#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58393  int8_t __s0 = __p0; \
58394  int8_t __ret; \
58395  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58396  __ret; \
58397})
58398#else
58399#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58400  int8_t __s0 = __p0; \
58401  int8_t __ret; \
58402  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58403  __ret; \
58404})
58405#endif
58406
58407#ifdef __LITTLE_ENDIAN__
58408#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58409  int32_t __s0 = __p0; \
58410  int32_t __ret; \
58411  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58412  __ret; \
58413})
58414#else
58415#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58416  int32_t __s0 = __p0; \
58417  int32_t __ret; \
58418  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58419  __ret; \
58420})
58421#endif
58422
58423#ifdef __LITTLE_ENDIAN__
58424#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58425  int64_t __s0 = __p0; \
58426  int64_t __ret; \
58427  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58428  __ret; \
58429})
58430#else
58431#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58432  int64_t __s0 = __p0; \
58433  int64_t __ret; \
58434  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58435  __ret; \
58436})
58437#endif
58438
58439#ifdef __LITTLE_ENDIAN__
58440#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58441  int16_t __s0 = __p0; \
58442  int16_t __ret; \
58443  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58444  __ret; \
58445})
58446#else
58447#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58448  int16_t __s0 = __p0; \
58449  int16_t __ret; \
58450  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58451  __ret; \
58452})
58453#endif
58454
58455#ifdef __LITTLE_ENDIAN__
58456#define vqshrn_high_n_u32(__p0_188, __p1_188, __p2_188) __extension__ ({ \
58457  uint16x4_t __s0_188 = __p0_188; \
58458  uint32x4_t __s1_188 = __p1_188; \
58459  uint16x8_t __ret_188; \
58460  __ret_188 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_188), (uint16x4_t)(vqshrn_n_u32(__s1_188, __p2_188)))); \
58461  __ret_188; \
58462})
58463#else
58464#define vqshrn_high_n_u32(__p0_189, __p1_189, __p2_189) __extension__ ({ \
58465  uint16x4_t __s0_189 = __p0_189; \
58466  uint32x4_t __s1_189 = __p1_189; \
58467  uint16x4_t __rev0_189;  __rev0_189 = __builtin_shufflevector(__s0_189, __s0_189, 3, 2, 1, 0); \
58468  uint32x4_t __rev1_189;  __rev1_189 = __builtin_shufflevector(__s1_189, __s1_189, 3, 2, 1, 0); \
58469  uint16x8_t __ret_189; \
58470  __ret_189 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_189), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_189, __p2_189)))); \
58471  __ret_189 = __builtin_shufflevector(__ret_189, __ret_189, 7, 6, 5, 4, 3, 2, 1, 0); \
58472  __ret_189; \
58473})
58474#endif
58475
58476#ifdef __LITTLE_ENDIAN__
58477#define vqshrn_high_n_u64(__p0_190, __p1_190, __p2_190) __extension__ ({ \
58478  uint32x2_t __s0_190 = __p0_190; \
58479  uint64x2_t __s1_190 = __p1_190; \
58480  uint32x4_t __ret_190; \
58481  __ret_190 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_190), (uint32x2_t)(vqshrn_n_u64(__s1_190, __p2_190)))); \
58482  __ret_190; \
58483})
58484#else
58485#define vqshrn_high_n_u64(__p0_191, __p1_191, __p2_191) __extension__ ({ \
58486  uint32x2_t __s0_191 = __p0_191; \
58487  uint64x2_t __s1_191 = __p1_191; \
58488  uint32x2_t __rev0_191;  __rev0_191 = __builtin_shufflevector(__s0_191, __s0_191, 1, 0); \
58489  uint64x2_t __rev1_191;  __rev1_191 = __builtin_shufflevector(__s1_191, __s1_191, 1, 0); \
58490  uint32x4_t __ret_191; \
58491  __ret_191 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_191), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_191, __p2_191)))); \
58492  __ret_191 = __builtin_shufflevector(__ret_191, __ret_191, 3, 2, 1, 0); \
58493  __ret_191; \
58494})
58495#endif
58496
58497#ifdef __LITTLE_ENDIAN__
58498#define vqshrn_high_n_u16(__p0_192, __p1_192, __p2_192) __extension__ ({ \
58499  uint8x8_t __s0_192 = __p0_192; \
58500  uint16x8_t __s1_192 = __p1_192; \
58501  uint8x16_t __ret_192; \
58502  __ret_192 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_192), (uint8x8_t)(vqshrn_n_u16(__s1_192, __p2_192)))); \
58503  __ret_192; \
58504})
58505#else
58506#define vqshrn_high_n_u16(__p0_193, __p1_193, __p2_193) __extension__ ({ \
58507  uint8x8_t __s0_193 = __p0_193; \
58508  uint16x8_t __s1_193 = __p1_193; \
58509  uint8x8_t __rev0_193;  __rev0_193 = __builtin_shufflevector(__s0_193, __s0_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58510  uint16x8_t __rev1_193;  __rev1_193 = __builtin_shufflevector(__s1_193, __s1_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58511  uint8x16_t __ret_193; \
58512  __ret_193 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_193), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_193, __p2_193)))); \
58513  __ret_193 = __builtin_shufflevector(__ret_193, __ret_193, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58514  __ret_193; \
58515})
58516#endif
58517
58518#ifdef __LITTLE_ENDIAN__
58519#define vqshrn_high_n_s32(__p0_194, __p1_194, __p2_194) __extension__ ({ \
58520  int16x4_t __s0_194 = __p0_194; \
58521  int32x4_t __s1_194 = __p1_194; \
58522  int16x8_t __ret_194; \
58523  __ret_194 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_194), (int16x4_t)(vqshrn_n_s32(__s1_194, __p2_194)))); \
58524  __ret_194; \
58525})
58526#else
58527#define vqshrn_high_n_s32(__p0_195, __p1_195, __p2_195) __extension__ ({ \
58528  int16x4_t __s0_195 = __p0_195; \
58529  int32x4_t __s1_195 = __p1_195; \
58530  int16x4_t __rev0_195;  __rev0_195 = __builtin_shufflevector(__s0_195, __s0_195, 3, 2, 1, 0); \
58531  int32x4_t __rev1_195;  __rev1_195 = __builtin_shufflevector(__s1_195, __s1_195, 3, 2, 1, 0); \
58532  int16x8_t __ret_195; \
58533  __ret_195 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_195), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_195, __p2_195)))); \
58534  __ret_195 = __builtin_shufflevector(__ret_195, __ret_195, 7, 6, 5, 4, 3, 2, 1, 0); \
58535  __ret_195; \
58536})
58537#endif
58538
58539#ifdef __LITTLE_ENDIAN__
58540#define vqshrn_high_n_s64(__p0_196, __p1_196, __p2_196) __extension__ ({ \
58541  int32x2_t __s0_196 = __p0_196; \
58542  int64x2_t __s1_196 = __p1_196; \
58543  int32x4_t __ret_196; \
58544  __ret_196 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_196), (int32x2_t)(vqshrn_n_s64(__s1_196, __p2_196)))); \
58545  __ret_196; \
58546})
58547#else
58548#define vqshrn_high_n_s64(__p0_197, __p1_197, __p2_197) __extension__ ({ \
58549  int32x2_t __s0_197 = __p0_197; \
58550  int64x2_t __s1_197 = __p1_197; \
58551  int32x2_t __rev0_197;  __rev0_197 = __builtin_shufflevector(__s0_197, __s0_197, 1, 0); \
58552  int64x2_t __rev1_197;  __rev1_197 = __builtin_shufflevector(__s1_197, __s1_197, 1, 0); \
58553  int32x4_t __ret_197; \
58554  __ret_197 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_197), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_197, __p2_197)))); \
58555  __ret_197 = __builtin_shufflevector(__ret_197, __ret_197, 3, 2, 1, 0); \
58556  __ret_197; \
58557})
58558#endif
58559
58560#ifdef __LITTLE_ENDIAN__
58561#define vqshrn_high_n_s16(__p0_198, __p1_198, __p2_198) __extension__ ({ \
58562  int8x8_t __s0_198 = __p0_198; \
58563  int16x8_t __s1_198 = __p1_198; \
58564  int8x16_t __ret_198; \
58565  __ret_198 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_198), (int8x8_t)(vqshrn_n_s16(__s1_198, __p2_198)))); \
58566  __ret_198; \
58567})
58568#else
58569#define vqshrn_high_n_s16(__p0_199, __p1_199, __p2_199) __extension__ ({ \
58570  int8x8_t __s0_199 = __p0_199; \
58571  int16x8_t __s1_199 = __p1_199; \
58572  int8x8_t __rev0_199;  __rev0_199 = __builtin_shufflevector(__s0_199, __s0_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58573  int16x8_t __rev1_199;  __rev1_199 = __builtin_shufflevector(__s1_199, __s1_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58574  int8x16_t __ret_199; \
58575  __ret_199 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_199), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_199, __p2_199)))); \
58576  __ret_199 = __builtin_shufflevector(__ret_199, __ret_199, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58577  __ret_199; \
58578})
58579#endif
58580
58581#ifdef __LITTLE_ENDIAN__
58582#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58583  uint32_t __s0 = __p0; \
58584  uint16_t __ret; \
58585  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58586  __ret; \
58587})
58588#else
58589#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58590  uint32_t __s0 = __p0; \
58591  uint16_t __ret; \
58592  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58593  __ret; \
58594})
58595#endif
58596
58597#ifdef __LITTLE_ENDIAN__
58598#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58599  uint64_t __s0 = __p0; \
58600  uint32_t __ret; \
58601  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58602  __ret; \
58603})
58604#else
58605#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58606  uint64_t __s0 = __p0; \
58607  uint32_t __ret; \
58608  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58609  __ret; \
58610})
58611#endif
58612
58613#ifdef __LITTLE_ENDIAN__
58614#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58615  uint16_t __s0 = __p0; \
58616  uint8_t __ret; \
58617  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58618  __ret; \
58619})
58620#else
58621#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58622  uint16_t __s0 = __p0; \
58623  uint8_t __ret; \
58624  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58625  __ret; \
58626})
58627#endif
58628
58629#ifdef __LITTLE_ENDIAN__
58630#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58631  int32_t __s0 = __p0; \
58632  int16_t __ret; \
58633  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58634  __ret; \
58635})
58636#else
58637#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58638  int32_t __s0 = __p0; \
58639  int16_t __ret; \
58640  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58641  __ret; \
58642})
58643#endif
58644
58645#ifdef __LITTLE_ENDIAN__
58646#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58647  int64_t __s0 = __p0; \
58648  int32_t __ret; \
58649  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58650  __ret; \
58651})
58652#else
58653#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58654  int64_t __s0 = __p0; \
58655  int32_t __ret; \
58656  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58657  __ret; \
58658})
58659#endif
58660
58661#ifdef __LITTLE_ENDIAN__
58662#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58663  int16_t __s0 = __p0; \
58664  int8_t __ret; \
58665  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58666  __ret; \
58667})
58668#else
58669#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58670  int16_t __s0 = __p0; \
58671  int8_t __ret; \
58672  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58673  __ret; \
58674})
58675#endif
58676
58677#ifdef __LITTLE_ENDIAN__
58678#define vqshrun_high_n_s32(__p0_200, __p1_200, __p2_200) __extension__ ({ \
58679  int16x4_t __s0_200 = __p0_200; \
58680  int32x4_t __s1_200 = __p1_200; \
58681  int16x8_t __ret_200; \
58682  __ret_200 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_200), (int16x4_t)(vqshrun_n_s32(__s1_200, __p2_200)))); \
58683  __ret_200; \
58684})
58685#else
58686#define vqshrun_high_n_s32(__p0_201, __p1_201, __p2_201) __extension__ ({ \
58687  int16x4_t __s0_201 = __p0_201; \
58688  int32x4_t __s1_201 = __p1_201; \
58689  int16x4_t __rev0_201;  __rev0_201 = __builtin_shufflevector(__s0_201, __s0_201, 3, 2, 1, 0); \
58690  int32x4_t __rev1_201;  __rev1_201 = __builtin_shufflevector(__s1_201, __s1_201, 3, 2, 1, 0); \
58691  int16x8_t __ret_201; \
58692  __ret_201 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_201), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_201, __p2_201)))); \
58693  __ret_201 = __builtin_shufflevector(__ret_201, __ret_201, 7, 6, 5, 4, 3, 2, 1, 0); \
58694  __ret_201; \
58695})
58696#endif
58697
58698#ifdef __LITTLE_ENDIAN__
58699#define vqshrun_high_n_s64(__p0_202, __p1_202, __p2_202) __extension__ ({ \
58700  int32x2_t __s0_202 = __p0_202; \
58701  int64x2_t __s1_202 = __p1_202; \
58702  int32x4_t __ret_202; \
58703  __ret_202 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_202), (int32x2_t)(vqshrun_n_s64(__s1_202, __p2_202)))); \
58704  __ret_202; \
58705})
58706#else
58707#define vqshrun_high_n_s64(__p0_203, __p1_203, __p2_203) __extension__ ({ \
58708  int32x2_t __s0_203 = __p0_203; \
58709  int64x2_t __s1_203 = __p1_203; \
58710  int32x2_t __rev0_203;  __rev0_203 = __builtin_shufflevector(__s0_203, __s0_203, 1, 0); \
58711  int64x2_t __rev1_203;  __rev1_203 = __builtin_shufflevector(__s1_203, __s1_203, 1, 0); \
58712  int32x4_t __ret_203; \
58713  __ret_203 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_203), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_203, __p2_203)))); \
58714  __ret_203 = __builtin_shufflevector(__ret_203, __ret_203, 3, 2, 1, 0); \
58715  __ret_203; \
58716})
58717#endif
58718
58719#ifdef __LITTLE_ENDIAN__
58720#define vqshrun_high_n_s16(__p0_204, __p1_204, __p2_204) __extension__ ({ \
58721  int8x8_t __s0_204 = __p0_204; \
58722  int16x8_t __s1_204 = __p1_204; \
58723  int8x16_t __ret_204; \
58724  __ret_204 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_204), (int8x8_t)(vqshrun_n_s16(__s1_204, __p2_204)))); \
58725  __ret_204; \
58726})
58727#else
58728#define vqshrun_high_n_s16(__p0_205, __p1_205, __p2_205) __extension__ ({ \
58729  int8x8_t __s0_205 = __p0_205; \
58730  int16x8_t __s1_205 = __p1_205; \
58731  int8x8_t __rev0_205;  __rev0_205 = __builtin_shufflevector(__s0_205, __s0_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58732  int16x8_t __rev1_205;  __rev1_205 = __builtin_shufflevector(__s1_205, __s1_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58733  int8x16_t __ret_205; \
58734  __ret_205 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_205), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_205, __p2_205)))); \
58735  __ret_205 = __builtin_shufflevector(__ret_205, __ret_205, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58736  __ret_205; \
58737})
58738#endif
58739
58740#ifdef __LITTLE_ENDIAN__
58741#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58742  int32_t __s0 = __p0; \
58743  int16_t __ret; \
58744  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58745  __ret; \
58746})
58747#else
58748#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58749  int32_t __s0 = __p0; \
58750  int16_t __ret; \
58751  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58752  __ret; \
58753})
58754#endif
58755
58756#ifdef __LITTLE_ENDIAN__
58757#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58758  int64_t __s0 = __p0; \
58759  int32_t __ret; \
58760  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58761  __ret; \
58762})
58763#else
58764#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58765  int64_t __s0 = __p0; \
58766  int32_t __ret; \
58767  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58768  __ret; \
58769})
58770#endif
58771
58772#ifdef __LITTLE_ENDIAN__
58773#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58774  int16_t __s0 = __p0; \
58775  int8_t __ret; \
58776  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58777  __ret; \
58778})
58779#else
58780#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58781  int16_t __s0 = __p0; \
58782  int8_t __ret; \
58783  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58784  __ret; \
58785})
58786#endif
58787
58788#ifdef __LITTLE_ENDIAN__
58789__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58790  uint8_t __ret;
58791  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58792  return __ret;
58793}
58794#else
58795__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58796  uint8_t __ret;
58797  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58798  return __ret;
58799}
58800#endif
58801
58802#ifdef __LITTLE_ENDIAN__
58803__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58804  uint32_t __ret;
58805  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58806  return __ret;
58807}
58808#else
58809__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58810  uint32_t __ret;
58811  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58812  return __ret;
58813}
58814#endif
58815
58816#ifdef __LITTLE_ENDIAN__
58817__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58818  uint64_t __ret;
58819  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58820  return __ret;
58821}
58822#else
58823__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58824  uint64_t __ret;
58825  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58826  return __ret;
58827}
58828#endif
58829
58830#ifdef __LITTLE_ENDIAN__
58831__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58832  uint16_t __ret;
58833  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58834  return __ret;
58835}
58836#else
58837__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58838  uint16_t __ret;
58839  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58840  return __ret;
58841}
58842#endif
58843
58844#ifdef __LITTLE_ENDIAN__
58845__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58846  int8_t __ret;
58847  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58848  return __ret;
58849}
58850#else
58851__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58852  int8_t __ret;
58853  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58854  return __ret;
58855}
58856#endif
58857
58858#ifdef __LITTLE_ENDIAN__
58859__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58860  int32_t __ret;
58861  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58862  return __ret;
58863}
58864#else
58865__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58866  int32_t __ret;
58867  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58868  return __ret;
58869}
58870__ai int32_t __noswap_vqsubs_s32(int32_t __p0, int32_t __p1) {
58871  int32_t __ret;
58872  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58873  return __ret;
58874}
58875#endif
58876
58877#ifdef __LITTLE_ENDIAN__
58878__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58879  int64_t __ret;
58880  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58881  return __ret;
58882}
58883#else
58884__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58885  int64_t __ret;
58886  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58887  return __ret;
58888}
58889#endif
58890
58891#ifdef __LITTLE_ENDIAN__
58892__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58893  int16_t __ret;
58894  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58895  return __ret;
58896}
58897#else
58898__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58899  int16_t __ret;
58900  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58901  return __ret;
58902}
58903__ai int16_t __noswap_vqsubh_s16(int16_t __p0, int16_t __p1) {
58904  int16_t __ret;
58905  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58906  return __ret;
58907}
58908#endif
58909
58910#ifdef __LITTLE_ENDIAN__
58911__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58912  poly8x8_t __ret;
58913  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
58914  return __ret;
58915}
58916#else
58917__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58918  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58919  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58920  poly8x8_t __ret;
58921  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
58922  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58923  return __ret;
58924}
58925#endif
58926
58927#ifdef __LITTLE_ENDIAN__
58928__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58929  poly8x16_t __ret;
58930  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
58931  return __ret;
58932}
58933#else
58934__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58935  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58936  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58937  poly8x16_t __ret;
58938  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
58939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58940  return __ret;
58941}
58942#endif
58943
58944#ifdef __LITTLE_ENDIAN__
58945__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58946  uint8x16_t __ret;
58947  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
58948  return __ret;
58949}
58950#else
58951__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58952  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58953  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58954  uint8x16_t __ret;
58955  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
58956  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58957  return __ret;
58958}
58959#endif
58960
58961#ifdef __LITTLE_ENDIAN__
58962__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58963  int8x16_t __ret;
58964  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
58965  return __ret;
58966}
58967#else
58968__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58969  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58970  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58971  int8x16_t __ret;
58972  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
58973  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58974  return __ret;
58975}
58976#endif
58977
58978#ifdef __LITTLE_ENDIAN__
58979__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58980  uint8x8_t __ret;
58981  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
58982  return __ret;
58983}
58984#else
58985__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58986  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58987  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58988  uint8x8_t __ret;
58989  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
58990  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58991  return __ret;
58992}
58993#endif
58994
58995#ifdef __LITTLE_ENDIAN__
58996__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
58997  int8x8_t __ret;
58998  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
58999  return __ret;
59000}
59001#else
59002__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
59003  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59004  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59005  int8x8_t __ret;
59006  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
59007  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59008  return __ret;
59009}
59010#endif
59011
59012#ifdef __LITTLE_ENDIAN__
59013__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59014  poly8x8_t __ret;
59015  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
59016  return __ret;
59017}
59018#else
59019__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59020  poly8x16x2_t __rev0;
59021  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59022  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59023  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59024  poly8x8_t __ret;
59025  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
59026  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59027  return __ret;
59028}
59029#endif
59030
59031#ifdef __LITTLE_ENDIAN__
59032__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59033  poly8x16_t __ret;
59034  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
59035  return __ret;
59036}
59037#else
59038__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59039  poly8x16x2_t __rev0;
59040  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59041  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59042  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59043  poly8x16_t __ret;
59044  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
59045  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59046  return __ret;
59047}
59048#endif
59049
59050#ifdef __LITTLE_ENDIAN__
59051__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59052  uint8x16_t __ret;
59053  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
59054  return __ret;
59055}
59056#else
59057__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59058  uint8x16x2_t __rev0;
59059  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59060  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59061  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59062  uint8x16_t __ret;
59063  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
59064  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59065  return __ret;
59066}
59067#endif
59068
59069#ifdef __LITTLE_ENDIAN__
59070__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59071  int8x16_t __ret;
59072  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
59073  return __ret;
59074}
59075#else
59076__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59077  int8x16x2_t __rev0;
59078  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59079  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59080  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59081  int8x16_t __ret;
59082  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
59083  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59084  return __ret;
59085}
59086#endif
59087
59088#ifdef __LITTLE_ENDIAN__
59089__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59090  uint8x8_t __ret;
59091  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
59092  return __ret;
59093}
59094#else
59095__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59096  uint8x16x2_t __rev0;
59097  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59098  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59099  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59100  uint8x8_t __ret;
59101  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
59102  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59103  return __ret;
59104}
59105#endif
59106
59107#ifdef __LITTLE_ENDIAN__
59108__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59109  int8x8_t __ret;
59110  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
59111  return __ret;
59112}
59113#else
59114__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59115  int8x16x2_t __rev0;
59116  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59117  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59118  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59119  int8x8_t __ret;
59120  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
59121  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59122  return __ret;
59123}
59124#endif
59125
59126#ifdef __LITTLE_ENDIAN__
59127__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59128  poly8x8_t __ret;
59129  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
59130  return __ret;
59131}
59132#else
59133__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59134  poly8x16x3_t __rev0;
59135  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59136  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59137  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59138  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59139  poly8x8_t __ret;
59140  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
59141  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59142  return __ret;
59143}
59144#endif
59145
59146#ifdef __LITTLE_ENDIAN__
59147__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59148  poly8x16_t __ret;
59149  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
59150  return __ret;
59151}
59152#else
59153__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59154  poly8x16x3_t __rev0;
59155  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59156  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59157  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59158  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59159  poly8x16_t __ret;
59160  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
59161  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59162  return __ret;
59163}
59164#endif
59165
59166#ifdef __LITTLE_ENDIAN__
59167__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59168  uint8x16_t __ret;
59169  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
59170  return __ret;
59171}
59172#else
59173__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59174  uint8x16x3_t __rev0;
59175  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59176  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59177  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59178  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59179  uint8x16_t __ret;
59180  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
59181  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59182  return __ret;
59183}
59184#endif
59185
59186#ifdef __LITTLE_ENDIAN__
59187__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59188  int8x16_t __ret;
59189  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
59190  return __ret;
59191}
59192#else
59193__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59194  int8x16x3_t __rev0;
59195  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59196  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59197  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59198  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59199  int8x16_t __ret;
59200  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
59201  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59202  return __ret;
59203}
59204#endif
59205
59206#ifdef __LITTLE_ENDIAN__
59207__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59208  uint8x8_t __ret;
59209  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
59210  return __ret;
59211}
59212#else
59213__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59214  uint8x16x3_t __rev0;
59215  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59216  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59217  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59218  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59219  uint8x8_t __ret;
59220  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
59221  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59222  return __ret;
59223}
59224#endif
59225
59226#ifdef __LITTLE_ENDIAN__
59227__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59228  int8x8_t __ret;
59229  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
59230  return __ret;
59231}
59232#else
59233__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59234  int8x16x3_t __rev0;
59235  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59236  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59237  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59238  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59239  int8x8_t __ret;
59240  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
59241  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59242  return __ret;
59243}
59244#endif
59245
59246#ifdef __LITTLE_ENDIAN__
59247__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59248  poly8x8_t __ret;
59249  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
59250  return __ret;
59251}
59252#else
59253__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59254  poly8x16x4_t __rev0;
59255  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59256  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59257  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59258  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59259  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59260  poly8x8_t __ret;
59261  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
59262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59263  return __ret;
59264}
59265#endif
59266
59267#ifdef __LITTLE_ENDIAN__
59268__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59269  poly8x16_t __ret;
59270  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
59271  return __ret;
59272}
59273#else
59274__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59275  poly8x16x4_t __rev0;
59276  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59277  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59278  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59279  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59280  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59281  poly8x16_t __ret;
59282  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
59283  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59284  return __ret;
59285}
59286#endif
59287
59288#ifdef __LITTLE_ENDIAN__
59289__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59290  uint8x16_t __ret;
59291  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
59292  return __ret;
59293}
59294#else
59295__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59296  uint8x16x4_t __rev0;
59297  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59298  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59299  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59300  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59301  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59302  uint8x16_t __ret;
59303  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
59304  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59305  return __ret;
59306}
59307#endif
59308
59309#ifdef __LITTLE_ENDIAN__
59310__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59311  int8x16_t __ret;
59312  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
59313  return __ret;
59314}
59315#else
59316__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59317  int8x16x4_t __rev0;
59318  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59319  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59320  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59321  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59322  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59323  int8x16_t __ret;
59324  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
59325  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59326  return __ret;
59327}
59328#endif
59329
59330#ifdef __LITTLE_ENDIAN__
59331__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59332  uint8x8_t __ret;
59333  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
59334  return __ret;
59335}
59336#else
59337__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59338  uint8x16x4_t __rev0;
59339  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59340  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59341  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59342  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59343  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59344  uint8x8_t __ret;
59345  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
59346  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59347  return __ret;
59348}
59349#endif
59350
59351#ifdef __LITTLE_ENDIAN__
59352__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59353  int8x8_t __ret;
59354  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
59355  return __ret;
59356}
59357#else
59358__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59359  int8x16x4_t __rev0;
59360  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59361  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59362  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59363  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59364  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59365  int8x8_t __ret;
59366  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
59367  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59368  return __ret;
59369}
59370#endif
59371
59372#ifdef __LITTLE_ENDIAN__
59373__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59374  poly8x8_t __ret;
59375  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
59376  return __ret;
59377}
59378#else
59379__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59380  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59381  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59382  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59383  poly8x8_t __ret;
59384  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
59385  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59386  return __ret;
59387}
59388#endif
59389
59390#ifdef __LITTLE_ENDIAN__
59391__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59392  poly8x16_t __ret;
59393  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
59394  return __ret;
59395}
59396#else
59397__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59398  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59399  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59400  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59401  poly8x16_t __ret;
59402  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
59403  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59404  return __ret;
59405}
59406#endif
59407
59408#ifdef __LITTLE_ENDIAN__
59409__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59410  uint8x16_t __ret;
59411  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
59412  return __ret;
59413}
59414#else
59415__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59416  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59417  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59418  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59419  uint8x16_t __ret;
59420  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
59421  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59422  return __ret;
59423}
59424#endif
59425
59426#ifdef __LITTLE_ENDIAN__
59427__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59428  int8x16_t __ret;
59429  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
59430  return __ret;
59431}
59432#else
59433__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59434  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59435  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59436  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59437  int8x16_t __ret;
59438  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
59439  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59440  return __ret;
59441}
59442#endif
59443
59444#ifdef __LITTLE_ENDIAN__
59445__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59446  uint8x8_t __ret;
59447  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
59448  return __ret;
59449}
59450#else
59451__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59452  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59453  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59454  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59455  uint8x8_t __ret;
59456  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
59457  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59458  return __ret;
59459}
59460#endif
59461
59462#ifdef __LITTLE_ENDIAN__
59463__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59464  int8x8_t __ret;
59465  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
59466  return __ret;
59467}
59468#else
59469__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59470  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59471  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59472  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59473  int8x8_t __ret;
59474  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
59475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59476  return __ret;
59477}
59478#endif
59479
59480#ifdef __LITTLE_ENDIAN__
59481__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59482  poly8x8_t __ret;
59483  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
59484  return __ret;
59485}
59486#else
59487__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59488  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59489  poly8x16x2_t __rev1;
59490  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59491  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59492  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59493  poly8x8_t __ret;
59494  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
59495  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59496  return __ret;
59497}
59498#endif
59499
59500#ifdef __LITTLE_ENDIAN__
59501__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59502  poly8x16_t __ret;
59503  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
59504  return __ret;
59505}
59506#else
59507__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59508  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59509  poly8x16x2_t __rev1;
59510  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59511  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59512  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59513  poly8x16_t __ret;
59514  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
59515  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59516  return __ret;
59517}
59518#endif
59519
59520#ifdef __LITTLE_ENDIAN__
59521__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59522  uint8x16_t __ret;
59523  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
59524  return __ret;
59525}
59526#else
59527__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59528  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59529  uint8x16x2_t __rev1;
59530  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59531  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59532  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59533  uint8x16_t __ret;
59534  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
59535  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59536  return __ret;
59537}
59538#endif
59539
59540#ifdef __LITTLE_ENDIAN__
59541__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59542  int8x16_t __ret;
59543  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
59544  return __ret;
59545}
59546#else
59547__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59548  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59549  int8x16x2_t __rev1;
59550  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59551  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59552  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59553  int8x16_t __ret;
59554  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
59555  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59556  return __ret;
59557}
59558#endif
59559
59560#ifdef __LITTLE_ENDIAN__
59561__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59562  uint8x8_t __ret;
59563  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
59564  return __ret;
59565}
59566#else
59567__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59568  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59569  uint8x16x2_t __rev1;
59570  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59571  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59572  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59573  uint8x8_t __ret;
59574  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
59575  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59576  return __ret;
59577}
59578#endif
59579
59580#ifdef __LITTLE_ENDIAN__
59581__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59582  int8x8_t __ret;
59583  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
59584  return __ret;
59585}
59586#else
59587__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59588  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59589  int8x16x2_t __rev1;
59590  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59591  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59592  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59593  int8x8_t __ret;
59594  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
59595  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59596  return __ret;
59597}
59598#endif
59599
59600#ifdef __LITTLE_ENDIAN__
59601__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59602  poly8x8_t __ret;
59603  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
59604  return __ret;
59605}
59606#else
59607__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59608  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59609  poly8x16x3_t __rev1;
59610  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59611  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59612  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59613  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59614  poly8x8_t __ret;
59615  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
59616  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59617  return __ret;
59618}
59619#endif
59620
59621#ifdef __LITTLE_ENDIAN__
59622__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59623  poly8x16_t __ret;
59624  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
59625  return __ret;
59626}
59627#else
59628__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59629  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59630  poly8x16x3_t __rev1;
59631  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59632  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59633  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59634  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59635  poly8x16_t __ret;
59636  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
59637  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59638  return __ret;
59639}
59640#endif
59641
59642#ifdef __LITTLE_ENDIAN__
59643__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59644  uint8x16_t __ret;
59645  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
59646  return __ret;
59647}
59648#else
59649__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59650  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59651  uint8x16x3_t __rev1;
59652  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59653  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59654  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59655  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59656  uint8x16_t __ret;
59657  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
59658  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59659  return __ret;
59660}
59661#endif
59662
59663#ifdef __LITTLE_ENDIAN__
59664__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59665  int8x16_t __ret;
59666  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
59667  return __ret;
59668}
59669#else
59670__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59671  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59672  int8x16x3_t __rev1;
59673  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59674  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59675  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59676  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59677  int8x16_t __ret;
59678  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
59679  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59680  return __ret;
59681}
59682#endif
59683
59684#ifdef __LITTLE_ENDIAN__
59685__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59686  uint8x8_t __ret;
59687  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
59688  return __ret;
59689}
59690#else
59691__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59692  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59693  uint8x16x3_t __rev1;
59694  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59695  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59696  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59697  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59698  uint8x8_t __ret;
59699  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
59700  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59701  return __ret;
59702}
59703#endif
59704
59705#ifdef __LITTLE_ENDIAN__
59706__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59707  int8x8_t __ret;
59708  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
59709  return __ret;
59710}
59711#else
59712__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59713  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59714  int8x16x3_t __rev1;
59715  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59716  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59717  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59718  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59719  int8x8_t __ret;
59720  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
59721  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59722  return __ret;
59723}
59724#endif
59725
59726#ifdef __LITTLE_ENDIAN__
59727__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59728  poly8x8_t __ret;
59729  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
59730  return __ret;
59731}
59732#else
59733__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59734  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59735  poly8x16x4_t __rev1;
59736  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59737  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59738  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59739  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59740  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59741  poly8x8_t __ret;
59742  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
59743  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59744  return __ret;
59745}
59746#endif
59747
59748#ifdef __LITTLE_ENDIAN__
59749__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59750  poly8x16_t __ret;
59751  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
59752  return __ret;
59753}
59754#else
59755__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59756  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59757  poly8x16x4_t __rev1;
59758  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59759  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59760  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59761  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59762  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59763  poly8x16_t __ret;
59764  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
59765  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59766  return __ret;
59767}
59768#endif
59769
59770#ifdef __LITTLE_ENDIAN__
59771__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59772  uint8x16_t __ret;
59773  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
59774  return __ret;
59775}
59776#else
59777__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59778  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59779  uint8x16x4_t __rev1;
59780  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59781  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59782  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59783  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59784  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59785  uint8x16_t __ret;
59786  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
59787  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59788  return __ret;
59789}
59790#endif
59791
59792#ifdef __LITTLE_ENDIAN__
59793__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59794  int8x16_t __ret;
59795  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
59796  return __ret;
59797}
59798#else
59799__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59800  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59801  int8x16x4_t __rev1;
59802  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59803  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59804  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59805  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59806  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59807  int8x16_t __ret;
59808  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
59809  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59810  return __ret;
59811}
59812#endif
59813
59814#ifdef __LITTLE_ENDIAN__
59815__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59816  uint8x8_t __ret;
59817  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
59818  return __ret;
59819}
59820#else
59821__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59822  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59823  uint8x16x4_t __rev1;
59824  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59825  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59826  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59827  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59828  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59829  uint8x8_t __ret;
59830  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
59831  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59832  return __ret;
59833}
59834#endif
59835
59836#ifdef __LITTLE_ENDIAN__
59837__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59838  int8x8_t __ret;
59839  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
59840  return __ret;
59841}
59842#else
59843__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59844  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59845  int8x16x4_t __rev1;
59846  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59847  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59848  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59849  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59850  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59851  int8x8_t __ret;
59852  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
59853  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59854  return __ret;
59855}
59856#endif
59857
59858#ifdef __LITTLE_ENDIAN__
59859__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59860  uint16x8_t __ret;
59861  __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
59862  return __ret;
59863}
59864#else
59865__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59866  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59867  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59868  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59869  uint16x8_t __ret;
59870  __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
59871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59872  return __ret;
59873}
59874#endif
59875
59876#ifdef __LITTLE_ENDIAN__
59877__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59878  uint32x4_t __ret;
59879  __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
59880  return __ret;
59881}
59882#else
59883__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59884  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59885  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59886  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59887  uint32x4_t __ret;
59888  __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
59889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59890  return __ret;
59891}
59892#endif
59893
59894#ifdef __LITTLE_ENDIAN__
59895__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59896  uint8x16_t __ret;
59897  __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
59898  return __ret;
59899}
59900#else
59901__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59902  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59903  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59904  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59905  uint8x16_t __ret;
59906  __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
59907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59908  return __ret;
59909}
59910#endif
59911
59912#ifdef __LITTLE_ENDIAN__
59913__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59914  int16x8_t __ret;
59915  __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
59916  return __ret;
59917}
59918#else
59919__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59920  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59921  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59922  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59923  int16x8_t __ret;
59924  __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
59925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59926  return __ret;
59927}
59928#endif
59929
59930#ifdef __LITTLE_ENDIAN__
59931__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59932  int32x4_t __ret;
59933  __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
59934  return __ret;
59935}
59936#else
59937__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59938  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59939  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59940  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59941  int32x4_t __ret;
59942  __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
59943  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59944  return __ret;
59945}
59946#endif
59947
59948#ifdef __LITTLE_ENDIAN__
59949__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59950  int8x16_t __ret;
59951  __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
59952  return __ret;
59953}
59954#else
59955__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59956  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59957  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59958  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59959  int8x16_t __ret;
59960  __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
59961  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59962  return __ret;
59963}
59964#endif
59965
59966#ifdef __LITTLE_ENDIAN__
59967__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59968  poly8x8_t __ret;
59969  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
59970  return __ret;
59971}
59972#else
59973__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59974  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59975  poly8x8_t __ret;
59976  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
59977  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59978  return __ret;
59979}
59980#endif
59981
59982#ifdef __LITTLE_ENDIAN__
59983__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59984  poly8x16_t __ret;
59985  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
59986  return __ret;
59987}
59988#else
59989__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59990  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59991  poly8x16_t __ret;
59992  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
59993  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59994  return __ret;
59995}
59996#endif
59997
59998#ifdef __LITTLE_ENDIAN__
59999__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60000  uint8x16_t __ret;
60001  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
60002  return __ret;
60003}
60004#else
60005__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60006  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60007  uint8x16_t __ret;
60008  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
60009  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60010  return __ret;
60011}
60012#endif
60013
60014#ifdef __LITTLE_ENDIAN__
60015__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60016  int8x16_t __ret;
60017  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
60018  return __ret;
60019}
60020#else
60021__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60022  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60023  int8x16_t __ret;
60024  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
60025  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60026  return __ret;
60027}
60028#endif
60029
60030#ifdef __LITTLE_ENDIAN__
60031__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60032  uint8x8_t __ret;
60033  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
60034  return __ret;
60035}
60036#else
60037__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60038  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60039  uint8x8_t __ret;
60040  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
60041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60042  return __ret;
60043}
60044#endif
60045
60046#ifdef __LITTLE_ENDIAN__
60047__ai int8x8_t vrbit_s8(int8x8_t __p0) {
60048  int8x8_t __ret;
60049  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
60050  return __ret;
60051}
60052#else
60053__ai int8x8_t vrbit_s8(int8x8_t __p0) {
60054  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60055  int8x8_t __ret;
60056  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
60057  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60058  return __ret;
60059}
60060#endif
60061
60062#ifdef __LITTLE_ENDIAN__
60063__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60064  float64x2_t __ret;
60065  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
60066  return __ret;
60067}
60068#else
60069__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60070  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60071  float64x2_t __ret;
60072  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
60073  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60074  return __ret;
60075}
60076#endif
60077
60078#ifdef __LITTLE_ENDIAN__
60079__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60080  float64x1_t __ret;
60081  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60082  return __ret;
60083}
60084#else
60085__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60086  float64x1_t __ret;
60087  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60088  return __ret;
60089}
60090#endif
60091
60092#ifdef __LITTLE_ENDIAN__
60093__ai float64_t vrecped_f64(float64_t __p0) {
60094  float64_t __ret;
60095  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60096  return __ret;
60097}
60098#else
60099__ai float64_t vrecped_f64(float64_t __p0) {
60100  float64_t __ret;
60101  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60102  return __ret;
60103}
60104#endif
60105
60106#ifdef __LITTLE_ENDIAN__
60107__ai float32_t vrecpes_f32(float32_t __p0) {
60108  float32_t __ret;
60109  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60110  return __ret;
60111}
60112#else
60113__ai float32_t vrecpes_f32(float32_t __p0) {
60114  float32_t __ret;
60115  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60116  return __ret;
60117}
60118#endif
60119
60120#ifdef __LITTLE_ENDIAN__
60121__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60122  float64x2_t __ret;
60123  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60124  return __ret;
60125}
60126#else
60127__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60128  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60129  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60130  float64x2_t __ret;
60131  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60132  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60133  return __ret;
60134}
60135#endif
60136
60137#ifdef __LITTLE_ENDIAN__
60138__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60139  float64x1_t __ret;
60140  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60141  return __ret;
60142}
60143#else
60144__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60145  float64x1_t __ret;
60146  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60147  return __ret;
60148}
60149#endif
60150
60151#ifdef __LITTLE_ENDIAN__
60152__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60153  float64_t __ret;
60154  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60155  return __ret;
60156}
60157#else
60158__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60159  float64_t __ret;
60160  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60161  return __ret;
60162}
60163#endif
60164
60165#ifdef __LITTLE_ENDIAN__
60166__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60167  float32_t __ret;
60168  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60169  return __ret;
60170}
60171#else
60172__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60173  float32_t __ret;
60174  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60175  return __ret;
60176}
60177#endif
60178
60179#ifdef __LITTLE_ENDIAN__
60180__ai float64_t vrecpxd_f64(float64_t __p0) {
60181  float64_t __ret;
60182  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60183  return __ret;
60184}
60185#else
60186__ai float64_t vrecpxd_f64(float64_t __p0) {
60187  float64_t __ret;
60188  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60189  return __ret;
60190}
60191#endif
60192
60193#ifdef __LITTLE_ENDIAN__
60194__ai float32_t vrecpxs_f32(float32_t __p0) {
60195  float32_t __ret;
60196  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60197  return __ret;
60198}
60199#else
60200__ai float32_t vrecpxs_f32(float32_t __p0) {
60201  float32_t __ret;
60202  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60203  return __ret;
60204}
60205#endif
60206
60207#ifdef __LITTLE_ENDIAN__
60208__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60209  uint64_t __ret;
60210  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60211  return __ret;
60212}
60213#else
60214__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60215  uint64_t __ret;
60216  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60217  return __ret;
60218}
60219#endif
60220
60221#ifdef __LITTLE_ENDIAN__
60222__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60223  int64_t __ret;
60224  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60225  return __ret;
60226}
60227#else
60228__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60229  int64_t __ret;
60230  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60231  return __ret;
60232}
60233#endif
60234
60235#ifdef __LITTLE_ENDIAN__
60236#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60237  uint64_t __s0 = __p0; \
60238  uint64_t __ret; \
60239  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60240  __ret; \
60241})
60242#else
60243#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60244  uint64_t __s0 = __p0; \
60245  uint64_t __ret; \
60246  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60247  __ret; \
60248})
60249#endif
60250
60251#ifdef __LITTLE_ENDIAN__
60252#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60253  int64_t __s0 = __p0; \
60254  int64_t __ret; \
60255  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60256  __ret; \
60257})
60258#else
60259#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60260  int64_t __s0 = __p0; \
60261  int64_t __ret; \
60262  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60263  __ret; \
60264})
60265#endif
60266
60267#ifdef __LITTLE_ENDIAN__
60268#define vrshrn_high_n_u32(__p0_206, __p1_206, __p2_206) __extension__ ({ \
60269  uint16x4_t __s0_206 = __p0_206; \
60270  uint32x4_t __s1_206 = __p1_206; \
60271  uint16x8_t __ret_206; \
60272  __ret_206 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_206), (uint16x4_t)(vrshrn_n_u32(__s1_206, __p2_206)))); \
60273  __ret_206; \
60274})
60275#else
60276#define vrshrn_high_n_u32(__p0_207, __p1_207, __p2_207) __extension__ ({ \
60277  uint16x4_t __s0_207 = __p0_207; \
60278  uint32x4_t __s1_207 = __p1_207; \
60279  uint16x4_t __rev0_207;  __rev0_207 = __builtin_shufflevector(__s0_207, __s0_207, 3, 2, 1, 0); \
60280  uint32x4_t __rev1_207;  __rev1_207 = __builtin_shufflevector(__s1_207, __s1_207, 3, 2, 1, 0); \
60281  uint16x8_t __ret_207; \
60282  __ret_207 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_207), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_207, __p2_207)))); \
60283  __ret_207 = __builtin_shufflevector(__ret_207, __ret_207, 7, 6, 5, 4, 3, 2, 1, 0); \
60284  __ret_207; \
60285})
60286#endif
60287
60288#ifdef __LITTLE_ENDIAN__
60289#define vrshrn_high_n_u64(__p0_208, __p1_208, __p2_208) __extension__ ({ \
60290  uint32x2_t __s0_208 = __p0_208; \
60291  uint64x2_t __s1_208 = __p1_208; \
60292  uint32x4_t __ret_208; \
60293  __ret_208 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_208), (uint32x2_t)(vrshrn_n_u64(__s1_208, __p2_208)))); \
60294  __ret_208; \
60295})
60296#else
60297#define vrshrn_high_n_u64(__p0_209, __p1_209, __p2_209) __extension__ ({ \
60298  uint32x2_t __s0_209 = __p0_209; \
60299  uint64x2_t __s1_209 = __p1_209; \
60300  uint32x2_t __rev0_209;  __rev0_209 = __builtin_shufflevector(__s0_209, __s0_209, 1, 0); \
60301  uint64x2_t __rev1_209;  __rev1_209 = __builtin_shufflevector(__s1_209, __s1_209, 1, 0); \
60302  uint32x4_t __ret_209; \
60303  __ret_209 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_209), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_209, __p2_209)))); \
60304  __ret_209 = __builtin_shufflevector(__ret_209, __ret_209, 3, 2, 1, 0); \
60305  __ret_209; \
60306})
60307#endif
60308
60309#ifdef __LITTLE_ENDIAN__
60310#define vrshrn_high_n_u16(__p0_210, __p1_210, __p2_210) __extension__ ({ \
60311  uint8x8_t __s0_210 = __p0_210; \
60312  uint16x8_t __s1_210 = __p1_210; \
60313  uint8x16_t __ret_210; \
60314  __ret_210 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_210), (uint8x8_t)(vrshrn_n_u16(__s1_210, __p2_210)))); \
60315  __ret_210; \
60316})
60317#else
60318#define vrshrn_high_n_u16(__p0_211, __p1_211, __p2_211) __extension__ ({ \
60319  uint8x8_t __s0_211 = __p0_211; \
60320  uint16x8_t __s1_211 = __p1_211; \
60321  uint8x8_t __rev0_211;  __rev0_211 = __builtin_shufflevector(__s0_211, __s0_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60322  uint16x8_t __rev1_211;  __rev1_211 = __builtin_shufflevector(__s1_211, __s1_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60323  uint8x16_t __ret_211; \
60324  __ret_211 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_211), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_211, __p2_211)))); \
60325  __ret_211 = __builtin_shufflevector(__ret_211, __ret_211, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60326  __ret_211; \
60327})
60328#endif
60329
60330#ifdef __LITTLE_ENDIAN__
60331#define vrshrn_high_n_s32(__p0_212, __p1_212, __p2_212) __extension__ ({ \
60332  int16x4_t __s0_212 = __p0_212; \
60333  int32x4_t __s1_212 = __p1_212; \
60334  int16x8_t __ret_212; \
60335  __ret_212 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_212), (int16x4_t)(vrshrn_n_s32(__s1_212, __p2_212)))); \
60336  __ret_212; \
60337})
60338#else
60339#define vrshrn_high_n_s32(__p0_213, __p1_213, __p2_213) __extension__ ({ \
60340  int16x4_t __s0_213 = __p0_213; \
60341  int32x4_t __s1_213 = __p1_213; \
60342  int16x4_t __rev0_213;  __rev0_213 = __builtin_shufflevector(__s0_213, __s0_213, 3, 2, 1, 0); \
60343  int32x4_t __rev1_213;  __rev1_213 = __builtin_shufflevector(__s1_213, __s1_213, 3, 2, 1, 0); \
60344  int16x8_t __ret_213; \
60345  __ret_213 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_213), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_213, __p2_213)))); \
60346  __ret_213 = __builtin_shufflevector(__ret_213, __ret_213, 7, 6, 5, 4, 3, 2, 1, 0); \
60347  __ret_213; \
60348})
60349#endif
60350
60351#ifdef __LITTLE_ENDIAN__
60352#define vrshrn_high_n_s64(__p0_214, __p1_214, __p2_214) __extension__ ({ \
60353  int32x2_t __s0_214 = __p0_214; \
60354  int64x2_t __s1_214 = __p1_214; \
60355  int32x4_t __ret_214; \
60356  __ret_214 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_214), (int32x2_t)(vrshrn_n_s64(__s1_214, __p2_214)))); \
60357  __ret_214; \
60358})
60359#else
60360#define vrshrn_high_n_s64(__p0_215, __p1_215, __p2_215) __extension__ ({ \
60361  int32x2_t __s0_215 = __p0_215; \
60362  int64x2_t __s1_215 = __p1_215; \
60363  int32x2_t __rev0_215;  __rev0_215 = __builtin_shufflevector(__s0_215, __s0_215, 1, 0); \
60364  int64x2_t __rev1_215;  __rev1_215 = __builtin_shufflevector(__s1_215, __s1_215, 1, 0); \
60365  int32x4_t __ret_215; \
60366  __ret_215 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_215), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_215, __p2_215)))); \
60367  __ret_215 = __builtin_shufflevector(__ret_215, __ret_215, 3, 2, 1, 0); \
60368  __ret_215; \
60369})
60370#endif
60371
60372#ifdef __LITTLE_ENDIAN__
60373#define vrshrn_high_n_s16(__p0_216, __p1_216, __p2_216) __extension__ ({ \
60374  int8x8_t __s0_216 = __p0_216; \
60375  int16x8_t __s1_216 = __p1_216; \
60376  int8x16_t __ret_216; \
60377  __ret_216 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_216), (int8x8_t)(vrshrn_n_s16(__s1_216, __p2_216)))); \
60378  __ret_216; \
60379})
60380#else
60381#define vrshrn_high_n_s16(__p0_217, __p1_217, __p2_217) __extension__ ({ \
60382  int8x8_t __s0_217 = __p0_217; \
60383  int16x8_t __s1_217 = __p1_217; \
60384  int8x8_t __rev0_217;  __rev0_217 = __builtin_shufflevector(__s0_217, __s0_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60385  int16x8_t __rev1_217;  __rev1_217 = __builtin_shufflevector(__s1_217, __s1_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60386  int8x16_t __ret_217; \
60387  __ret_217 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_217), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_217, __p2_217)))); \
60388  __ret_217 = __builtin_shufflevector(__ret_217, __ret_217, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60389  __ret_217; \
60390})
60391#endif
60392
60393#ifdef __LITTLE_ENDIAN__
60394__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60395  float64x2_t __ret;
60396  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
60397  return __ret;
60398}
60399#else
60400__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60401  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60402  float64x2_t __ret;
60403  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
60404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60405  return __ret;
60406}
60407#endif
60408
60409#ifdef __LITTLE_ENDIAN__
60410__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60411  float64x1_t __ret;
60412  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60413  return __ret;
60414}
60415#else
60416__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60417  float64x1_t __ret;
60418  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60419  return __ret;
60420}
60421#endif
60422
60423#ifdef __LITTLE_ENDIAN__
60424__ai float64_t vrsqrted_f64(float64_t __p0) {
60425  float64_t __ret;
60426  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60427  return __ret;
60428}
60429#else
60430__ai float64_t vrsqrted_f64(float64_t __p0) {
60431  float64_t __ret;
60432  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60433  return __ret;
60434}
60435#endif
60436
60437#ifdef __LITTLE_ENDIAN__
60438__ai float32_t vrsqrtes_f32(float32_t __p0) {
60439  float32_t __ret;
60440  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60441  return __ret;
60442}
60443#else
60444__ai float32_t vrsqrtes_f32(float32_t __p0) {
60445  float32_t __ret;
60446  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60447  return __ret;
60448}
60449#endif
60450
60451#ifdef __LITTLE_ENDIAN__
60452__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60453  float64x2_t __ret;
60454  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60455  return __ret;
60456}
60457#else
60458__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60459  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60460  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60461  float64x2_t __ret;
60462  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60463  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60464  return __ret;
60465}
60466#endif
60467
60468#ifdef __LITTLE_ENDIAN__
60469__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60470  float64x1_t __ret;
60471  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60472  return __ret;
60473}
60474#else
60475__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60476  float64x1_t __ret;
60477  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60478  return __ret;
60479}
60480#endif
60481
60482#ifdef __LITTLE_ENDIAN__
60483__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60484  float64_t __ret;
60485  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60486  return __ret;
60487}
60488#else
60489__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60490  float64_t __ret;
60491  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60492  return __ret;
60493}
60494#endif
60495
60496#ifdef __LITTLE_ENDIAN__
60497__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60498  float32_t __ret;
60499  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60500  return __ret;
60501}
60502#else
60503__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60504  float32_t __ret;
60505  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60506  return __ret;
60507}
60508#endif
60509
60510#ifdef __LITTLE_ENDIAN__
60511#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60512  uint64_t __s0 = __p0; \
60513  uint64_t __s1 = __p1; \
60514  uint64_t __ret; \
60515  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60516  __ret; \
60517})
60518#else
60519#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60520  uint64_t __s0 = __p0; \
60521  uint64_t __s1 = __p1; \
60522  uint64_t __ret; \
60523  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60524  __ret; \
60525})
60526#endif
60527
60528#ifdef __LITTLE_ENDIAN__
60529#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60530  int64_t __s0 = __p0; \
60531  int64_t __s1 = __p1; \
60532  int64_t __ret; \
60533  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60534  __ret; \
60535})
60536#else
60537#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60538  int64_t __s0 = __p0; \
60539  int64_t __s1 = __p1; \
60540  int64_t __ret; \
60541  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60542  __ret; \
60543})
60544#endif
60545
60546#ifdef __LITTLE_ENDIAN__
60547__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60548  uint16x8_t __ret;
60549  __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
60550  return __ret;
60551}
60552#else
60553__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60554  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60555  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60556  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60557  uint16x8_t __ret;
60558  __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
60559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60560  return __ret;
60561}
60562#endif
60563
60564#ifdef __LITTLE_ENDIAN__
60565__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60566  uint32x4_t __ret;
60567  __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
60568  return __ret;
60569}
60570#else
60571__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60572  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60573  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60574  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60575  uint32x4_t __ret;
60576  __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
60577  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60578  return __ret;
60579}
60580#endif
60581
60582#ifdef __LITTLE_ENDIAN__
60583__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60584  uint8x16_t __ret;
60585  __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
60586  return __ret;
60587}
60588#else
60589__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60590  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60591  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60592  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60593  uint8x16_t __ret;
60594  __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
60595  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60596  return __ret;
60597}
60598#endif
60599
60600#ifdef __LITTLE_ENDIAN__
60601__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60602  int16x8_t __ret;
60603  __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
60604  return __ret;
60605}
60606#else
60607__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60608  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60609  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60610  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60611  int16x8_t __ret;
60612  __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
60613  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60614  return __ret;
60615}
60616#endif
60617
60618#ifdef __LITTLE_ENDIAN__
60619__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60620  int32x4_t __ret;
60621  __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
60622  return __ret;
60623}
60624#else
60625__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60626  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60627  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60628  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60629  int32x4_t __ret;
60630  __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
60631  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60632  return __ret;
60633}
60634#endif
60635
60636#ifdef __LITTLE_ENDIAN__
60637__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60638  int8x16_t __ret;
60639  __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
60640  return __ret;
60641}
60642#else
60643__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60644  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60645  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60646  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60647  int8x16_t __ret;
60648  __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
60649  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60650  return __ret;
60651}
60652#endif
60653
60654#ifdef __LITTLE_ENDIAN__
60655#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60656  poly64_t __s0 = __p0; \
60657  poly64x1_t __s1 = __p1; \
60658  poly64x1_t __ret; \
60659  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60660  __ret; \
60661})
60662#else
60663#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60664  poly64_t __s0 = __p0; \
60665  poly64x1_t __s1 = __p1; \
60666  poly64x1_t __ret; \
60667  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60668  __ret; \
60669})
60670#define __noswap_vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60671  poly64_t __s0 = __p0; \
60672  poly64x1_t __s1 = __p1; \
60673  poly64x1_t __ret; \
60674  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60675  __ret; \
60676})
60677#endif
60678
60679#ifdef __LITTLE_ENDIAN__
60680#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60681  poly64_t __s0 = __p0; \
60682  poly64x2_t __s1 = __p1; \
60683  poly64x2_t __ret; \
60684  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60685  __ret; \
60686})
60687#else
60688#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60689  poly64_t __s0 = __p0; \
60690  poly64x2_t __s1 = __p1; \
60691  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60692  poly64x2_t __ret; \
60693  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
60694  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60695  __ret; \
60696})
60697#define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60698  poly64_t __s0 = __p0; \
60699  poly64x2_t __s1 = __p1; \
60700  poly64x2_t __ret; \
60701  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60702  __ret; \
60703})
60704#endif
60705
60706#ifdef __LITTLE_ENDIAN__
60707#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60708  float64_t __s0 = __p0; \
60709  float64x2_t __s1 = __p1; \
60710  float64x2_t __ret; \
60711  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60712  __ret; \
60713})
60714#else
60715#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60716  float64_t __s0 = __p0; \
60717  float64x2_t __s1 = __p1; \
60718  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60719  float64x2_t __ret; \
60720  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__rev1, __p2); \
60721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60722  __ret; \
60723})
60724#define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60725  float64_t __s0 = __p0; \
60726  float64x2_t __s1 = __p1; \
60727  float64x2_t __ret; \
60728  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60729  __ret; \
60730})
60731#endif
60732
60733#ifdef __LITTLE_ENDIAN__
60734#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60735  float64_t __s0 = __p0; \
60736  float64x1_t __s1 = __p1; \
60737  float64x1_t __ret; \
60738  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60739  __ret; \
60740})
60741#else
60742#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60743  float64_t __s0 = __p0; \
60744  float64x1_t __s1 = __p1; \
60745  float64x1_t __ret; \
60746  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60747  __ret; \
60748})
60749#define __noswap_vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60750  float64_t __s0 = __p0; \
60751  float64x1_t __s1 = __p1; \
60752  float64x1_t __ret; \
60753  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60754  __ret; \
60755})
60756#endif
60757
60758#ifdef __LITTLE_ENDIAN__
60759__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60760  uint64_t __ret;
60761  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60762  return __ret;
60763}
60764#else
60765__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60766  uint64_t __ret;
60767  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60768  return __ret;
60769}
60770#endif
60771
60772#ifdef __LITTLE_ENDIAN__
60773__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60774  int64_t __ret;
60775  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60776  return __ret;
60777}
60778#else
60779__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60780  int64_t __ret;
60781  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60782  return __ret;
60783}
60784#endif
60785
60786#ifdef __LITTLE_ENDIAN__
60787#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60788  uint64_t __s0 = __p0; \
60789  uint64_t __ret; \
60790  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60791  __ret; \
60792})
60793#else
60794#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60795  uint64_t __s0 = __p0; \
60796  uint64_t __ret; \
60797  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60798  __ret; \
60799})
60800#endif
60801
60802#ifdef __LITTLE_ENDIAN__
60803#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60804  int64_t __s0 = __p0; \
60805  int64_t __ret; \
60806  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60807  __ret; \
60808})
60809#else
60810#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60811  int64_t __s0 = __p0; \
60812  int64_t __ret; \
60813  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60814  __ret; \
60815})
60816#endif
60817
60818#ifdef __LITTLE_ENDIAN__
60819#define vshll_high_n_u8(__p0_218, __p1_218) __extension__ ({ \
60820  uint8x16_t __s0_218 = __p0_218; \
60821  uint16x8_t __ret_218; \
60822  __ret_218 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_218), __p1_218)); \
60823  __ret_218; \
60824})
60825#else
60826#define vshll_high_n_u8(__p0_219, __p1_219) __extension__ ({ \
60827  uint8x16_t __s0_219 = __p0_219; \
60828  uint8x16_t __rev0_219;  __rev0_219 = __builtin_shufflevector(__s0_219, __s0_219, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60829  uint16x8_t __ret_219; \
60830  __ret_219 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_219), __p1_219)); \
60831  __ret_219 = __builtin_shufflevector(__ret_219, __ret_219, 7, 6, 5, 4, 3, 2, 1, 0); \
60832  __ret_219; \
60833})
60834#endif
60835
60836#ifdef __LITTLE_ENDIAN__
60837#define vshll_high_n_u32(__p0_220, __p1_220) __extension__ ({ \
60838  uint32x4_t __s0_220 = __p0_220; \
60839  uint64x2_t __ret_220; \
60840  __ret_220 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_220), __p1_220)); \
60841  __ret_220; \
60842})
60843#else
60844#define vshll_high_n_u32(__p0_221, __p1_221) __extension__ ({ \
60845  uint32x4_t __s0_221 = __p0_221; \
60846  uint32x4_t __rev0_221;  __rev0_221 = __builtin_shufflevector(__s0_221, __s0_221, 3, 2, 1, 0); \
60847  uint64x2_t __ret_221; \
60848  __ret_221 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_221), __p1_221)); \
60849  __ret_221 = __builtin_shufflevector(__ret_221, __ret_221, 1, 0); \
60850  __ret_221; \
60851})
60852#endif
60853
60854#ifdef __LITTLE_ENDIAN__
60855#define vshll_high_n_u16(__p0_222, __p1_222) __extension__ ({ \
60856  uint16x8_t __s0_222 = __p0_222; \
60857  uint32x4_t __ret_222; \
60858  __ret_222 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_222), __p1_222)); \
60859  __ret_222; \
60860})
60861#else
60862#define vshll_high_n_u16(__p0_223, __p1_223) __extension__ ({ \
60863  uint16x8_t __s0_223 = __p0_223; \
60864  uint16x8_t __rev0_223;  __rev0_223 = __builtin_shufflevector(__s0_223, __s0_223, 7, 6, 5, 4, 3, 2, 1, 0); \
60865  uint32x4_t __ret_223; \
60866  __ret_223 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_223), __p1_223)); \
60867  __ret_223 = __builtin_shufflevector(__ret_223, __ret_223, 3, 2, 1, 0); \
60868  __ret_223; \
60869})
60870#endif
60871
60872#ifdef __LITTLE_ENDIAN__
60873#define vshll_high_n_s8(__p0_224, __p1_224) __extension__ ({ \
60874  int8x16_t __s0_224 = __p0_224; \
60875  int16x8_t __ret_224; \
60876  __ret_224 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_224), __p1_224)); \
60877  __ret_224; \
60878})
60879#else
60880#define vshll_high_n_s8(__p0_225, __p1_225) __extension__ ({ \
60881  int8x16_t __s0_225 = __p0_225; \
60882  int8x16_t __rev0_225;  __rev0_225 = __builtin_shufflevector(__s0_225, __s0_225, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60883  int16x8_t __ret_225; \
60884  __ret_225 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_225), __p1_225)); \
60885  __ret_225 = __builtin_shufflevector(__ret_225, __ret_225, 7, 6, 5, 4, 3, 2, 1, 0); \
60886  __ret_225; \
60887})
60888#endif
60889
60890#ifdef __LITTLE_ENDIAN__
60891#define vshll_high_n_s32(__p0_226, __p1_226) __extension__ ({ \
60892  int32x4_t __s0_226 = __p0_226; \
60893  int64x2_t __ret_226; \
60894  __ret_226 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_226), __p1_226)); \
60895  __ret_226; \
60896})
60897#else
60898#define vshll_high_n_s32(__p0_227, __p1_227) __extension__ ({ \
60899  int32x4_t __s0_227 = __p0_227; \
60900  int32x4_t __rev0_227;  __rev0_227 = __builtin_shufflevector(__s0_227, __s0_227, 3, 2, 1, 0); \
60901  int64x2_t __ret_227; \
60902  __ret_227 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_227), __p1_227)); \
60903  __ret_227 = __builtin_shufflevector(__ret_227, __ret_227, 1, 0); \
60904  __ret_227; \
60905})
60906#endif
60907
60908#ifdef __LITTLE_ENDIAN__
60909#define vshll_high_n_s16(__p0_228, __p1_228) __extension__ ({ \
60910  int16x8_t __s0_228 = __p0_228; \
60911  int32x4_t __ret_228; \
60912  __ret_228 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_228), __p1_228)); \
60913  __ret_228; \
60914})
60915#else
60916#define vshll_high_n_s16(__p0_229, __p1_229) __extension__ ({ \
60917  int16x8_t __s0_229 = __p0_229; \
60918  int16x8_t __rev0_229;  __rev0_229 = __builtin_shufflevector(__s0_229, __s0_229, 7, 6, 5, 4, 3, 2, 1, 0); \
60919  int32x4_t __ret_229; \
60920  __ret_229 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_229), __p1_229)); \
60921  __ret_229 = __builtin_shufflevector(__ret_229, __ret_229, 3, 2, 1, 0); \
60922  __ret_229; \
60923})
60924#endif
60925
60926#ifdef __LITTLE_ENDIAN__
60927#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60928  uint64_t __s0 = __p0; \
60929  uint64_t __ret; \
60930  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60931  __ret; \
60932})
60933#else
60934#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60935  uint64_t __s0 = __p0; \
60936  uint64_t __ret; \
60937  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60938  __ret; \
60939})
60940#endif
60941
60942#ifdef __LITTLE_ENDIAN__
60943#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60944  int64_t __s0 = __p0; \
60945  int64_t __ret; \
60946  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60947  __ret; \
60948})
60949#else
60950#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60951  int64_t __s0 = __p0; \
60952  int64_t __ret; \
60953  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60954  __ret; \
60955})
60956#endif
60957
60958#ifdef __LITTLE_ENDIAN__
60959#define vshrn_high_n_u32(__p0_230, __p1_230, __p2_230) __extension__ ({ \
60960  uint16x4_t __s0_230 = __p0_230; \
60961  uint32x4_t __s1_230 = __p1_230; \
60962  uint16x8_t __ret_230; \
60963  __ret_230 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_230), (uint16x4_t)(vshrn_n_u32(__s1_230, __p2_230)))); \
60964  __ret_230; \
60965})
60966#else
60967#define vshrn_high_n_u32(__p0_231, __p1_231, __p2_231) __extension__ ({ \
60968  uint16x4_t __s0_231 = __p0_231; \
60969  uint32x4_t __s1_231 = __p1_231; \
60970  uint16x4_t __rev0_231;  __rev0_231 = __builtin_shufflevector(__s0_231, __s0_231, 3, 2, 1, 0); \
60971  uint32x4_t __rev1_231;  __rev1_231 = __builtin_shufflevector(__s1_231, __s1_231, 3, 2, 1, 0); \
60972  uint16x8_t __ret_231; \
60973  __ret_231 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_231), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_231, __p2_231)))); \
60974  __ret_231 = __builtin_shufflevector(__ret_231, __ret_231, 7, 6, 5, 4, 3, 2, 1, 0); \
60975  __ret_231; \
60976})
60977#endif
60978
60979#ifdef __LITTLE_ENDIAN__
60980#define vshrn_high_n_u64(__p0_232, __p1_232, __p2_232) __extension__ ({ \
60981  uint32x2_t __s0_232 = __p0_232; \
60982  uint64x2_t __s1_232 = __p1_232; \
60983  uint32x4_t __ret_232; \
60984  __ret_232 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_232), (uint32x2_t)(vshrn_n_u64(__s1_232, __p2_232)))); \
60985  __ret_232; \
60986})
60987#else
60988#define vshrn_high_n_u64(__p0_233, __p1_233, __p2_233) __extension__ ({ \
60989  uint32x2_t __s0_233 = __p0_233; \
60990  uint64x2_t __s1_233 = __p1_233; \
60991  uint32x2_t __rev0_233;  __rev0_233 = __builtin_shufflevector(__s0_233, __s0_233, 1, 0); \
60992  uint64x2_t __rev1_233;  __rev1_233 = __builtin_shufflevector(__s1_233, __s1_233, 1, 0); \
60993  uint32x4_t __ret_233; \
60994  __ret_233 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_233), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_233, __p2_233)))); \
60995  __ret_233 = __builtin_shufflevector(__ret_233, __ret_233, 3, 2, 1, 0); \
60996  __ret_233; \
60997})
60998#endif
60999
61000#ifdef __LITTLE_ENDIAN__
61001#define vshrn_high_n_u16(__p0_234, __p1_234, __p2_234) __extension__ ({ \
61002  uint8x8_t __s0_234 = __p0_234; \
61003  uint16x8_t __s1_234 = __p1_234; \
61004  uint8x16_t __ret_234; \
61005  __ret_234 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_234), (uint8x8_t)(vshrn_n_u16(__s1_234, __p2_234)))); \
61006  __ret_234; \
61007})
61008#else
61009#define vshrn_high_n_u16(__p0_235, __p1_235, __p2_235) __extension__ ({ \
61010  uint8x8_t __s0_235 = __p0_235; \
61011  uint16x8_t __s1_235 = __p1_235; \
61012  uint8x8_t __rev0_235;  __rev0_235 = __builtin_shufflevector(__s0_235, __s0_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61013  uint16x8_t __rev1_235;  __rev1_235 = __builtin_shufflevector(__s1_235, __s1_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61014  uint8x16_t __ret_235; \
61015  __ret_235 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_235), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_235, __p2_235)))); \
61016  __ret_235 = __builtin_shufflevector(__ret_235, __ret_235, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61017  __ret_235; \
61018})
61019#endif
61020
61021#ifdef __LITTLE_ENDIAN__
61022#define vshrn_high_n_s32(__p0_236, __p1_236, __p2_236) __extension__ ({ \
61023  int16x4_t __s0_236 = __p0_236; \
61024  int32x4_t __s1_236 = __p1_236; \
61025  int16x8_t __ret_236; \
61026  __ret_236 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_236), (int16x4_t)(vshrn_n_s32(__s1_236, __p2_236)))); \
61027  __ret_236; \
61028})
61029#else
61030#define vshrn_high_n_s32(__p0_237, __p1_237, __p2_237) __extension__ ({ \
61031  int16x4_t __s0_237 = __p0_237; \
61032  int32x4_t __s1_237 = __p1_237; \
61033  int16x4_t __rev0_237;  __rev0_237 = __builtin_shufflevector(__s0_237, __s0_237, 3, 2, 1, 0); \
61034  int32x4_t __rev1_237;  __rev1_237 = __builtin_shufflevector(__s1_237, __s1_237, 3, 2, 1, 0); \
61035  int16x8_t __ret_237; \
61036  __ret_237 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_237), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_237, __p2_237)))); \
61037  __ret_237 = __builtin_shufflevector(__ret_237, __ret_237, 7, 6, 5, 4, 3, 2, 1, 0); \
61038  __ret_237; \
61039})
61040#endif
61041
61042#ifdef __LITTLE_ENDIAN__
61043#define vshrn_high_n_s64(__p0_238, __p1_238, __p2_238) __extension__ ({ \
61044  int32x2_t __s0_238 = __p0_238; \
61045  int64x2_t __s1_238 = __p1_238; \
61046  int32x4_t __ret_238; \
61047  __ret_238 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_238), (int32x2_t)(vshrn_n_s64(__s1_238, __p2_238)))); \
61048  __ret_238; \
61049})
61050#else
61051#define vshrn_high_n_s64(__p0_239, __p1_239, __p2_239) __extension__ ({ \
61052  int32x2_t __s0_239 = __p0_239; \
61053  int64x2_t __s1_239 = __p1_239; \
61054  int32x2_t __rev0_239;  __rev0_239 = __builtin_shufflevector(__s0_239, __s0_239, 1, 0); \
61055  int64x2_t __rev1_239;  __rev1_239 = __builtin_shufflevector(__s1_239, __s1_239, 1, 0); \
61056  int32x4_t __ret_239; \
61057  __ret_239 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_239), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_239, __p2_239)))); \
61058  __ret_239 = __builtin_shufflevector(__ret_239, __ret_239, 3, 2, 1, 0); \
61059  __ret_239; \
61060})
61061#endif
61062
61063#ifdef __LITTLE_ENDIAN__
61064#define vshrn_high_n_s16(__p0_240, __p1_240, __p2_240) __extension__ ({ \
61065  int8x8_t __s0_240 = __p0_240; \
61066  int16x8_t __s1_240 = __p1_240; \
61067  int8x16_t __ret_240; \
61068  __ret_240 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_240), (int8x8_t)(vshrn_n_s16(__s1_240, __p2_240)))); \
61069  __ret_240; \
61070})
61071#else
61072#define vshrn_high_n_s16(__p0_241, __p1_241, __p2_241) __extension__ ({ \
61073  int8x8_t __s0_241 = __p0_241; \
61074  int16x8_t __s1_241 = __p1_241; \
61075  int8x8_t __rev0_241;  __rev0_241 = __builtin_shufflevector(__s0_241, __s0_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61076  int16x8_t __rev1_241;  __rev1_241 = __builtin_shufflevector(__s1_241, __s1_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61077  int8x16_t __ret_241; \
61078  __ret_241 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_241), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_241, __p2_241)))); \
61079  __ret_241 = __builtin_shufflevector(__ret_241, __ret_241, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61080  __ret_241; \
61081})
61082#endif
61083
61084#ifdef __LITTLE_ENDIAN__
61085#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61086  uint64_t __s0 = __p0; \
61087  uint64_t __s1 = __p1; \
61088  uint64_t __ret; \
61089  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61090  __ret; \
61091})
61092#else
61093#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61094  uint64_t __s0 = __p0; \
61095  uint64_t __s1 = __p1; \
61096  uint64_t __ret; \
61097  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61098  __ret; \
61099})
61100#endif
61101
61102#ifdef __LITTLE_ENDIAN__
61103#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61104  int64_t __s0 = __p0; \
61105  int64_t __s1 = __p1; \
61106  int64_t __ret; \
61107  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61108  __ret; \
61109})
61110#else
61111#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61112  int64_t __s0 = __p0; \
61113  int64_t __s1 = __p1; \
61114  int64_t __ret; \
61115  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61116  __ret; \
61117})
61118#endif
61119
61120#ifdef __LITTLE_ENDIAN__
61121#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61122  poly64x1_t __s0 = __p0; \
61123  poly64x1_t __s1 = __p1; \
61124  poly64x1_t __ret; \
61125  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61126  __ret; \
61127})
61128#else
61129#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61130  poly64x1_t __s0 = __p0; \
61131  poly64x1_t __s1 = __p1; \
61132  poly64x1_t __ret; \
61133  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61134  __ret; \
61135})
61136#endif
61137
61138#ifdef __LITTLE_ENDIAN__
61139#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61140  poly64x2_t __s0 = __p0; \
61141  poly64x2_t __s1 = __p1; \
61142  poly64x2_t __ret; \
61143  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61144  __ret; \
61145})
61146#else
61147#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61148  poly64x2_t __s0 = __p0; \
61149  poly64x2_t __s1 = __p1; \
61150  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61151  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61152  poly64x2_t __ret; \
61153  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61154  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61155  __ret; \
61156})
61157#endif
61158
61159#ifdef __LITTLE_ENDIAN__
61160__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61161  uint8_t __ret;
61162  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61163  return __ret;
61164}
61165#else
61166__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61167  uint8_t __ret;
61168  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61169  return __ret;
61170}
61171#endif
61172
61173#ifdef __LITTLE_ENDIAN__
61174__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61175  uint32_t __ret;
61176  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61177  return __ret;
61178}
61179#else
61180__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61181  uint32_t __ret;
61182  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61183  return __ret;
61184}
61185#endif
61186
61187#ifdef __LITTLE_ENDIAN__
61188__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61189  uint64_t __ret;
61190  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61191  return __ret;
61192}
61193#else
61194__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61195  uint64_t __ret;
61196  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61197  return __ret;
61198}
61199#endif
61200
61201#ifdef __LITTLE_ENDIAN__
61202__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61203  uint16_t __ret;
61204  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61205  return __ret;
61206}
61207#else
61208__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61209  uint16_t __ret;
61210  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61211  return __ret;
61212}
61213#endif
61214
61215#ifdef __LITTLE_ENDIAN__
61216__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61217  uint8x16_t __ret;
61218  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
61219  return __ret;
61220}
61221#else
61222__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61223  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61224  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61225  uint8x16_t __ret;
61226  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
61227  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61228  return __ret;
61229}
61230#endif
61231
61232#ifdef __LITTLE_ENDIAN__
61233__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61234  uint32x4_t __ret;
61235  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
61236  return __ret;
61237}
61238#else
61239__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61240  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61241  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61242  uint32x4_t __ret;
61243  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
61244  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61245  return __ret;
61246}
61247#endif
61248
61249#ifdef __LITTLE_ENDIAN__
61250__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61251  uint64x2_t __ret;
61252  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
61253  return __ret;
61254}
61255#else
61256__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61257  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61258  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61259  uint64x2_t __ret;
61260  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
61261  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61262  return __ret;
61263}
61264#endif
61265
61266#ifdef __LITTLE_ENDIAN__
61267__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61268  uint16x8_t __ret;
61269  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
61270  return __ret;
61271}
61272#else
61273__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61274  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61275  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61276  uint16x8_t __ret;
61277  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
61278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61279  return __ret;
61280}
61281#endif
61282
61283#ifdef __LITTLE_ENDIAN__
61284__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61285  uint8x8_t __ret;
61286  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
61287  return __ret;
61288}
61289#else
61290__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61291  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61292  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61293  uint8x8_t __ret;
61294  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
61295  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61296  return __ret;
61297}
61298#endif
61299
61300#ifdef __LITTLE_ENDIAN__
61301__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61302  uint32x2_t __ret;
61303  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
61304  return __ret;
61305}
61306#else
61307__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61308  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61309  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61310  uint32x2_t __ret;
61311  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
61312  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61313  return __ret;
61314}
61315#endif
61316
61317#ifdef __LITTLE_ENDIAN__
61318__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61319  uint64x1_t __ret;
61320  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61321  return __ret;
61322}
61323#else
61324__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61325  uint64x1_t __ret;
61326  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61327  return __ret;
61328}
61329#endif
61330
61331#ifdef __LITTLE_ENDIAN__
61332__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61333  uint16x4_t __ret;
61334  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
61335  return __ret;
61336}
61337#else
61338__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61339  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61340  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61341  uint16x4_t __ret;
61342  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
61343  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61344  return __ret;
61345}
61346#endif
61347
61348#ifdef __LITTLE_ENDIAN__
61349__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61350  float64x2_t __ret;
61351  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
61352  return __ret;
61353}
61354#else
61355__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61356  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61357  float64x2_t __ret;
61358  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
61359  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61360  return __ret;
61361}
61362#endif
61363
61364#ifdef __LITTLE_ENDIAN__
61365__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61366  float32x4_t __ret;
61367  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
61368  return __ret;
61369}
61370#else
61371__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61372  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61373  float32x4_t __ret;
61374  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
61375  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61376  return __ret;
61377}
61378#endif
61379
61380#ifdef __LITTLE_ENDIAN__
61381__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61382  float64x1_t __ret;
61383  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61384  return __ret;
61385}
61386#else
61387__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61388  float64x1_t __ret;
61389  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61390  return __ret;
61391}
61392#endif
61393
61394#ifdef __LITTLE_ENDIAN__
61395__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61396  float32x2_t __ret;
61397  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
61398  return __ret;
61399}
61400#else
61401__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61402  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61403  float32x2_t __ret;
61404  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
61405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61406  return __ret;
61407}
61408#endif
61409
61410#ifdef __LITTLE_ENDIAN__
61411#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61412  uint64_t __s0 = __p0; \
61413  uint64_t __s1 = __p1; \
61414  uint64_t __ret; \
61415  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61416  __ret; \
61417})
61418#else
61419#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61420  uint64_t __s0 = __p0; \
61421  uint64_t __s1 = __p1; \
61422  uint64_t __ret; \
61423  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61424  __ret; \
61425})
61426#endif
61427
61428#ifdef __LITTLE_ENDIAN__
61429#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61430  int64_t __s0 = __p0; \
61431  int64_t __s1 = __p1; \
61432  int64_t __ret; \
61433  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61434  __ret; \
61435})
61436#else
61437#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61438  int64_t __s0 = __p0; \
61439  int64_t __s1 = __p1; \
61440  int64_t __ret; \
61441  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61442  __ret; \
61443})
61444#endif
61445
61446#ifdef __LITTLE_ENDIAN__
61447#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61448  uint64_t __s0 = __p0; \
61449  uint64_t __s1 = __p1; \
61450  uint64_t __ret; \
61451  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61452  __ret; \
61453})
61454#else
61455#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61456  uint64_t __s0 = __p0; \
61457  uint64_t __s1 = __p1; \
61458  uint64_t __ret; \
61459  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61460  __ret; \
61461})
61462#endif
61463
61464#ifdef __LITTLE_ENDIAN__
61465#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61466  int64_t __s0 = __p0; \
61467  int64_t __s1 = __p1; \
61468  int64_t __ret; \
61469  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61470  __ret; \
61471})
61472#else
61473#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61474  int64_t __s0 = __p0; \
61475  int64_t __s1 = __p1; \
61476  int64_t __ret; \
61477  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61478  __ret; \
61479})
61480#endif
61481
61482#ifdef __LITTLE_ENDIAN__
61483#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61484  poly64x1_t __s0 = __p0; \
61485  poly64x1_t __s1 = __p1; \
61486  poly64x1_t __ret; \
61487  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61488  __ret; \
61489})
61490#else
61491#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61492  poly64x1_t __s0 = __p0; \
61493  poly64x1_t __s1 = __p1; \
61494  poly64x1_t __ret; \
61495  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61496  __ret; \
61497})
61498#endif
61499
61500#ifdef __LITTLE_ENDIAN__
61501#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61502  poly64x2_t __s0 = __p0; \
61503  poly64x2_t __s1 = __p1; \
61504  poly64x2_t __ret; \
61505  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61506  __ret; \
61507})
61508#else
61509#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61510  poly64x2_t __s0 = __p0; \
61511  poly64x2_t __s1 = __p1; \
61512  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61513  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61514  poly64x2_t __ret; \
61515  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61517  __ret; \
61518})
61519#endif
61520
61521#ifdef __LITTLE_ENDIAN__
61522#define vst1_p64(__p0, __p1) __extension__ ({ \
61523  poly64x1_t __s1 = __p1; \
61524  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61525})
61526#else
61527#define vst1_p64(__p0, __p1) __extension__ ({ \
61528  poly64x1_t __s1 = __p1; \
61529  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61530})
61531#endif
61532
61533#ifdef __LITTLE_ENDIAN__
61534#define vst1q_p64(__p0, __p1) __extension__ ({ \
61535  poly64x2_t __s1 = __p1; \
61536  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
61537})
61538#else
61539#define vst1q_p64(__p0, __p1) __extension__ ({ \
61540  poly64x2_t __s1 = __p1; \
61541  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61542  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
61543})
61544#endif
61545
61546#ifdef __LITTLE_ENDIAN__
61547#define vst1q_f64(__p0, __p1) __extension__ ({ \
61548  float64x2_t __s1 = __p1; \
61549  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
61550})
61551#else
61552#define vst1q_f64(__p0, __p1) __extension__ ({ \
61553  float64x2_t __s1 = __p1; \
61554  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61555  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
61556})
61557#endif
61558
61559#ifdef __LITTLE_ENDIAN__
61560#define vst1_f64(__p0, __p1) __extension__ ({ \
61561  float64x1_t __s1 = __p1; \
61562  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61563})
61564#else
61565#define vst1_f64(__p0, __p1) __extension__ ({ \
61566  float64x1_t __s1 = __p1; \
61567  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61568})
61569#endif
61570
61571#ifdef __LITTLE_ENDIAN__
61572#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61573  poly64x1_t __s1 = __p1; \
61574  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61575})
61576#else
61577#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61578  poly64x1_t __s1 = __p1; \
61579  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61580})
61581#endif
61582
61583#ifdef __LITTLE_ENDIAN__
61584#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61585  poly64x2_t __s1 = __p1; \
61586  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
61587})
61588#else
61589#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61590  poly64x2_t __s1 = __p1; \
61591  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61592  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
61593})
61594#endif
61595
61596#ifdef __LITTLE_ENDIAN__
61597#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61598  float64x2_t __s1 = __p1; \
61599  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
61600})
61601#else
61602#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61603  float64x2_t __s1 = __p1; \
61604  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61605  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
61606})
61607#endif
61608
61609#ifdef __LITTLE_ENDIAN__
61610#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61611  float64x1_t __s1 = __p1; \
61612  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61613})
61614#else
61615#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61616  float64x1_t __s1 = __p1; \
61617  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61618})
61619#endif
61620
61621#ifdef __LITTLE_ENDIAN__
61622#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61623  poly8x8x2_t __s1 = __p1; \
61624  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
61625})
61626#else
61627#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61628  poly8x8x2_t __s1 = __p1; \
61629  poly8x8x2_t __rev1; \
61630  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61631  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61632  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
61633})
61634#endif
61635
61636#ifdef __LITTLE_ENDIAN__
61637#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61638  poly64x1x2_t __s1 = __p1; \
61639  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61640})
61641#else
61642#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61643  poly64x1x2_t __s1 = __p1; \
61644  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61645})
61646#endif
61647
61648#ifdef __LITTLE_ENDIAN__
61649#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61650  poly16x4x2_t __s1 = __p1; \
61651  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
61652})
61653#else
61654#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61655  poly16x4x2_t __s1 = __p1; \
61656  poly16x4x2_t __rev1; \
61657  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61658  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61659  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
61660})
61661#endif
61662
61663#ifdef __LITTLE_ENDIAN__
61664#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61665  poly8x16x2_t __s1 = __p1; \
61666  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
61667})
61668#else
61669#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61670  poly8x16x2_t __s1 = __p1; \
61671  poly8x16x2_t __rev1; \
61672  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61673  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61674  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
61675})
61676#endif
61677
61678#ifdef __LITTLE_ENDIAN__
61679#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61680  poly64x2x2_t __s1 = __p1; \
61681  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
61682})
61683#else
61684#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61685  poly64x2x2_t __s1 = __p1; \
61686  poly64x2x2_t __rev1; \
61687  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61688  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61689  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
61690})
61691#endif
61692
61693#ifdef __LITTLE_ENDIAN__
61694#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61695  poly16x8x2_t __s1 = __p1; \
61696  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
61697})
61698#else
61699#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61700  poly16x8x2_t __s1 = __p1; \
61701  poly16x8x2_t __rev1; \
61702  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61703  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61704  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
61705})
61706#endif
61707
61708#ifdef __LITTLE_ENDIAN__
61709#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61710  uint8x16x2_t __s1 = __p1; \
61711  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
61712})
61713#else
61714#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61715  uint8x16x2_t __s1 = __p1; \
61716  uint8x16x2_t __rev1; \
61717  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61718  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61719  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
61720})
61721#endif
61722
61723#ifdef __LITTLE_ENDIAN__
61724#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61725  uint32x4x2_t __s1 = __p1; \
61726  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
61727})
61728#else
61729#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61730  uint32x4x2_t __s1 = __p1; \
61731  uint32x4x2_t __rev1; \
61732  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61733  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61734  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
61735})
61736#endif
61737
61738#ifdef __LITTLE_ENDIAN__
61739#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61740  uint64x2x2_t __s1 = __p1; \
61741  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
61742})
61743#else
61744#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61745  uint64x2x2_t __s1 = __p1; \
61746  uint64x2x2_t __rev1; \
61747  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61748  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61749  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
61750})
61751#endif
61752
61753#ifdef __LITTLE_ENDIAN__
61754#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61755  uint16x8x2_t __s1 = __p1; \
61756  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
61757})
61758#else
61759#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61760  uint16x8x2_t __s1 = __p1; \
61761  uint16x8x2_t __rev1; \
61762  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61763  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61764  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
61765})
61766#endif
61767
61768#ifdef __LITTLE_ENDIAN__
61769#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61770  int8x16x2_t __s1 = __p1; \
61771  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
61772})
61773#else
61774#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61775  int8x16x2_t __s1 = __p1; \
61776  int8x16x2_t __rev1; \
61777  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61778  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61779  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
61780})
61781#endif
61782
61783#ifdef __LITTLE_ENDIAN__
61784#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61785  float64x2x2_t __s1 = __p1; \
61786  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 42); \
61787})
61788#else
61789#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61790  float64x2x2_t __s1 = __p1; \
61791  float64x2x2_t __rev1; \
61792  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61793  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61794  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
61795})
61796#endif
61797
61798#ifdef __LITTLE_ENDIAN__
61799#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61800  float32x4x2_t __s1 = __p1; \
61801  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 41); \
61802})
61803#else
61804#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61805  float32x4x2_t __s1 = __p1; \
61806  float32x4x2_t __rev1; \
61807  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61808  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61809  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
61810})
61811#endif
61812
61813#ifdef __LITTLE_ENDIAN__
61814#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61815  float16x8x2_t __s1 = __p1; \
61816  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 40); \
61817})
61818#else
61819#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61820  float16x8x2_t __s1 = __p1; \
61821  float16x8x2_t __rev1; \
61822  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61823  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61824  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
61825})
61826#endif
61827
61828#ifdef __LITTLE_ENDIAN__
61829#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61830  int32x4x2_t __s1 = __p1; \
61831  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 34); \
61832})
61833#else
61834#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61835  int32x4x2_t __s1 = __p1; \
61836  int32x4x2_t __rev1; \
61837  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61838  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61839  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
61840})
61841#endif
61842
61843#ifdef __LITTLE_ENDIAN__
61844#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61845  int64x2x2_t __s1 = __p1; \
61846  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 35); \
61847})
61848#else
61849#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61850  int64x2x2_t __s1 = __p1; \
61851  int64x2x2_t __rev1; \
61852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61854  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
61855})
61856#endif
61857
61858#ifdef __LITTLE_ENDIAN__
61859#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61860  int16x8x2_t __s1 = __p1; \
61861  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 33); \
61862})
61863#else
61864#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61865  int16x8x2_t __s1 = __p1; \
61866  int16x8x2_t __rev1; \
61867  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61868  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61869  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
61870})
61871#endif
61872
61873#ifdef __LITTLE_ENDIAN__
61874#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61875  uint8x8x2_t __s1 = __p1; \
61876  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
61877})
61878#else
61879#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61880  uint8x8x2_t __s1 = __p1; \
61881  uint8x8x2_t __rev1; \
61882  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61883  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61884  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
61885})
61886#endif
61887
61888#ifdef __LITTLE_ENDIAN__
61889#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61890  uint32x2x2_t __s1 = __p1; \
61891  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
61892})
61893#else
61894#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61895  uint32x2x2_t __s1 = __p1; \
61896  uint32x2x2_t __rev1; \
61897  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61898  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61899  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
61900})
61901#endif
61902
61903#ifdef __LITTLE_ENDIAN__
61904#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61905  uint64x1x2_t __s1 = __p1; \
61906  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61907})
61908#else
61909#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61910  uint64x1x2_t __s1 = __p1; \
61911  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61912})
61913#endif
61914
61915#ifdef __LITTLE_ENDIAN__
61916#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61917  uint16x4x2_t __s1 = __p1; \
61918  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
61919})
61920#else
61921#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61922  uint16x4x2_t __s1 = __p1; \
61923  uint16x4x2_t __rev1; \
61924  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61925  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61926  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
61927})
61928#endif
61929
61930#ifdef __LITTLE_ENDIAN__
61931#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61932  int8x8x2_t __s1 = __p1; \
61933  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
61934})
61935#else
61936#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61937  int8x8x2_t __s1 = __p1; \
61938  int8x8x2_t __rev1; \
61939  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61940  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61941  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
61942})
61943#endif
61944
61945#ifdef __LITTLE_ENDIAN__
61946#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61947  float64x1x2_t __s1 = __p1; \
61948  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61949})
61950#else
61951#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61952  float64x1x2_t __s1 = __p1; \
61953  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61954})
61955#endif
61956
61957#ifdef __LITTLE_ENDIAN__
61958#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61959  float32x2x2_t __s1 = __p1; \
61960  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 9); \
61961})
61962#else
61963#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61964  float32x2x2_t __s1 = __p1; \
61965  float32x2x2_t __rev1; \
61966  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61967  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61968  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
61969})
61970#endif
61971
61972#ifdef __LITTLE_ENDIAN__
61973#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61974  float16x4x2_t __s1 = __p1; \
61975  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 8); \
61976})
61977#else
61978#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61979  float16x4x2_t __s1 = __p1; \
61980  float16x4x2_t __rev1; \
61981  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61982  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61983  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
61984})
61985#endif
61986
61987#ifdef __LITTLE_ENDIAN__
61988#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61989  int32x2x2_t __s1 = __p1; \
61990  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 2); \
61991})
61992#else
61993#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61994  int32x2x2_t __s1 = __p1; \
61995  int32x2x2_t __rev1; \
61996  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61997  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61998  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
61999})
62000#endif
62001
62002#ifdef __LITTLE_ENDIAN__
62003#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62004  int64x1x2_t __s1 = __p1; \
62005  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62006})
62007#else
62008#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62009  int64x1x2_t __s1 = __p1; \
62010  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62011})
62012#endif
62013
62014#ifdef __LITTLE_ENDIAN__
62015#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62016  int16x4x2_t __s1 = __p1; \
62017  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 1); \
62018})
62019#else
62020#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62021  int16x4x2_t __s1 = __p1; \
62022  int16x4x2_t __rev1; \
62023  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62024  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62025  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
62026})
62027#endif
62028
62029#ifdef __LITTLE_ENDIAN__
62030#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62031  poly8x8x3_t __s1 = __p1; \
62032  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
62033})
62034#else
62035#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62036  poly8x8x3_t __s1 = __p1; \
62037  poly8x8x3_t __rev1; \
62038  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62039  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62040  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62041  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
62042})
62043#endif
62044
62045#ifdef __LITTLE_ENDIAN__
62046#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62047  poly64x1x3_t __s1 = __p1; \
62048  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62049})
62050#else
62051#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62052  poly64x1x3_t __s1 = __p1; \
62053  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62054})
62055#endif
62056
62057#ifdef __LITTLE_ENDIAN__
62058#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62059  poly16x4x3_t __s1 = __p1; \
62060  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
62061})
62062#else
62063#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62064  poly16x4x3_t __s1 = __p1; \
62065  poly16x4x3_t __rev1; \
62066  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62067  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62068  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62069  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
62070})
62071#endif
62072
62073#ifdef __LITTLE_ENDIAN__
62074#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62075  poly8x16x3_t __s1 = __p1; \
62076  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
62077})
62078#else
62079#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62080  poly8x16x3_t __s1 = __p1; \
62081  poly8x16x3_t __rev1; \
62082  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62083  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62084  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62085  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
62086})
62087#endif
62088
62089#ifdef __LITTLE_ENDIAN__
62090#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62091  poly64x2x3_t __s1 = __p1; \
62092  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
62093})
62094#else
62095#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62096  poly64x2x3_t __s1 = __p1; \
62097  poly64x2x3_t __rev1; \
62098  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62099  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62100  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62101  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
62102})
62103#endif
62104
62105#ifdef __LITTLE_ENDIAN__
62106#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62107  poly16x8x3_t __s1 = __p1; \
62108  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
62109})
62110#else
62111#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62112  poly16x8x3_t __s1 = __p1; \
62113  poly16x8x3_t __rev1; \
62114  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62115  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62116  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62117  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
62118})
62119#endif
62120
62121#ifdef __LITTLE_ENDIAN__
62122#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62123  uint8x16x3_t __s1 = __p1; \
62124  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
62125})
62126#else
62127#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62128  uint8x16x3_t __s1 = __p1; \
62129  uint8x16x3_t __rev1; \
62130  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62131  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62132  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62133  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
62134})
62135#endif
62136
62137#ifdef __LITTLE_ENDIAN__
62138#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62139  uint32x4x3_t __s1 = __p1; \
62140  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
62141})
62142#else
62143#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62144  uint32x4x3_t __s1 = __p1; \
62145  uint32x4x3_t __rev1; \
62146  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62147  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62148  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62149  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
62150})
62151#endif
62152
62153#ifdef __LITTLE_ENDIAN__
62154#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62155  uint64x2x3_t __s1 = __p1; \
62156  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
62157})
62158#else
62159#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62160  uint64x2x3_t __s1 = __p1; \
62161  uint64x2x3_t __rev1; \
62162  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62163  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62164  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62165  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
62166})
62167#endif
62168
62169#ifdef __LITTLE_ENDIAN__
62170#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62171  uint16x8x3_t __s1 = __p1; \
62172  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
62173})
62174#else
62175#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62176  uint16x8x3_t __s1 = __p1; \
62177  uint16x8x3_t __rev1; \
62178  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62179  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62180  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62181  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
62182})
62183#endif
62184
62185#ifdef __LITTLE_ENDIAN__
62186#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62187  int8x16x3_t __s1 = __p1; \
62188  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
62189})
62190#else
62191#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62192  int8x16x3_t __s1 = __p1; \
62193  int8x16x3_t __rev1; \
62194  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62195  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62196  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62197  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
62198})
62199#endif
62200
62201#ifdef __LITTLE_ENDIAN__
62202#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62203  float64x2x3_t __s1 = __p1; \
62204  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
62205})
62206#else
62207#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62208  float64x2x3_t __s1 = __p1; \
62209  float64x2x3_t __rev1; \
62210  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62211  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62212  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62213  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
62214})
62215#endif
62216
62217#ifdef __LITTLE_ENDIAN__
62218#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62219  float32x4x3_t __s1 = __p1; \
62220  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
62221})
62222#else
62223#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62224  float32x4x3_t __s1 = __p1; \
62225  float32x4x3_t __rev1; \
62226  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62227  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62228  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62229  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
62230})
62231#endif
62232
62233#ifdef __LITTLE_ENDIAN__
62234#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62235  float16x8x3_t __s1 = __p1; \
62236  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
62237})
62238#else
62239#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62240  float16x8x3_t __s1 = __p1; \
62241  float16x8x3_t __rev1; \
62242  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62243  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62244  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62245  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
62246})
62247#endif
62248
62249#ifdef __LITTLE_ENDIAN__
62250#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62251  int32x4x3_t __s1 = __p1; \
62252  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
62253})
62254#else
62255#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62256  int32x4x3_t __s1 = __p1; \
62257  int32x4x3_t __rev1; \
62258  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62259  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62260  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62261  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
62262})
62263#endif
62264
62265#ifdef __LITTLE_ENDIAN__
62266#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62267  int64x2x3_t __s1 = __p1; \
62268  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
62269})
62270#else
62271#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62272  int64x2x3_t __s1 = __p1; \
62273  int64x2x3_t __rev1; \
62274  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62275  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62276  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62277  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
62278})
62279#endif
62280
62281#ifdef __LITTLE_ENDIAN__
62282#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62283  int16x8x3_t __s1 = __p1; \
62284  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
62285})
62286#else
62287#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62288  int16x8x3_t __s1 = __p1; \
62289  int16x8x3_t __rev1; \
62290  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62291  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62292  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62293  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
62294})
62295#endif
62296
62297#ifdef __LITTLE_ENDIAN__
62298#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62299  uint8x8x3_t __s1 = __p1; \
62300  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
62301})
62302#else
62303#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62304  uint8x8x3_t __s1 = __p1; \
62305  uint8x8x3_t __rev1; \
62306  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62307  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62308  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62309  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
62310})
62311#endif
62312
62313#ifdef __LITTLE_ENDIAN__
62314#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62315  uint32x2x3_t __s1 = __p1; \
62316  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
62317})
62318#else
62319#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62320  uint32x2x3_t __s1 = __p1; \
62321  uint32x2x3_t __rev1; \
62322  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62323  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62324  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62325  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
62326})
62327#endif
62328
62329#ifdef __LITTLE_ENDIAN__
62330#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62331  uint64x1x3_t __s1 = __p1; \
62332  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62333})
62334#else
62335#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62336  uint64x1x3_t __s1 = __p1; \
62337  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62338})
62339#endif
62340
62341#ifdef __LITTLE_ENDIAN__
62342#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62343  uint16x4x3_t __s1 = __p1; \
62344  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
62345})
62346#else
62347#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62348  uint16x4x3_t __s1 = __p1; \
62349  uint16x4x3_t __rev1; \
62350  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62351  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62352  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62353  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
62354})
62355#endif
62356
62357#ifdef __LITTLE_ENDIAN__
62358#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62359  int8x8x3_t __s1 = __p1; \
62360  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
62361})
62362#else
62363#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62364  int8x8x3_t __s1 = __p1; \
62365  int8x8x3_t __rev1; \
62366  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62367  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62368  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62369  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
62370})
62371#endif
62372
62373#ifdef __LITTLE_ENDIAN__
62374#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62375  float64x1x3_t __s1 = __p1; \
62376  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62377})
62378#else
62379#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62380  float64x1x3_t __s1 = __p1; \
62381  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62382})
62383#endif
62384
62385#ifdef __LITTLE_ENDIAN__
62386#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62387  float32x2x3_t __s1 = __p1; \
62388  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
62389})
62390#else
62391#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62392  float32x2x3_t __s1 = __p1; \
62393  float32x2x3_t __rev1; \
62394  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62395  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62396  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62397  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
62398})
62399#endif
62400
62401#ifdef __LITTLE_ENDIAN__
62402#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62403  float16x4x3_t __s1 = __p1; \
62404  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
62405})
62406#else
62407#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62408  float16x4x3_t __s1 = __p1; \
62409  float16x4x3_t __rev1; \
62410  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62411  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62412  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62413  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
62414})
62415#endif
62416
62417#ifdef __LITTLE_ENDIAN__
62418#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62419  int32x2x3_t __s1 = __p1; \
62420  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
62421})
62422#else
62423#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62424  int32x2x3_t __s1 = __p1; \
62425  int32x2x3_t __rev1; \
62426  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62427  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62428  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62429  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
62430})
62431#endif
62432
62433#ifdef __LITTLE_ENDIAN__
62434#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62435  int64x1x3_t __s1 = __p1; \
62436  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62437})
62438#else
62439#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62440  int64x1x3_t __s1 = __p1; \
62441  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62442})
62443#endif
62444
62445#ifdef __LITTLE_ENDIAN__
62446#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62447  int16x4x3_t __s1 = __p1; \
62448  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
62449})
62450#else
62451#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62452  int16x4x3_t __s1 = __p1; \
62453  int16x4x3_t __rev1; \
62454  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62455  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62456  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62457  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
62458})
62459#endif
62460
62461#ifdef __LITTLE_ENDIAN__
62462#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62463  poly8x8x4_t __s1 = __p1; \
62464  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
62465})
62466#else
62467#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62468  poly8x8x4_t __s1 = __p1; \
62469  poly8x8x4_t __rev1; \
62470  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62471  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62472  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62473  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62474  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
62475})
62476#endif
62477
62478#ifdef __LITTLE_ENDIAN__
62479#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62480  poly64x1x4_t __s1 = __p1; \
62481  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62482})
62483#else
62484#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62485  poly64x1x4_t __s1 = __p1; \
62486  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62487})
62488#endif
62489
62490#ifdef __LITTLE_ENDIAN__
62491#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62492  poly16x4x4_t __s1 = __p1; \
62493  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
62494})
62495#else
62496#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62497  poly16x4x4_t __s1 = __p1; \
62498  poly16x4x4_t __rev1; \
62499  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62500  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62501  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62502  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62503  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
62504})
62505#endif
62506
62507#ifdef __LITTLE_ENDIAN__
62508#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62509  poly8x16x4_t __s1 = __p1; \
62510  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
62511})
62512#else
62513#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62514  poly8x16x4_t __s1 = __p1; \
62515  poly8x16x4_t __rev1; \
62516  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62517  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62518  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62519  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62520  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
62521})
62522#endif
62523
62524#ifdef __LITTLE_ENDIAN__
62525#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62526  poly64x2x4_t __s1 = __p1; \
62527  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
62528})
62529#else
62530#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62531  poly64x2x4_t __s1 = __p1; \
62532  poly64x2x4_t __rev1; \
62533  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62534  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62535  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62536  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62537  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
62538})
62539#endif
62540
62541#ifdef __LITTLE_ENDIAN__
62542#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62543  poly16x8x4_t __s1 = __p1; \
62544  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
62545})
62546#else
62547#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62548  poly16x8x4_t __s1 = __p1; \
62549  poly16x8x4_t __rev1; \
62550  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62551  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62552  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62553  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62554  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
62555})
62556#endif
62557
62558#ifdef __LITTLE_ENDIAN__
62559#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62560  uint8x16x4_t __s1 = __p1; \
62561  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
62562})
62563#else
62564#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62565  uint8x16x4_t __s1 = __p1; \
62566  uint8x16x4_t __rev1; \
62567  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62568  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62569  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62570  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62571  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
62572})
62573#endif
62574
62575#ifdef __LITTLE_ENDIAN__
62576#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62577  uint32x4x4_t __s1 = __p1; \
62578  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
62579})
62580#else
62581#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62582  uint32x4x4_t __s1 = __p1; \
62583  uint32x4x4_t __rev1; \
62584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62586  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62587  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62588  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
62589})
62590#endif
62591
62592#ifdef __LITTLE_ENDIAN__
62593#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62594  uint64x2x4_t __s1 = __p1; \
62595  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
62596})
62597#else
62598#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62599  uint64x2x4_t __s1 = __p1; \
62600  uint64x2x4_t __rev1; \
62601  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62602  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62603  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62604  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62605  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
62606})
62607#endif
62608
62609#ifdef __LITTLE_ENDIAN__
62610#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62611  uint16x8x4_t __s1 = __p1; \
62612  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
62613})
62614#else
62615#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62616  uint16x8x4_t __s1 = __p1; \
62617  uint16x8x4_t __rev1; \
62618  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62619  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62620  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62621  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62622  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
62623})
62624#endif
62625
62626#ifdef __LITTLE_ENDIAN__
62627#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62628  int8x16x4_t __s1 = __p1; \
62629  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
62630})
62631#else
62632#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62633  int8x16x4_t __s1 = __p1; \
62634  int8x16x4_t __rev1; \
62635  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62636  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62637  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62638  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62639  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
62640})
62641#endif
62642
62643#ifdef __LITTLE_ENDIAN__
62644#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62645  float64x2x4_t __s1 = __p1; \
62646  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
62647})
62648#else
62649#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62650  float64x2x4_t __s1 = __p1; \
62651  float64x2x4_t __rev1; \
62652  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62653  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62654  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62655  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62656  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
62657})
62658#endif
62659
62660#ifdef __LITTLE_ENDIAN__
62661#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62662  float32x4x4_t __s1 = __p1; \
62663  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
62664})
62665#else
62666#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62667  float32x4x4_t __s1 = __p1; \
62668  float32x4x4_t __rev1; \
62669  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62670  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62671  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62672  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62673  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
62674})
62675#endif
62676
62677#ifdef __LITTLE_ENDIAN__
62678#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62679  float16x8x4_t __s1 = __p1; \
62680  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
62681})
62682#else
62683#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62684  float16x8x4_t __s1 = __p1; \
62685  float16x8x4_t __rev1; \
62686  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62687  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62688  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62689  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62690  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
62691})
62692#endif
62693
62694#ifdef __LITTLE_ENDIAN__
62695#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62696  int32x4x4_t __s1 = __p1; \
62697  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
62698})
62699#else
62700#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62701  int32x4x4_t __s1 = __p1; \
62702  int32x4x4_t __rev1; \
62703  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62704  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62705  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62706  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62707  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
62708})
62709#endif
62710
62711#ifdef __LITTLE_ENDIAN__
62712#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62713  int64x2x4_t __s1 = __p1; \
62714  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
62715})
62716#else
62717#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62718  int64x2x4_t __s1 = __p1; \
62719  int64x2x4_t __rev1; \
62720  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62721  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62722  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62723  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62724  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
62725})
62726#endif
62727
62728#ifdef __LITTLE_ENDIAN__
62729#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62730  int16x8x4_t __s1 = __p1; \
62731  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
62732})
62733#else
62734#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62735  int16x8x4_t __s1 = __p1; \
62736  int16x8x4_t __rev1; \
62737  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62738  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62739  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62740  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62741  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
62742})
62743#endif
62744
62745#ifdef __LITTLE_ENDIAN__
62746#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62747  uint8x8x4_t __s1 = __p1; \
62748  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
62749})
62750#else
62751#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62752  uint8x8x4_t __s1 = __p1; \
62753  uint8x8x4_t __rev1; \
62754  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62755  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62756  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62757  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62758  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
62759})
62760#endif
62761
62762#ifdef __LITTLE_ENDIAN__
62763#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62764  uint32x2x4_t __s1 = __p1; \
62765  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
62766})
62767#else
62768#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62769  uint32x2x4_t __s1 = __p1; \
62770  uint32x2x4_t __rev1; \
62771  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62772  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62773  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62774  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62775  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
62776})
62777#endif
62778
62779#ifdef __LITTLE_ENDIAN__
62780#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62781  uint64x1x4_t __s1 = __p1; \
62782  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62783})
62784#else
62785#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62786  uint64x1x4_t __s1 = __p1; \
62787  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62788})
62789#endif
62790
62791#ifdef __LITTLE_ENDIAN__
62792#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62793  uint16x4x4_t __s1 = __p1; \
62794  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
62795})
62796#else
62797#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62798  uint16x4x4_t __s1 = __p1; \
62799  uint16x4x4_t __rev1; \
62800  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62801  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62802  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62803  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62804  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
62805})
62806#endif
62807
62808#ifdef __LITTLE_ENDIAN__
62809#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62810  int8x8x4_t __s1 = __p1; \
62811  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
62812})
62813#else
62814#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62815  int8x8x4_t __s1 = __p1; \
62816  int8x8x4_t __rev1; \
62817  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62818  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62819  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62820  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62821  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
62822})
62823#endif
62824
62825#ifdef __LITTLE_ENDIAN__
62826#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62827  float64x1x4_t __s1 = __p1; \
62828  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62829})
62830#else
62831#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62832  float64x1x4_t __s1 = __p1; \
62833  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62834})
62835#endif
62836
62837#ifdef __LITTLE_ENDIAN__
62838#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62839  float32x2x4_t __s1 = __p1; \
62840  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
62841})
62842#else
62843#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62844  float32x2x4_t __s1 = __p1; \
62845  float32x2x4_t __rev1; \
62846  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62847  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62848  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62849  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62850  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
62851})
62852#endif
62853
62854#ifdef __LITTLE_ENDIAN__
62855#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62856  float16x4x4_t __s1 = __p1; \
62857  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
62858})
62859#else
62860#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62861  float16x4x4_t __s1 = __p1; \
62862  float16x4x4_t __rev1; \
62863  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62864  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62865  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62866  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62867  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
62868})
62869#endif
62870
62871#ifdef __LITTLE_ENDIAN__
62872#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62873  int32x2x4_t __s1 = __p1; \
62874  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
62875})
62876#else
62877#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62878  int32x2x4_t __s1 = __p1; \
62879  int32x2x4_t __rev1; \
62880  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62881  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62882  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62883  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62884  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
62885})
62886#endif
62887
62888#ifdef __LITTLE_ENDIAN__
62889#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62890  int64x1x4_t __s1 = __p1; \
62891  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62892})
62893#else
62894#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62895  int64x1x4_t __s1 = __p1; \
62896  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62897})
62898#endif
62899
62900#ifdef __LITTLE_ENDIAN__
62901#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62902  int16x4x4_t __s1 = __p1; \
62903  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
62904})
62905#else
62906#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62907  int16x4x4_t __s1 = __p1; \
62908  int16x4x4_t __rev1; \
62909  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62910  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62911  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62912  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62913  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
62914})
62915#endif
62916
62917#ifdef __LITTLE_ENDIAN__
62918#define vst2_p64(__p0, __p1) __extension__ ({ \
62919  poly64x1x2_t __s1 = __p1; \
62920  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62921})
62922#else
62923#define vst2_p64(__p0, __p1) __extension__ ({ \
62924  poly64x1x2_t __s1 = __p1; \
62925  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62926})
62927#endif
62928
62929#ifdef __LITTLE_ENDIAN__
62930#define vst2q_p64(__p0, __p1) __extension__ ({ \
62931  poly64x2x2_t __s1 = __p1; \
62932  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
62933})
62934#else
62935#define vst2q_p64(__p0, __p1) __extension__ ({ \
62936  poly64x2x2_t __s1 = __p1; \
62937  poly64x2x2_t __rev1; \
62938  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62939  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62940  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
62941})
62942#endif
62943
62944#ifdef __LITTLE_ENDIAN__
62945#define vst2q_u64(__p0, __p1) __extension__ ({ \
62946  uint64x2x2_t __s1 = __p1; \
62947  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
62948})
62949#else
62950#define vst2q_u64(__p0, __p1) __extension__ ({ \
62951  uint64x2x2_t __s1 = __p1; \
62952  uint64x2x2_t __rev1; \
62953  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62954  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62955  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
62956})
62957#endif
62958
62959#ifdef __LITTLE_ENDIAN__
62960#define vst2q_f64(__p0, __p1) __extension__ ({ \
62961  float64x2x2_t __s1 = __p1; \
62962  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 42); \
62963})
62964#else
62965#define vst2q_f64(__p0, __p1) __extension__ ({ \
62966  float64x2x2_t __s1 = __p1; \
62967  float64x2x2_t __rev1; \
62968  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62969  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62970  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
62971})
62972#endif
62973
62974#ifdef __LITTLE_ENDIAN__
62975#define vst2q_s64(__p0, __p1) __extension__ ({ \
62976  int64x2x2_t __s1 = __p1; \
62977  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 35); \
62978})
62979#else
62980#define vst2q_s64(__p0, __p1) __extension__ ({ \
62981  int64x2x2_t __s1 = __p1; \
62982  int64x2x2_t __rev1; \
62983  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62984  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62985  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
62986})
62987#endif
62988
62989#ifdef __LITTLE_ENDIAN__
62990#define vst2_f64(__p0, __p1) __extension__ ({ \
62991  float64x1x2_t __s1 = __p1; \
62992  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62993})
62994#else
62995#define vst2_f64(__p0, __p1) __extension__ ({ \
62996  float64x1x2_t __s1 = __p1; \
62997  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62998})
62999#endif
63000
63001#ifdef __LITTLE_ENDIAN__
63002#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63003  poly64x1x2_t __s1 = __p1; \
63004  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63005})
63006#else
63007#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63008  poly64x1x2_t __s1 = __p1; \
63009  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63010})
63011#endif
63012
63013#ifdef __LITTLE_ENDIAN__
63014#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63015  poly8x16x2_t __s1 = __p1; \
63016  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
63017})
63018#else
63019#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63020  poly8x16x2_t __s1 = __p1; \
63021  poly8x16x2_t __rev1; \
63022  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63023  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63024  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
63025})
63026#endif
63027
63028#ifdef __LITTLE_ENDIAN__
63029#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63030  poly64x2x2_t __s1 = __p1; \
63031  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
63032})
63033#else
63034#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63035  poly64x2x2_t __s1 = __p1; \
63036  poly64x2x2_t __rev1; \
63037  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63038  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63039  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
63040})
63041#endif
63042
63043#ifdef __LITTLE_ENDIAN__
63044#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63045  uint8x16x2_t __s1 = __p1; \
63046  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
63047})
63048#else
63049#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63050  uint8x16x2_t __s1 = __p1; \
63051  uint8x16x2_t __rev1; \
63052  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63053  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63054  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
63055})
63056#endif
63057
63058#ifdef __LITTLE_ENDIAN__
63059#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63060  uint64x2x2_t __s1 = __p1; \
63061  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
63062})
63063#else
63064#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63065  uint64x2x2_t __s1 = __p1; \
63066  uint64x2x2_t __rev1; \
63067  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63068  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63069  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
63070})
63071#endif
63072
63073#ifdef __LITTLE_ENDIAN__
63074#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63075  int8x16x2_t __s1 = __p1; \
63076  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
63077})
63078#else
63079#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63080  int8x16x2_t __s1 = __p1; \
63081  int8x16x2_t __rev1; \
63082  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63083  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63084  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
63085})
63086#endif
63087
63088#ifdef __LITTLE_ENDIAN__
63089#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63090  float64x2x2_t __s1 = __p1; \
63091  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 42); \
63092})
63093#else
63094#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63095  float64x2x2_t __s1 = __p1; \
63096  float64x2x2_t __rev1; \
63097  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63098  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63099  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
63100})
63101#endif
63102
63103#ifdef __LITTLE_ENDIAN__
63104#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63105  int64x2x2_t __s1 = __p1; \
63106  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 35); \
63107})
63108#else
63109#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63110  int64x2x2_t __s1 = __p1; \
63111  int64x2x2_t __rev1; \
63112  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63113  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63114  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
63115})
63116#endif
63117
63118#ifdef __LITTLE_ENDIAN__
63119#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63120  uint64x1x2_t __s1 = __p1; \
63121  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63122})
63123#else
63124#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63125  uint64x1x2_t __s1 = __p1; \
63126  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63127})
63128#endif
63129
63130#ifdef __LITTLE_ENDIAN__
63131#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63132  float64x1x2_t __s1 = __p1; \
63133  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63134})
63135#else
63136#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63137  float64x1x2_t __s1 = __p1; \
63138  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63139})
63140#endif
63141
63142#ifdef __LITTLE_ENDIAN__
63143#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63144  int64x1x2_t __s1 = __p1; \
63145  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63146})
63147#else
63148#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63149  int64x1x2_t __s1 = __p1; \
63150  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63151})
63152#endif
63153
63154#ifdef __LITTLE_ENDIAN__
63155#define vst3_p64(__p0, __p1) __extension__ ({ \
63156  poly64x1x3_t __s1 = __p1; \
63157  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63158})
63159#else
63160#define vst3_p64(__p0, __p1) __extension__ ({ \
63161  poly64x1x3_t __s1 = __p1; \
63162  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63163})
63164#endif
63165
63166#ifdef __LITTLE_ENDIAN__
63167#define vst3q_p64(__p0, __p1) __extension__ ({ \
63168  poly64x2x3_t __s1 = __p1; \
63169  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
63170})
63171#else
63172#define vst3q_p64(__p0, __p1) __extension__ ({ \
63173  poly64x2x3_t __s1 = __p1; \
63174  poly64x2x3_t __rev1; \
63175  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63176  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63177  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63178  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
63179})
63180#endif
63181
63182#ifdef __LITTLE_ENDIAN__
63183#define vst3q_u64(__p0, __p1) __extension__ ({ \
63184  uint64x2x3_t __s1 = __p1; \
63185  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
63186})
63187#else
63188#define vst3q_u64(__p0, __p1) __extension__ ({ \
63189  uint64x2x3_t __s1 = __p1; \
63190  uint64x2x3_t __rev1; \
63191  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63192  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63193  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63194  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
63195})
63196#endif
63197
63198#ifdef __LITTLE_ENDIAN__
63199#define vst3q_f64(__p0, __p1) __extension__ ({ \
63200  float64x2x3_t __s1 = __p1; \
63201  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
63202})
63203#else
63204#define vst3q_f64(__p0, __p1) __extension__ ({ \
63205  float64x2x3_t __s1 = __p1; \
63206  float64x2x3_t __rev1; \
63207  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63208  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63209  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63210  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
63211})
63212#endif
63213
63214#ifdef __LITTLE_ENDIAN__
63215#define vst3q_s64(__p0, __p1) __extension__ ({ \
63216  int64x2x3_t __s1 = __p1; \
63217  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
63218})
63219#else
63220#define vst3q_s64(__p0, __p1) __extension__ ({ \
63221  int64x2x3_t __s1 = __p1; \
63222  int64x2x3_t __rev1; \
63223  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63224  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63225  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63226  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
63227})
63228#endif
63229
63230#ifdef __LITTLE_ENDIAN__
63231#define vst3_f64(__p0, __p1) __extension__ ({ \
63232  float64x1x3_t __s1 = __p1; \
63233  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63234})
63235#else
63236#define vst3_f64(__p0, __p1) __extension__ ({ \
63237  float64x1x3_t __s1 = __p1; \
63238  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63239})
63240#endif
63241
63242#ifdef __LITTLE_ENDIAN__
63243#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63244  poly64x1x3_t __s1 = __p1; \
63245  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63246})
63247#else
63248#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63249  poly64x1x3_t __s1 = __p1; \
63250  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63251})
63252#endif
63253
63254#ifdef __LITTLE_ENDIAN__
63255#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63256  poly8x16x3_t __s1 = __p1; \
63257  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
63258})
63259#else
63260#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63261  poly8x16x3_t __s1 = __p1; \
63262  poly8x16x3_t __rev1; \
63263  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63264  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63265  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63266  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
63267})
63268#endif
63269
63270#ifdef __LITTLE_ENDIAN__
63271#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63272  poly64x2x3_t __s1 = __p1; \
63273  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
63274})
63275#else
63276#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63277  poly64x2x3_t __s1 = __p1; \
63278  poly64x2x3_t __rev1; \
63279  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63280  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63281  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63282  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
63283})
63284#endif
63285
63286#ifdef __LITTLE_ENDIAN__
63287#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63288  uint8x16x3_t __s1 = __p1; \
63289  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
63290})
63291#else
63292#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63293  uint8x16x3_t __s1 = __p1; \
63294  uint8x16x3_t __rev1; \
63295  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63296  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63297  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63298  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
63299})
63300#endif
63301
63302#ifdef __LITTLE_ENDIAN__
63303#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63304  uint64x2x3_t __s1 = __p1; \
63305  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
63306})
63307#else
63308#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63309  uint64x2x3_t __s1 = __p1; \
63310  uint64x2x3_t __rev1; \
63311  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63312  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63313  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63314  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
63315})
63316#endif
63317
63318#ifdef __LITTLE_ENDIAN__
63319#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63320  int8x16x3_t __s1 = __p1; \
63321  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
63322})
63323#else
63324#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63325  int8x16x3_t __s1 = __p1; \
63326  int8x16x3_t __rev1; \
63327  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63328  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63329  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63330  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
63331})
63332#endif
63333
63334#ifdef __LITTLE_ENDIAN__
63335#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63336  float64x2x3_t __s1 = __p1; \
63337  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
63338})
63339#else
63340#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63341  float64x2x3_t __s1 = __p1; \
63342  float64x2x3_t __rev1; \
63343  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63344  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63345  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63346  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
63347})
63348#endif
63349
63350#ifdef __LITTLE_ENDIAN__
63351#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63352  int64x2x3_t __s1 = __p1; \
63353  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
63354})
63355#else
63356#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63357  int64x2x3_t __s1 = __p1; \
63358  int64x2x3_t __rev1; \
63359  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63360  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63361  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63362  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
63363})
63364#endif
63365
63366#ifdef __LITTLE_ENDIAN__
63367#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63368  uint64x1x3_t __s1 = __p1; \
63369  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63370})
63371#else
63372#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63373  uint64x1x3_t __s1 = __p1; \
63374  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63375})
63376#endif
63377
63378#ifdef __LITTLE_ENDIAN__
63379#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63380  float64x1x3_t __s1 = __p1; \
63381  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63382})
63383#else
63384#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63385  float64x1x3_t __s1 = __p1; \
63386  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63387})
63388#endif
63389
63390#ifdef __LITTLE_ENDIAN__
63391#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63392  int64x1x3_t __s1 = __p1; \
63393  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63394})
63395#else
63396#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63397  int64x1x3_t __s1 = __p1; \
63398  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63399})
63400#endif
63401
63402#ifdef __LITTLE_ENDIAN__
63403#define vst4_p64(__p0, __p1) __extension__ ({ \
63404  poly64x1x4_t __s1 = __p1; \
63405  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63406})
63407#else
63408#define vst4_p64(__p0, __p1) __extension__ ({ \
63409  poly64x1x4_t __s1 = __p1; \
63410  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63411})
63412#endif
63413
63414#ifdef __LITTLE_ENDIAN__
63415#define vst4q_p64(__p0, __p1) __extension__ ({ \
63416  poly64x2x4_t __s1 = __p1; \
63417  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
63418})
63419#else
63420#define vst4q_p64(__p0, __p1) __extension__ ({ \
63421  poly64x2x4_t __s1 = __p1; \
63422  poly64x2x4_t __rev1; \
63423  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63424  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63425  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63426  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63427  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
63428})
63429#endif
63430
63431#ifdef __LITTLE_ENDIAN__
63432#define vst4q_u64(__p0, __p1) __extension__ ({ \
63433  uint64x2x4_t __s1 = __p1; \
63434  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
63435})
63436#else
63437#define vst4q_u64(__p0, __p1) __extension__ ({ \
63438  uint64x2x4_t __s1 = __p1; \
63439  uint64x2x4_t __rev1; \
63440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63443  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63444  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
63445})
63446#endif
63447
63448#ifdef __LITTLE_ENDIAN__
63449#define vst4q_f64(__p0, __p1) __extension__ ({ \
63450  float64x2x4_t __s1 = __p1; \
63451  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
63452})
63453#else
63454#define vst4q_f64(__p0, __p1) __extension__ ({ \
63455  float64x2x4_t __s1 = __p1; \
63456  float64x2x4_t __rev1; \
63457  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63458  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63459  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63460  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63461  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
63462})
63463#endif
63464
63465#ifdef __LITTLE_ENDIAN__
63466#define vst4q_s64(__p0, __p1) __extension__ ({ \
63467  int64x2x4_t __s1 = __p1; \
63468  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
63469})
63470#else
63471#define vst4q_s64(__p0, __p1) __extension__ ({ \
63472  int64x2x4_t __s1 = __p1; \
63473  int64x2x4_t __rev1; \
63474  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63475  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63476  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63477  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63478  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
63479})
63480#endif
63481
63482#ifdef __LITTLE_ENDIAN__
63483#define vst4_f64(__p0, __p1) __extension__ ({ \
63484  float64x1x4_t __s1 = __p1; \
63485  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63486})
63487#else
63488#define vst4_f64(__p0, __p1) __extension__ ({ \
63489  float64x1x4_t __s1 = __p1; \
63490  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63491})
63492#endif
63493
63494#ifdef __LITTLE_ENDIAN__
63495#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63496  poly64x1x4_t __s1 = __p1; \
63497  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63498})
63499#else
63500#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63501  poly64x1x4_t __s1 = __p1; \
63502  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63503})
63504#endif
63505
63506#ifdef __LITTLE_ENDIAN__
63507#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63508  poly8x16x4_t __s1 = __p1; \
63509  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
63510})
63511#else
63512#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63513  poly8x16x4_t __s1 = __p1; \
63514  poly8x16x4_t __rev1; \
63515  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63516  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63517  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63518  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63519  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
63520})
63521#endif
63522
63523#ifdef __LITTLE_ENDIAN__
63524#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63525  poly64x2x4_t __s1 = __p1; \
63526  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
63527})
63528#else
63529#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63530  poly64x2x4_t __s1 = __p1; \
63531  poly64x2x4_t __rev1; \
63532  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63533  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63534  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63535  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63536  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
63537})
63538#endif
63539
63540#ifdef __LITTLE_ENDIAN__
63541#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63542  uint8x16x4_t __s1 = __p1; \
63543  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
63544})
63545#else
63546#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63547  uint8x16x4_t __s1 = __p1; \
63548  uint8x16x4_t __rev1; \
63549  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63550  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63551  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63552  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63553  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
63554})
63555#endif
63556
63557#ifdef __LITTLE_ENDIAN__
63558#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63559  uint64x2x4_t __s1 = __p1; \
63560  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
63561})
63562#else
63563#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63564  uint64x2x4_t __s1 = __p1; \
63565  uint64x2x4_t __rev1; \
63566  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63567  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63568  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63569  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63570  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
63571})
63572#endif
63573
63574#ifdef __LITTLE_ENDIAN__
63575#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63576  int8x16x4_t __s1 = __p1; \
63577  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
63578})
63579#else
63580#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63581  int8x16x4_t __s1 = __p1; \
63582  int8x16x4_t __rev1; \
63583  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63584  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63585  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63586  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63587  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
63588})
63589#endif
63590
63591#ifdef __LITTLE_ENDIAN__
63592#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63593  float64x2x4_t __s1 = __p1; \
63594  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
63595})
63596#else
63597#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63598  float64x2x4_t __s1 = __p1; \
63599  float64x2x4_t __rev1; \
63600  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63601  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63602  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63603  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63604  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
63605})
63606#endif
63607
63608#ifdef __LITTLE_ENDIAN__
63609#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63610  int64x2x4_t __s1 = __p1; \
63611  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
63612})
63613#else
63614#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63615  int64x2x4_t __s1 = __p1; \
63616  int64x2x4_t __rev1; \
63617  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63618  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63619  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63620  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63621  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
63622})
63623#endif
63624
63625#ifdef __LITTLE_ENDIAN__
63626#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63627  uint64x1x4_t __s1 = __p1; \
63628  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63629})
63630#else
63631#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63632  uint64x1x4_t __s1 = __p1; \
63633  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63634})
63635#endif
63636
63637#ifdef __LITTLE_ENDIAN__
63638#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63639  float64x1x4_t __s1 = __p1; \
63640  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63641})
63642#else
63643#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63644  float64x1x4_t __s1 = __p1; \
63645  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63646})
63647#endif
63648
63649#ifdef __LITTLE_ENDIAN__
63650#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63651  int64x1x4_t __s1 = __p1; \
63652  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63653})
63654#else
63655#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63656  int64x1x4_t __s1 = __p1; \
63657  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63658})
63659#endif
63660
63661#ifdef __LITTLE_ENDIAN__
63662#define vstrq_p128(__p0, __p1) __extension__ ({ \
63663  poly128_t __s1 = __p1; \
63664  __builtin_neon_vstrq_p128(__p0, __s1); \
63665})
63666#else
63667#define vstrq_p128(__p0, __p1) __extension__ ({ \
63668  poly128_t __s1 = __p1; \
63669  __builtin_neon_vstrq_p128(__p0, __s1); \
63670})
63671#endif
63672
63673#ifdef __LITTLE_ENDIAN__
63674__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63675  uint64_t __ret;
63676  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63677  return __ret;
63678}
63679#else
63680__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63681  uint64_t __ret;
63682  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63683  return __ret;
63684}
63685#endif
63686
63687#ifdef __LITTLE_ENDIAN__
63688__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63689  int64_t __ret;
63690  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63691  return __ret;
63692}
63693#else
63694__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63695  int64_t __ret;
63696  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63697  return __ret;
63698}
63699#endif
63700
63701#ifdef __LITTLE_ENDIAN__
63702__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63703  float64x2_t __ret;
63704  __ret = __p0 - __p1;
63705  return __ret;
63706}
63707#else
63708__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63709  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63710  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63711  float64x2_t __ret;
63712  __ret = __rev0 - __rev1;
63713  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63714  return __ret;
63715}
63716#endif
63717
63718#ifdef __LITTLE_ENDIAN__
63719__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63720  float64x1_t __ret;
63721  __ret = __p0 - __p1;
63722  return __ret;
63723}
63724#else
63725__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63726  float64x1_t __ret;
63727  __ret = __p0 - __p1;
63728  return __ret;
63729}
63730#endif
63731
63732#ifdef __LITTLE_ENDIAN__
63733__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63734  uint16x8_t __ret;
63735  __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
63736  return __ret;
63737}
63738#else
63739__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63740  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63741  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63742  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63743  uint16x8_t __ret;
63744  __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
63745  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63746  return __ret;
63747}
63748#endif
63749
63750#ifdef __LITTLE_ENDIAN__
63751__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63752  uint32x4_t __ret;
63753  __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
63754  return __ret;
63755}
63756#else
63757__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63758  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63759  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63760  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63761  uint32x4_t __ret;
63762  __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
63763  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63764  return __ret;
63765}
63766#endif
63767
63768#ifdef __LITTLE_ENDIAN__
63769__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63770  uint8x16_t __ret;
63771  __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
63772  return __ret;
63773}
63774#else
63775__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63776  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63777  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63778  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63779  uint8x16_t __ret;
63780  __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
63781  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63782  return __ret;
63783}
63784#endif
63785
63786#ifdef __LITTLE_ENDIAN__
63787__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63788  int16x8_t __ret;
63789  __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
63790  return __ret;
63791}
63792#else
63793__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63794  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63795  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63796  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63797  int16x8_t __ret;
63798  __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
63799  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63800  return __ret;
63801}
63802#endif
63803
63804#ifdef __LITTLE_ENDIAN__
63805__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63806  int32x4_t __ret;
63807  __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
63808  return __ret;
63809}
63810#else
63811__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63812  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63813  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63814  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63815  int32x4_t __ret;
63816  __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
63817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63818  return __ret;
63819}
63820#endif
63821
63822#ifdef __LITTLE_ENDIAN__
63823__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63824  int8x16_t __ret;
63825  __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
63826  return __ret;
63827}
63828#else
63829__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63830  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63831  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63832  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63833  int8x16_t __ret;
63834  __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
63835  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63836  return __ret;
63837}
63838#endif
63839
63840#ifdef __LITTLE_ENDIAN__
63841__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63842  uint16x8_t __ret;
63843  __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
63844  return __ret;
63845}
63846#else
63847__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63848  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63849  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63850  uint16x8_t __ret;
63851  __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
63852  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63853  return __ret;
63854}
63855#endif
63856
63857#ifdef __LITTLE_ENDIAN__
63858__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63859  uint64x2_t __ret;
63860  __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
63861  return __ret;
63862}
63863#else
63864__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63865  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63866  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63867  uint64x2_t __ret;
63868  __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
63869  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63870  return __ret;
63871}
63872#endif
63873
63874#ifdef __LITTLE_ENDIAN__
63875__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63876  uint32x4_t __ret;
63877  __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
63878  return __ret;
63879}
63880#else
63881__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63882  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63883  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63884  uint32x4_t __ret;
63885  __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
63886  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63887  return __ret;
63888}
63889#endif
63890
63891#ifdef __LITTLE_ENDIAN__
63892__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63893  int16x8_t __ret;
63894  __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
63895  return __ret;
63896}
63897#else
63898__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63899  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63900  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63901  int16x8_t __ret;
63902  __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
63903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63904  return __ret;
63905}
63906#endif
63907
63908#ifdef __LITTLE_ENDIAN__
63909__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63910  int64x2_t __ret;
63911  __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
63912  return __ret;
63913}
63914#else
63915__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63916  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63917  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63918  int64x2_t __ret;
63919  __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
63920  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63921  return __ret;
63922}
63923#endif
63924
63925#ifdef __LITTLE_ENDIAN__
63926__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63927  int32x4_t __ret;
63928  __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
63929  return __ret;
63930}
63931#else
63932__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63933  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63934  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63935  int32x4_t __ret;
63936  __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
63937  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63938  return __ret;
63939}
63940#endif
63941
63942#ifdef __LITTLE_ENDIAN__
63943__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63944  uint16x8_t __ret;
63945  __ret = __p0 - vmovl_high_u8(__p1);
63946  return __ret;
63947}
63948#else
63949__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63950  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63951  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63952  uint16x8_t __ret;
63953  __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
63954  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63955  return __ret;
63956}
63957#endif
63958
63959#ifdef __LITTLE_ENDIAN__
63960__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63961  uint64x2_t __ret;
63962  __ret = __p0 - vmovl_high_u32(__p1);
63963  return __ret;
63964}
63965#else
63966__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63967  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63968  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63969  uint64x2_t __ret;
63970  __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
63971  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63972  return __ret;
63973}
63974#endif
63975
63976#ifdef __LITTLE_ENDIAN__
63977__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63978  uint32x4_t __ret;
63979  __ret = __p0 - vmovl_high_u16(__p1);
63980  return __ret;
63981}
63982#else
63983__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63984  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63985  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63986  uint32x4_t __ret;
63987  __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
63988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63989  return __ret;
63990}
63991#endif
63992
63993#ifdef __LITTLE_ENDIAN__
63994__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
63995  int16x8_t __ret;
63996  __ret = __p0 - vmovl_high_s8(__p1);
63997  return __ret;
63998}
63999#else
64000__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
64001  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64002  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64003  int16x8_t __ret;
64004  __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
64005  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64006  return __ret;
64007}
64008#endif
64009
64010#ifdef __LITTLE_ENDIAN__
64011__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64012  int64x2_t __ret;
64013  __ret = __p0 - vmovl_high_s32(__p1);
64014  return __ret;
64015}
64016#else
64017__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64018  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64019  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64020  int64x2_t __ret;
64021  __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
64022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64023  return __ret;
64024}
64025#endif
64026
64027#ifdef __LITTLE_ENDIAN__
64028__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64029  int32x4_t __ret;
64030  __ret = __p0 - vmovl_high_s16(__p1);
64031  return __ret;
64032}
64033#else
64034__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64035  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64036  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64037  int32x4_t __ret;
64038  __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
64039  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64040  return __ret;
64041}
64042#endif
64043
64044#ifdef __LITTLE_ENDIAN__
64045__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64046  poly8x8_t __ret;
64047  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64048  return __ret;
64049}
64050#else
64051__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64052  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64053  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64054  poly8x8_t __ret;
64055  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64056  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64057  return __ret;
64058}
64059#endif
64060
64061#ifdef __LITTLE_ENDIAN__
64062__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64063  poly16x4_t __ret;
64064  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64065  return __ret;
64066}
64067#else
64068__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64069  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64070  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64071  poly16x4_t __ret;
64072  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64073  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64074  return __ret;
64075}
64076#endif
64077
64078#ifdef __LITTLE_ENDIAN__
64079__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64080  poly8x16_t __ret;
64081  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64082  return __ret;
64083}
64084#else
64085__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64086  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64087  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64088  poly8x16_t __ret;
64089  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64090  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64091  return __ret;
64092}
64093#endif
64094
64095#ifdef __LITTLE_ENDIAN__
64096__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64097  poly64x2_t __ret;
64098  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64099  return __ret;
64100}
64101#else
64102__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64103  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64104  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64105  poly64x2_t __ret;
64106  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64108  return __ret;
64109}
64110#endif
64111
64112#ifdef __LITTLE_ENDIAN__
64113__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64114  poly16x8_t __ret;
64115  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64116  return __ret;
64117}
64118#else
64119__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64120  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64121  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64122  poly16x8_t __ret;
64123  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64124  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64125  return __ret;
64126}
64127#endif
64128
64129#ifdef __LITTLE_ENDIAN__
64130__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64131  uint8x16_t __ret;
64132  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64133  return __ret;
64134}
64135#else
64136__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64137  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64138  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64139  uint8x16_t __ret;
64140  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64141  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64142  return __ret;
64143}
64144#endif
64145
64146#ifdef __LITTLE_ENDIAN__
64147__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64148  uint32x4_t __ret;
64149  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64150  return __ret;
64151}
64152#else
64153__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64154  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64155  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64156  uint32x4_t __ret;
64157  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64158  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64159  return __ret;
64160}
64161#endif
64162
64163#ifdef __LITTLE_ENDIAN__
64164__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64165  uint64x2_t __ret;
64166  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64167  return __ret;
64168}
64169#else
64170__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64171  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64172  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64173  uint64x2_t __ret;
64174  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64175  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64176  return __ret;
64177}
64178#endif
64179
64180#ifdef __LITTLE_ENDIAN__
64181__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64182  uint16x8_t __ret;
64183  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64184  return __ret;
64185}
64186#else
64187__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64188  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64189  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64190  uint16x8_t __ret;
64191  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64192  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64193  return __ret;
64194}
64195#endif
64196
64197#ifdef __LITTLE_ENDIAN__
64198__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64199  int8x16_t __ret;
64200  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64201  return __ret;
64202}
64203#else
64204__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64205  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64206  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64207  int8x16_t __ret;
64208  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64209  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64210  return __ret;
64211}
64212#endif
64213
64214#ifdef __LITTLE_ENDIAN__
64215__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64216  float64x2_t __ret;
64217  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64218  return __ret;
64219}
64220#else
64221__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64222  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64223  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64224  float64x2_t __ret;
64225  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64227  return __ret;
64228}
64229#endif
64230
64231#ifdef __LITTLE_ENDIAN__
64232__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64233  float32x4_t __ret;
64234  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64235  return __ret;
64236}
64237#else
64238__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64239  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64240  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64241  float32x4_t __ret;
64242  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64244  return __ret;
64245}
64246#endif
64247
64248#ifdef __LITTLE_ENDIAN__
64249__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64250  int32x4_t __ret;
64251  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64252  return __ret;
64253}
64254#else
64255__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64256  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64257  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64258  int32x4_t __ret;
64259  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64260  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64261  return __ret;
64262}
64263#endif
64264
64265#ifdef __LITTLE_ENDIAN__
64266__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64267  int64x2_t __ret;
64268  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64269  return __ret;
64270}
64271#else
64272__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64273  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64274  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64275  int64x2_t __ret;
64276  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64277  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64278  return __ret;
64279}
64280#endif
64281
64282#ifdef __LITTLE_ENDIAN__
64283__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64284  int16x8_t __ret;
64285  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64286  return __ret;
64287}
64288#else
64289__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64290  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64291  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64292  int16x8_t __ret;
64293  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64295  return __ret;
64296}
64297#endif
64298
64299#ifdef __LITTLE_ENDIAN__
64300__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64301  uint8x8_t __ret;
64302  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64303  return __ret;
64304}
64305#else
64306__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64307  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64308  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64309  uint8x8_t __ret;
64310  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64311  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64312  return __ret;
64313}
64314#endif
64315
64316#ifdef __LITTLE_ENDIAN__
64317__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64318  uint32x2_t __ret;
64319  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64320  return __ret;
64321}
64322#else
64323__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64324  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64325  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64326  uint32x2_t __ret;
64327  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64329  return __ret;
64330}
64331#endif
64332
64333#ifdef __LITTLE_ENDIAN__
64334__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64335  uint16x4_t __ret;
64336  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64337  return __ret;
64338}
64339#else
64340__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64341  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64342  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64343  uint16x4_t __ret;
64344  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64345  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64346  return __ret;
64347}
64348#endif
64349
64350#ifdef __LITTLE_ENDIAN__
64351__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64352  int8x8_t __ret;
64353  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64354  return __ret;
64355}
64356#else
64357__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64358  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64359  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64360  int8x8_t __ret;
64361  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64362  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64363  return __ret;
64364}
64365#endif
64366
64367#ifdef __LITTLE_ENDIAN__
64368__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64369  float32x2_t __ret;
64370  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64371  return __ret;
64372}
64373#else
64374__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64375  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64376  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64377  float32x2_t __ret;
64378  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64379  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64380  return __ret;
64381}
64382#endif
64383
64384#ifdef __LITTLE_ENDIAN__
64385__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64386  int32x2_t __ret;
64387  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64388  return __ret;
64389}
64390#else
64391__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64392  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64393  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64394  int32x2_t __ret;
64395  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64396  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64397  return __ret;
64398}
64399#endif
64400
64401#ifdef __LITTLE_ENDIAN__
64402__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64403  int16x4_t __ret;
64404  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64405  return __ret;
64406}
64407#else
64408__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64409  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64410  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64411  int16x4_t __ret;
64412  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64413  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64414  return __ret;
64415}
64416#endif
64417
64418#ifdef __LITTLE_ENDIAN__
64419__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64420  poly8x8_t __ret;
64421  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64422  return __ret;
64423}
64424#else
64425__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64426  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64427  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64428  poly8x8_t __ret;
64429  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64430  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64431  return __ret;
64432}
64433#endif
64434
64435#ifdef __LITTLE_ENDIAN__
64436__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64437  poly16x4_t __ret;
64438  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64439  return __ret;
64440}
64441#else
64442__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64443  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64444  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64445  poly16x4_t __ret;
64446  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64447  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64448  return __ret;
64449}
64450#endif
64451
64452#ifdef __LITTLE_ENDIAN__
64453__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64454  poly8x16_t __ret;
64455  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64456  return __ret;
64457}
64458#else
64459__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64460  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64461  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64462  poly8x16_t __ret;
64463  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64464  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64465  return __ret;
64466}
64467#endif
64468
64469#ifdef __LITTLE_ENDIAN__
64470__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64471  poly64x2_t __ret;
64472  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64473  return __ret;
64474}
64475#else
64476__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64477  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64478  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64479  poly64x2_t __ret;
64480  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64481  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64482  return __ret;
64483}
64484#endif
64485
64486#ifdef __LITTLE_ENDIAN__
64487__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64488  poly16x8_t __ret;
64489  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64490  return __ret;
64491}
64492#else
64493__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64494  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64495  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64496  poly16x8_t __ret;
64497  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64498  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64499  return __ret;
64500}
64501#endif
64502
64503#ifdef __LITTLE_ENDIAN__
64504__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64505  uint8x16_t __ret;
64506  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64507  return __ret;
64508}
64509#else
64510__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64511  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64512  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64513  uint8x16_t __ret;
64514  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64515  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64516  return __ret;
64517}
64518#endif
64519
64520#ifdef __LITTLE_ENDIAN__
64521__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64522  uint32x4_t __ret;
64523  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64524  return __ret;
64525}
64526#else
64527__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64528  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64529  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64530  uint32x4_t __ret;
64531  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64533  return __ret;
64534}
64535#endif
64536
64537#ifdef __LITTLE_ENDIAN__
64538__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64539  uint64x2_t __ret;
64540  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64541  return __ret;
64542}
64543#else
64544__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64545  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64546  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64547  uint64x2_t __ret;
64548  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64549  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64550  return __ret;
64551}
64552#endif
64553
64554#ifdef __LITTLE_ENDIAN__
64555__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64556  uint16x8_t __ret;
64557  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64558  return __ret;
64559}
64560#else
64561__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64562  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64563  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64564  uint16x8_t __ret;
64565  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64566  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64567  return __ret;
64568}
64569#endif
64570
64571#ifdef __LITTLE_ENDIAN__
64572__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64573  int8x16_t __ret;
64574  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64575  return __ret;
64576}
64577#else
64578__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64579  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64580  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64581  int8x16_t __ret;
64582  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64583  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64584  return __ret;
64585}
64586#endif
64587
64588#ifdef __LITTLE_ENDIAN__
64589__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64590  float64x2_t __ret;
64591  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64592  return __ret;
64593}
64594#else
64595__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64596  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64597  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64598  float64x2_t __ret;
64599  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64601  return __ret;
64602}
64603#endif
64604
64605#ifdef __LITTLE_ENDIAN__
64606__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64607  float32x4_t __ret;
64608  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64609  return __ret;
64610}
64611#else
64612__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64613  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64614  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64615  float32x4_t __ret;
64616  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64617  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64618  return __ret;
64619}
64620#endif
64621
64622#ifdef __LITTLE_ENDIAN__
64623__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64624  int32x4_t __ret;
64625  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64626  return __ret;
64627}
64628#else
64629__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64630  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64631  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64632  int32x4_t __ret;
64633  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64635  return __ret;
64636}
64637#endif
64638
64639#ifdef __LITTLE_ENDIAN__
64640__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64641  int64x2_t __ret;
64642  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64643  return __ret;
64644}
64645#else
64646__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64647  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64648  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64649  int64x2_t __ret;
64650  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64651  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64652  return __ret;
64653}
64654#endif
64655
64656#ifdef __LITTLE_ENDIAN__
64657__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64658  int16x8_t __ret;
64659  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64660  return __ret;
64661}
64662#else
64663__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64664  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64665  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64666  int16x8_t __ret;
64667  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64668  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64669  return __ret;
64670}
64671#endif
64672
64673#ifdef __LITTLE_ENDIAN__
64674__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64675  uint8x8_t __ret;
64676  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64677  return __ret;
64678}
64679#else
64680__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64681  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64682  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64683  uint8x8_t __ret;
64684  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64685  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64686  return __ret;
64687}
64688#endif
64689
64690#ifdef __LITTLE_ENDIAN__
64691__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64692  uint32x2_t __ret;
64693  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64694  return __ret;
64695}
64696#else
64697__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64698  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64699  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64700  uint32x2_t __ret;
64701  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64702  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64703  return __ret;
64704}
64705#endif
64706
64707#ifdef __LITTLE_ENDIAN__
64708__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64709  uint16x4_t __ret;
64710  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64711  return __ret;
64712}
64713#else
64714__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64715  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64716  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64717  uint16x4_t __ret;
64718  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64719  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64720  return __ret;
64721}
64722#endif
64723
64724#ifdef __LITTLE_ENDIAN__
64725__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64726  int8x8_t __ret;
64727  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64728  return __ret;
64729}
64730#else
64731__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64732  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64733  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64734  int8x8_t __ret;
64735  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64736  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64737  return __ret;
64738}
64739#endif
64740
64741#ifdef __LITTLE_ENDIAN__
64742__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64743  float32x2_t __ret;
64744  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64745  return __ret;
64746}
64747#else
64748__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64749  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64750  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64751  float32x2_t __ret;
64752  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64753  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64754  return __ret;
64755}
64756#endif
64757
64758#ifdef __LITTLE_ENDIAN__
64759__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64760  int32x2_t __ret;
64761  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64762  return __ret;
64763}
64764#else
64765__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64766  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64767  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64768  int32x2_t __ret;
64769  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64770  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64771  return __ret;
64772}
64773#endif
64774
64775#ifdef __LITTLE_ENDIAN__
64776__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64777  int16x4_t __ret;
64778  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64779  return __ret;
64780}
64781#else
64782__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64783  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64784  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64785  int16x4_t __ret;
64786  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64787  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64788  return __ret;
64789}
64790#endif
64791
64792#ifdef __LITTLE_ENDIAN__
64793__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64794  uint64x1_t __ret;
64795  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64796  return __ret;
64797}
64798#else
64799__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64800  uint64x1_t __ret;
64801  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64802  return __ret;
64803}
64804#endif
64805
64806#ifdef __LITTLE_ENDIAN__
64807__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64808  uint64x2_t __ret;
64809  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64810  return __ret;
64811}
64812#else
64813__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64814  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64815  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64816  uint64x2_t __ret;
64817  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64818  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64819  return __ret;
64820}
64821#endif
64822
64823#ifdef __LITTLE_ENDIAN__
64824__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64825  uint64x2_t __ret;
64826  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64827  return __ret;
64828}
64829#else
64830__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64831  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64832  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64833  uint64x2_t __ret;
64834  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64835  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64836  return __ret;
64837}
64838#endif
64839
64840#ifdef __LITTLE_ENDIAN__
64841__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64842  uint64x2_t __ret;
64843  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64844  return __ret;
64845}
64846#else
64847__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64848  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64849  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64850  uint64x2_t __ret;
64851  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64852  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64853  return __ret;
64854}
64855#endif
64856
64857#ifdef __LITTLE_ENDIAN__
64858__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64859  uint64x1_t __ret;
64860  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64861  return __ret;
64862}
64863#else
64864__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64865  uint64x1_t __ret;
64866  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64867  return __ret;
64868}
64869#endif
64870
64871#ifdef __LITTLE_ENDIAN__
64872__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64873  uint64x1_t __ret;
64874  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64875  return __ret;
64876}
64877#else
64878__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64879  uint64x1_t __ret;
64880  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64881  return __ret;
64882}
64883#endif
64884
64885#ifdef __LITTLE_ENDIAN__
64886__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64887  uint64_t __ret;
64888  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64889  return __ret;
64890}
64891#else
64892__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64893  uint64_t __ret;
64894  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64895  return __ret;
64896}
64897#endif
64898
64899#ifdef __LITTLE_ENDIAN__
64900__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64901  int64_t __ret;
64902  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64903  return __ret;
64904}
64905#else
64906__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64907  int64_t __ret;
64908  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64909  return __ret;
64910}
64911#endif
64912
64913#ifdef __LITTLE_ENDIAN__
64914__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64915  int8_t __ret;
64916  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64917  return __ret;
64918}
64919#else
64920__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64921  int8_t __ret;
64922  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64923  return __ret;
64924}
64925#endif
64926
64927#ifdef __LITTLE_ENDIAN__
64928__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64929  int32_t __ret;
64930  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64931  return __ret;
64932}
64933#else
64934__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64935  int32_t __ret;
64936  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64937  return __ret;
64938}
64939#endif
64940
64941#ifdef __LITTLE_ENDIAN__
64942__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64943  int64_t __ret;
64944  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64945  return __ret;
64946}
64947#else
64948__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64949  int64_t __ret;
64950  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64951  return __ret;
64952}
64953#endif
64954
64955#ifdef __LITTLE_ENDIAN__
64956__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64957  int16_t __ret;
64958  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64959  return __ret;
64960}
64961#else
64962__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64963  int16_t __ret;
64964  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64965  return __ret;
64966}
64967#endif
64968
64969#ifdef __LITTLE_ENDIAN__
64970__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64971  int8x16_t __ret;
64972  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
64973  return __ret;
64974}
64975#else
64976__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64977  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64978  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64979  int8x16_t __ret;
64980  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
64981  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64982  return __ret;
64983}
64984#endif
64985
64986#ifdef __LITTLE_ENDIAN__
64987__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64988  int32x4_t __ret;
64989  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
64990  return __ret;
64991}
64992#else
64993__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64994  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64995  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64996  int32x4_t __ret;
64997  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
64998  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64999  return __ret;
65000}
65001#endif
65002
65003#ifdef __LITTLE_ENDIAN__
65004__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65005  int64x2_t __ret;
65006  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
65007  return __ret;
65008}
65009#else
65010__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65011  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65012  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65013  int64x2_t __ret;
65014  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
65015  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65016  return __ret;
65017}
65018#endif
65019
65020#ifdef __LITTLE_ENDIAN__
65021__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65022  int16x8_t __ret;
65023  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
65024  return __ret;
65025}
65026#else
65027__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65028  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65029  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65030  int16x8_t __ret;
65031  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
65032  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65033  return __ret;
65034}
65035#endif
65036
65037#ifdef __LITTLE_ENDIAN__
65038__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65039  int8x8_t __ret;
65040  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
65041  return __ret;
65042}
65043#else
65044__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65045  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65046  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65047  int8x8_t __ret;
65048  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
65049  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65050  return __ret;
65051}
65052#endif
65053
65054#ifdef __LITTLE_ENDIAN__
65055__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65056  int32x2_t __ret;
65057  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
65058  return __ret;
65059}
65060#else
65061__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65062  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65063  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65064  int32x2_t __ret;
65065  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
65066  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65067  return __ret;
65068}
65069#endif
65070
65071#ifdef __LITTLE_ENDIAN__
65072__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65073  int64x1_t __ret;
65074  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65075  return __ret;
65076}
65077#else
65078__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65079  int64x1_t __ret;
65080  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65081  return __ret;
65082}
65083#endif
65084
65085#ifdef __LITTLE_ENDIAN__
65086__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65087  int16x4_t __ret;
65088  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
65089  return __ret;
65090}
65091#else
65092__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65093  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65094  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65095  int16x4_t __ret;
65096  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
65097  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65098  return __ret;
65099}
65100#endif
65101
65102#ifdef __LITTLE_ENDIAN__
65103__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65104  poly8x8_t __ret;
65105  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65106  return __ret;
65107}
65108#else
65109__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65110  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65111  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65112  poly8x8_t __ret;
65113  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65114  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65115  return __ret;
65116}
65117#endif
65118
65119#ifdef __LITTLE_ENDIAN__
65120__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65121  poly16x4_t __ret;
65122  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65123  return __ret;
65124}
65125#else
65126__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65127  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65128  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65129  poly16x4_t __ret;
65130  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65131  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65132  return __ret;
65133}
65134#endif
65135
65136#ifdef __LITTLE_ENDIAN__
65137__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65138  poly8x16_t __ret;
65139  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65140  return __ret;
65141}
65142#else
65143__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65144  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65145  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65146  poly8x16_t __ret;
65147  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65148  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65149  return __ret;
65150}
65151#endif
65152
65153#ifdef __LITTLE_ENDIAN__
65154__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65155  poly64x2_t __ret;
65156  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65157  return __ret;
65158}
65159#else
65160__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65161  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65162  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65163  poly64x2_t __ret;
65164  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65166  return __ret;
65167}
65168#endif
65169
65170#ifdef __LITTLE_ENDIAN__
65171__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65172  poly16x8_t __ret;
65173  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65174  return __ret;
65175}
65176#else
65177__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65178  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65179  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65180  poly16x8_t __ret;
65181  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65182  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65183  return __ret;
65184}
65185#endif
65186
65187#ifdef __LITTLE_ENDIAN__
65188__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65189  uint8x16_t __ret;
65190  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65191  return __ret;
65192}
65193#else
65194__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65195  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65196  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65197  uint8x16_t __ret;
65198  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65199  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65200  return __ret;
65201}
65202#endif
65203
65204#ifdef __LITTLE_ENDIAN__
65205__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65206  uint32x4_t __ret;
65207  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65208  return __ret;
65209}
65210#else
65211__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65212  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65213  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65214  uint32x4_t __ret;
65215  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65216  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65217  return __ret;
65218}
65219#endif
65220
65221#ifdef __LITTLE_ENDIAN__
65222__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65223  uint64x2_t __ret;
65224  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65225  return __ret;
65226}
65227#else
65228__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65229  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65230  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65231  uint64x2_t __ret;
65232  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65233  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65234  return __ret;
65235}
65236#endif
65237
65238#ifdef __LITTLE_ENDIAN__
65239__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65240  uint16x8_t __ret;
65241  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65242  return __ret;
65243}
65244#else
65245__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65246  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65247  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65248  uint16x8_t __ret;
65249  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65250  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65251  return __ret;
65252}
65253#endif
65254
65255#ifdef __LITTLE_ENDIAN__
65256__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65257  int8x16_t __ret;
65258  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65259  return __ret;
65260}
65261#else
65262__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65263  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65264  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65265  int8x16_t __ret;
65266  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65267  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65268  return __ret;
65269}
65270#endif
65271
65272#ifdef __LITTLE_ENDIAN__
65273__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65274  float64x2_t __ret;
65275  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65276  return __ret;
65277}
65278#else
65279__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65280  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65281  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65282  float64x2_t __ret;
65283  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65284  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65285  return __ret;
65286}
65287#endif
65288
65289#ifdef __LITTLE_ENDIAN__
65290__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65291  float32x4_t __ret;
65292  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65293  return __ret;
65294}
65295#else
65296__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65297  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65298  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65299  float32x4_t __ret;
65300  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65301  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65302  return __ret;
65303}
65304#endif
65305
65306#ifdef __LITTLE_ENDIAN__
65307__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65308  int32x4_t __ret;
65309  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65310  return __ret;
65311}
65312#else
65313__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65314  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65315  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65316  int32x4_t __ret;
65317  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65318  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65319  return __ret;
65320}
65321#endif
65322
65323#ifdef __LITTLE_ENDIAN__
65324__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65325  int64x2_t __ret;
65326  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65327  return __ret;
65328}
65329#else
65330__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65331  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65332  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65333  int64x2_t __ret;
65334  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65335  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65336  return __ret;
65337}
65338#endif
65339
65340#ifdef __LITTLE_ENDIAN__
65341__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65342  int16x8_t __ret;
65343  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65344  return __ret;
65345}
65346#else
65347__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65348  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65349  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65350  int16x8_t __ret;
65351  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65353  return __ret;
65354}
65355#endif
65356
65357#ifdef __LITTLE_ENDIAN__
65358__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65359  uint8x8_t __ret;
65360  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65361  return __ret;
65362}
65363#else
65364__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65365  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65366  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65367  uint8x8_t __ret;
65368  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65369  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65370  return __ret;
65371}
65372#endif
65373
65374#ifdef __LITTLE_ENDIAN__
65375__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65376  uint32x2_t __ret;
65377  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65378  return __ret;
65379}
65380#else
65381__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65382  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65383  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65384  uint32x2_t __ret;
65385  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65386  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65387  return __ret;
65388}
65389#endif
65390
65391#ifdef __LITTLE_ENDIAN__
65392__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65393  uint16x4_t __ret;
65394  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65395  return __ret;
65396}
65397#else
65398__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65399  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65400  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65401  uint16x4_t __ret;
65402  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65403  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65404  return __ret;
65405}
65406#endif
65407
65408#ifdef __LITTLE_ENDIAN__
65409__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65410  int8x8_t __ret;
65411  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65412  return __ret;
65413}
65414#else
65415__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65416  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65417  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65418  int8x8_t __ret;
65419  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65420  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65421  return __ret;
65422}
65423#endif
65424
65425#ifdef __LITTLE_ENDIAN__
65426__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65427  float32x2_t __ret;
65428  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65429  return __ret;
65430}
65431#else
65432__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65433  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65434  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65435  float32x2_t __ret;
65436  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65437  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65438  return __ret;
65439}
65440#endif
65441
65442#ifdef __LITTLE_ENDIAN__
65443__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65444  int32x2_t __ret;
65445  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65446  return __ret;
65447}
65448#else
65449__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65450  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65451  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65452  int32x2_t __ret;
65453  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65454  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65455  return __ret;
65456}
65457#endif
65458
65459#ifdef __LITTLE_ENDIAN__
65460__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65461  int16x4_t __ret;
65462  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65463  return __ret;
65464}
65465#else
65466__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65467  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65468  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65469  int16x4_t __ret;
65470  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65471  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65472  return __ret;
65473}
65474#endif
65475
65476#ifdef __LITTLE_ENDIAN__
65477__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65478  poly8x8_t __ret;
65479  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65480  return __ret;
65481}
65482#else
65483__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65484  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65485  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65486  poly8x8_t __ret;
65487  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65488  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65489  return __ret;
65490}
65491#endif
65492
65493#ifdef __LITTLE_ENDIAN__
65494__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65495  poly16x4_t __ret;
65496  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65497  return __ret;
65498}
65499#else
65500__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65501  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65502  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65503  poly16x4_t __ret;
65504  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65505  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65506  return __ret;
65507}
65508#endif
65509
65510#ifdef __LITTLE_ENDIAN__
65511__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65512  poly8x16_t __ret;
65513  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65514  return __ret;
65515}
65516#else
65517__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65518  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65519  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65520  poly8x16_t __ret;
65521  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65522  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65523  return __ret;
65524}
65525#endif
65526
65527#ifdef __LITTLE_ENDIAN__
65528__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65529  poly64x2_t __ret;
65530  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65531  return __ret;
65532}
65533#else
65534__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65535  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65536  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65537  poly64x2_t __ret;
65538  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65539  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65540  return __ret;
65541}
65542#endif
65543
65544#ifdef __LITTLE_ENDIAN__
65545__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65546  poly16x8_t __ret;
65547  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65548  return __ret;
65549}
65550#else
65551__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65552  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65553  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65554  poly16x8_t __ret;
65555  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65556  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65557  return __ret;
65558}
65559#endif
65560
65561#ifdef __LITTLE_ENDIAN__
65562__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65563  uint8x16_t __ret;
65564  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65565  return __ret;
65566}
65567#else
65568__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65569  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65570  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65571  uint8x16_t __ret;
65572  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65573  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65574  return __ret;
65575}
65576#endif
65577
65578#ifdef __LITTLE_ENDIAN__
65579__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65580  uint32x4_t __ret;
65581  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65582  return __ret;
65583}
65584#else
65585__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65586  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65587  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65588  uint32x4_t __ret;
65589  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65590  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65591  return __ret;
65592}
65593#endif
65594
65595#ifdef __LITTLE_ENDIAN__
65596__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65597  uint64x2_t __ret;
65598  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65599  return __ret;
65600}
65601#else
65602__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65603  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65604  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65605  uint64x2_t __ret;
65606  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65607  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65608  return __ret;
65609}
65610#endif
65611
65612#ifdef __LITTLE_ENDIAN__
65613__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65614  uint16x8_t __ret;
65615  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65616  return __ret;
65617}
65618#else
65619__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65620  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65621  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65622  uint16x8_t __ret;
65623  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65625  return __ret;
65626}
65627#endif
65628
65629#ifdef __LITTLE_ENDIAN__
65630__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65631  int8x16_t __ret;
65632  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65633  return __ret;
65634}
65635#else
65636__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65637  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65638  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65639  int8x16_t __ret;
65640  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65641  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65642  return __ret;
65643}
65644#endif
65645
65646#ifdef __LITTLE_ENDIAN__
65647__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65648  float64x2_t __ret;
65649  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65650  return __ret;
65651}
65652#else
65653__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65654  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65655  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65656  float64x2_t __ret;
65657  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65659  return __ret;
65660}
65661#endif
65662
65663#ifdef __LITTLE_ENDIAN__
65664__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65665  float32x4_t __ret;
65666  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65667  return __ret;
65668}
65669#else
65670__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65671  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65672  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65673  float32x4_t __ret;
65674  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65675  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65676  return __ret;
65677}
65678#endif
65679
65680#ifdef __LITTLE_ENDIAN__
65681__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65682  int32x4_t __ret;
65683  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65684  return __ret;
65685}
65686#else
65687__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65688  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65689  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65690  int32x4_t __ret;
65691  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65692  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65693  return __ret;
65694}
65695#endif
65696
65697#ifdef __LITTLE_ENDIAN__
65698__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65699  int64x2_t __ret;
65700  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65701  return __ret;
65702}
65703#else
65704__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65705  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65706  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65707  int64x2_t __ret;
65708  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65709  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65710  return __ret;
65711}
65712#endif
65713
65714#ifdef __LITTLE_ENDIAN__
65715__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65716  int16x8_t __ret;
65717  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65718  return __ret;
65719}
65720#else
65721__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65722  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65723  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65724  int16x8_t __ret;
65725  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65726  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65727  return __ret;
65728}
65729#endif
65730
65731#ifdef __LITTLE_ENDIAN__
65732__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65733  uint8x8_t __ret;
65734  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65735  return __ret;
65736}
65737#else
65738__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65739  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65740  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65741  uint8x8_t __ret;
65742  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65743  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65744  return __ret;
65745}
65746#endif
65747
65748#ifdef __LITTLE_ENDIAN__
65749__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65750  uint32x2_t __ret;
65751  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65752  return __ret;
65753}
65754#else
65755__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65756  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65757  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65758  uint32x2_t __ret;
65759  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65760  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65761  return __ret;
65762}
65763#endif
65764
65765#ifdef __LITTLE_ENDIAN__
65766__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65767  uint16x4_t __ret;
65768  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65769  return __ret;
65770}
65771#else
65772__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65773  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65774  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65775  uint16x4_t __ret;
65776  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65777  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65778  return __ret;
65779}
65780#endif
65781
65782#ifdef __LITTLE_ENDIAN__
65783__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65784  int8x8_t __ret;
65785  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65786  return __ret;
65787}
65788#else
65789__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65790  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65791  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65792  int8x8_t __ret;
65793  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65794  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65795  return __ret;
65796}
65797#endif
65798
65799#ifdef __LITTLE_ENDIAN__
65800__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65801  float32x2_t __ret;
65802  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65803  return __ret;
65804}
65805#else
65806__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65807  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65808  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65809  float32x2_t __ret;
65810  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65812  return __ret;
65813}
65814#endif
65815
65816#ifdef __LITTLE_ENDIAN__
65817__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65818  int32x2_t __ret;
65819  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65820  return __ret;
65821}
65822#else
65823__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65824  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65825  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65826  int32x2_t __ret;
65827  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65828  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65829  return __ret;
65830}
65831#endif
65832
65833#ifdef __LITTLE_ENDIAN__
65834__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65835  int16x4_t __ret;
65836  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65837  return __ret;
65838}
65839#else
65840__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65841  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65842  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65843  int16x4_t __ret;
65844  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65845  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65846  return __ret;
65847}
65848#endif
65849
65850#ifdef __LITTLE_ENDIAN__
65851__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65852  poly8x8_t __ret;
65853  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65854  return __ret;
65855}
65856#else
65857__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65858  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65859  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65860  poly8x8_t __ret;
65861  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65863  return __ret;
65864}
65865#endif
65866
65867#ifdef __LITTLE_ENDIAN__
65868__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65869  poly16x4_t __ret;
65870  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65871  return __ret;
65872}
65873#else
65874__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65875  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65876  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65877  poly16x4_t __ret;
65878  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65879  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65880  return __ret;
65881}
65882#endif
65883
65884#ifdef __LITTLE_ENDIAN__
65885__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65886  poly8x16_t __ret;
65887  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65888  return __ret;
65889}
65890#else
65891__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65892  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65893  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65894  poly8x16_t __ret;
65895  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65896  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65897  return __ret;
65898}
65899#endif
65900
65901#ifdef __LITTLE_ENDIAN__
65902__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65903  poly64x2_t __ret;
65904  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65905  return __ret;
65906}
65907#else
65908__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65909  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65910  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65911  poly64x2_t __ret;
65912  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65914  return __ret;
65915}
65916#endif
65917
65918#ifdef __LITTLE_ENDIAN__
65919__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65920  poly16x8_t __ret;
65921  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65922  return __ret;
65923}
65924#else
65925__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65926  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65927  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65928  poly16x8_t __ret;
65929  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65930  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65931  return __ret;
65932}
65933#endif
65934
65935#ifdef __LITTLE_ENDIAN__
65936__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65937  uint8x16_t __ret;
65938  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65939  return __ret;
65940}
65941#else
65942__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65943  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65944  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65945  uint8x16_t __ret;
65946  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65947  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65948  return __ret;
65949}
65950#endif
65951
65952#ifdef __LITTLE_ENDIAN__
65953__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65954  uint32x4_t __ret;
65955  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65956  return __ret;
65957}
65958#else
65959__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65960  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65961  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65962  uint32x4_t __ret;
65963  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65964  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65965  return __ret;
65966}
65967#endif
65968
65969#ifdef __LITTLE_ENDIAN__
65970__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65971  uint64x2_t __ret;
65972  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65973  return __ret;
65974}
65975#else
65976__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65977  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65978  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65979  uint64x2_t __ret;
65980  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65981  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65982  return __ret;
65983}
65984#endif
65985
65986#ifdef __LITTLE_ENDIAN__
65987__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65988  uint16x8_t __ret;
65989  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65990  return __ret;
65991}
65992#else
65993__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65994  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65995  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65996  uint16x8_t __ret;
65997  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65998  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65999  return __ret;
66000}
66001#endif
66002
66003#ifdef __LITTLE_ENDIAN__
66004__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66005  int8x16_t __ret;
66006  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66007  return __ret;
66008}
66009#else
66010__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66011  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66012  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66013  int8x16_t __ret;
66014  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66015  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66016  return __ret;
66017}
66018#endif
66019
66020#ifdef __LITTLE_ENDIAN__
66021__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66022  float64x2_t __ret;
66023  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66024  return __ret;
66025}
66026#else
66027__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66028  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66029  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66030  float64x2_t __ret;
66031  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66032  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66033  return __ret;
66034}
66035#endif
66036
66037#ifdef __LITTLE_ENDIAN__
66038__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66039  float32x4_t __ret;
66040  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66041  return __ret;
66042}
66043#else
66044__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66045  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66046  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66047  float32x4_t __ret;
66048  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66049  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66050  return __ret;
66051}
66052#endif
66053
66054#ifdef __LITTLE_ENDIAN__
66055__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66056  int32x4_t __ret;
66057  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66058  return __ret;
66059}
66060#else
66061__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66062  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66063  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66064  int32x4_t __ret;
66065  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66066  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66067  return __ret;
66068}
66069#endif
66070
66071#ifdef __LITTLE_ENDIAN__
66072__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66073  int64x2_t __ret;
66074  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66075  return __ret;
66076}
66077#else
66078__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66079  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66080  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66081  int64x2_t __ret;
66082  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66083  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66084  return __ret;
66085}
66086#endif
66087
66088#ifdef __LITTLE_ENDIAN__
66089__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66090  int16x8_t __ret;
66091  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66092  return __ret;
66093}
66094#else
66095__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66096  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66097  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66098  int16x8_t __ret;
66099  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66100  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66101  return __ret;
66102}
66103#endif
66104
66105#ifdef __LITTLE_ENDIAN__
66106__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66107  uint8x8_t __ret;
66108  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66109  return __ret;
66110}
66111#else
66112__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66113  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66114  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66115  uint8x8_t __ret;
66116  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66117  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66118  return __ret;
66119}
66120#endif
66121
66122#ifdef __LITTLE_ENDIAN__
66123__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66124  uint32x2_t __ret;
66125  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66126  return __ret;
66127}
66128#else
66129__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66130  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66131  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66132  uint32x2_t __ret;
66133  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66134  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66135  return __ret;
66136}
66137#endif
66138
66139#ifdef __LITTLE_ENDIAN__
66140__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66141  uint16x4_t __ret;
66142  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66143  return __ret;
66144}
66145#else
66146__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66147  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66148  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66149  uint16x4_t __ret;
66150  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66151  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66152  return __ret;
66153}
66154#endif
66155
66156#ifdef __LITTLE_ENDIAN__
66157__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66158  int8x8_t __ret;
66159  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66160  return __ret;
66161}
66162#else
66163__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66164  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66165  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66166  int8x8_t __ret;
66167  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66168  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66169  return __ret;
66170}
66171#endif
66172
66173#ifdef __LITTLE_ENDIAN__
66174__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66175  float32x2_t __ret;
66176  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66177  return __ret;
66178}
66179#else
66180__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66181  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66182  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66183  float32x2_t __ret;
66184  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66185  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66186  return __ret;
66187}
66188#endif
66189
66190#ifdef __LITTLE_ENDIAN__
66191__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66192  int32x2_t __ret;
66193  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66194  return __ret;
66195}
66196#else
66197__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66198  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66199  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66200  int32x2_t __ret;
66201  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66203  return __ret;
66204}
66205#endif
66206
66207#ifdef __LITTLE_ENDIAN__
66208__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66209  int16x4_t __ret;
66210  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66211  return __ret;
66212}
66213#else
66214__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66215  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66216  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66217  int16x4_t __ret;
66218  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66220  return __ret;
66221}
66222#endif
66223
66224#ifdef __LITTLE_ENDIAN__
66225__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66226  poly8x8_t __ret;
66227  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66228  return __ret;
66229}
66230#else
66231__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66232  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66233  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66234  poly8x8_t __ret;
66235  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66236  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66237  return __ret;
66238}
66239#endif
66240
66241#ifdef __LITTLE_ENDIAN__
66242__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66243  poly16x4_t __ret;
66244  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66245  return __ret;
66246}
66247#else
66248__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66249  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66250  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66251  poly16x4_t __ret;
66252  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66253  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66254  return __ret;
66255}
66256#endif
66257
66258#ifdef __LITTLE_ENDIAN__
66259__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66260  poly8x16_t __ret;
66261  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66262  return __ret;
66263}
66264#else
66265__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66266  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66267  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66268  poly8x16_t __ret;
66269  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66270  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66271  return __ret;
66272}
66273#endif
66274
66275#ifdef __LITTLE_ENDIAN__
66276__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66277  poly64x2_t __ret;
66278  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66279  return __ret;
66280}
66281#else
66282__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66283  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66284  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66285  poly64x2_t __ret;
66286  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66287  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66288  return __ret;
66289}
66290#endif
66291
66292#ifdef __LITTLE_ENDIAN__
66293__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66294  poly16x8_t __ret;
66295  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66296  return __ret;
66297}
66298#else
66299__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66300  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66301  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66302  poly16x8_t __ret;
66303  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66305  return __ret;
66306}
66307#endif
66308
66309#ifdef __LITTLE_ENDIAN__
66310__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66311  uint8x16_t __ret;
66312  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66313  return __ret;
66314}
66315#else
66316__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66317  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66318  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66319  uint8x16_t __ret;
66320  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66321  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66322  return __ret;
66323}
66324#endif
66325
66326#ifdef __LITTLE_ENDIAN__
66327__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66328  uint32x4_t __ret;
66329  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66330  return __ret;
66331}
66332#else
66333__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66334  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66335  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66336  uint32x4_t __ret;
66337  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66338  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66339  return __ret;
66340}
66341#endif
66342
66343#ifdef __LITTLE_ENDIAN__
66344__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66345  uint64x2_t __ret;
66346  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66347  return __ret;
66348}
66349#else
66350__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66351  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66352  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66353  uint64x2_t __ret;
66354  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66356  return __ret;
66357}
66358#endif
66359
66360#ifdef __LITTLE_ENDIAN__
66361__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66362  uint16x8_t __ret;
66363  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66364  return __ret;
66365}
66366#else
66367__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66368  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66369  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66370  uint16x8_t __ret;
66371  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66372  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66373  return __ret;
66374}
66375#endif
66376
66377#ifdef __LITTLE_ENDIAN__
66378__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66379  int8x16_t __ret;
66380  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66381  return __ret;
66382}
66383#else
66384__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66385  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66386  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66387  int8x16_t __ret;
66388  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66389  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66390  return __ret;
66391}
66392#endif
66393
66394#ifdef __LITTLE_ENDIAN__
66395__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66396  float64x2_t __ret;
66397  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66398  return __ret;
66399}
66400#else
66401__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66402  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66403  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66404  float64x2_t __ret;
66405  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66407  return __ret;
66408}
66409#endif
66410
66411#ifdef __LITTLE_ENDIAN__
66412__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66413  float32x4_t __ret;
66414  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66415  return __ret;
66416}
66417#else
66418__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66419  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66420  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66421  float32x4_t __ret;
66422  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66424  return __ret;
66425}
66426#endif
66427
66428#ifdef __LITTLE_ENDIAN__
66429__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66430  int32x4_t __ret;
66431  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66432  return __ret;
66433}
66434#else
66435__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66436  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66437  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66438  int32x4_t __ret;
66439  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66441  return __ret;
66442}
66443#endif
66444
66445#ifdef __LITTLE_ENDIAN__
66446__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66447  int64x2_t __ret;
66448  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66449  return __ret;
66450}
66451#else
66452__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66453  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66454  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66455  int64x2_t __ret;
66456  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66457  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66458  return __ret;
66459}
66460#endif
66461
66462#ifdef __LITTLE_ENDIAN__
66463__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66464  int16x8_t __ret;
66465  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66466  return __ret;
66467}
66468#else
66469__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66470  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66471  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66472  int16x8_t __ret;
66473  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66474  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66475  return __ret;
66476}
66477#endif
66478
66479#ifdef __LITTLE_ENDIAN__
66480__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66481  uint8x8_t __ret;
66482  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66483  return __ret;
66484}
66485#else
66486__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66487  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66488  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66489  uint8x8_t __ret;
66490  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66492  return __ret;
66493}
66494#endif
66495
66496#ifdef __LITTLE_ENDIAN__
66497__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66498  uint32x2_t __ret;
66499  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66500  return __ret;
66501}
66502#else
66503__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66504  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66505  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66506  uint32x2_t __ret;
66507  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66508  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66509  return __ret;
66510}
66511#endif
66512
66513#ifdef __LITTLE_ENDIAN__
66514__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66515  uint16x4_t __ret;
66516  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66517  return __ret;
66518}
66519#else
66520__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66521  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66522  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66523  uint16x4_t __ret;
66524  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66526  return __ret;
66527}
66528#endif
66529
66530#ifdef __LITTLE_ENDIAN__
66531__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66532  int8x8_t __ret;
66533  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66534  return __ret;
66535}
66536#else
66537__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66538  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66539  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66540  int8x8_t __ret;
66541  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66543  return __ret;
66544}
66545#endif
66546
66547#ifdef __LITTLE_ENDIAN__
66548__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66549  float32x2_t __ret;
66550  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66551  return __ret;
66552}
66553#else
66554__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66555  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66556  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66557  float32x2_t __ret;
66558  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66559  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66560  return __ret;
66561}
66562#endif
66563
66564#ifdef __LITTLE_ENDIAN__
66565__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66566  int32x2_t __ret;
66567  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66568  return __ret;
66569}
66570#else
66571__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66572  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66573  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66574  int32x2_t __ret;
66575  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66577  return __ret;
66578}
66579#endif
66580
66581#ifdef __LITTLE_ENDIAN__
66582__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66583  int16x4_t __ret;
66584  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66585  return __ret;
66586}
66587#else
66588__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66589  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66590  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66591  int16x4_t __ret;
66592  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66594  return __ret;
66595}
66596#endif
66597
66598#endif
66599#ifdef __LITTLE_ENDIAN__
66600__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66601  uint8x16_t __ret;
66602  __ret = __p0 + vabdq_u8(__p1, __p2);
66603  return __ret;
66604}
66605#else
66606__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66607  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66608  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66609  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66610  uint8x16_t __ret;
66611  __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
66612  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66613  return __ret;
66614}
66615#endif
66616
66617#ifdef __LITTLE_ENDIAN__
66618__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66619  uint32x4_t __ret;
66620  __ret = __p0 + vabdq_u32(__p1, __p2);
66621  return __ret;
66622}
66623#else
66624__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66625  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66626  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66627  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66628  uint32x4_t __ret;
66629  __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
66630  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66631  return __ret;
66632}
66633#endif
66634
66635#ifdef __LITTLE_ENDIAN__
66636__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66637  uint16x8_t __ret;
66638  __ret = __p0 + vabdq_u16(__p1, __p2);
66639  return __ret;
66640}
66641#else
66642__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66643  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66644  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66645  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66646  uint16x8_t __ret;
66647  __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
66648  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66649  return __ret;
66650}
66651#endif
66652
66653#ifdef __LITTLE_ENDIAN__
66654__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66655  int8x16_t __ret;
66656  __ret = __p0 + vabdq_s8(__p1, __p2);
66657  return __ret;
66658}
66659#else
66660__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66661  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66662  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66663  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66664  int8x16_t __ret;
66665  __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
66666  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66667  return __ret;
66668}
66669#endif
66670
66671#ifdef __LITTLE_ENDIAN__
66672__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66673  int32x4_t __ret;
66674  __ret = __p0 + vabdq_s32(__p1, __p2);
66675  return __ret;
66676}
66677#else
66678__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66679  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66680  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66681  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66682  int32x4_t __ret;
66683  __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
66684  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66685  return __ret;
66686}
66687#endif
66688
66689#ifdef __LITTLE_ENDIAN__
66690__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66691  int16x8_t __ret;
66692  __ret = __p0 + vabdq_s16(__p1, __p2);
66693  return __ret;
66694}
66695#else
66696__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66697  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66698  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66699  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66700  int16x8_t __ret;
66701  __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
66702  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66703  return __ret;
66704}
66705#endif
66706
66707#ifdef __LITTLE_ENDIAN__
66708__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66709  uint8x8_t __ret;
66710  __ret = __p0 + vabd_u8(__p1, __p2);
66711  return __ret;
66712}
66713#else
66714__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66715  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66716  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66717  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66718  uint8x8_t __ret;
66719  __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
66720  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66721  return __ret;
66722}
66723#endif
66724
66725#ifdef __LITTLE_ENDIAN__
66726__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66727  uint32x2_t __ret;
66728  __ret = __p0 + vabd_u32(__p1, __p2);
66729  return __ret;
66730}
66731#else
66732__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66733  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66734  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66735  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66736  uint32x2_t __ret;
66737  __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
66738  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66739  return __ret;
66740}
66741#endif
66742
66743#ifdef __LITTLE_ENDIAN__
66744__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66745  uint16x4_t __ret;
66746  __ret = __p0 + vabd_u16(__p1, __p2);
66747  return __ret;
66748}
66749#else
66750__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66751  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66752  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66753  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66754  uint16x4_t __ret;
66755  __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
66756  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66757  return __ret;
66758}
66759#endif
66760
66761#ifdef __LITTLE_ENDIAN__
66762__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66763  int8x8_t __ret;
66764  __ret = __p0 + vabd_s8(__p1, __p2);
66765  return __ret;
66766}
66767#else
66768__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66769  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66770  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66771  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66772  int8x8_t __ret;
66773  __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
66774  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66775  return __ret;
66776}
66777#endif
66778
66779#ifdef __LITTLE_ENDIAN__
66780__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66781  int32x2_t __ret;
66782  __ret = __p0 + vabd_s32(__p1, __p2);
66783  return __ret;
66784}
66785#else
66786__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66787  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66788  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66789  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66790  int32x2_t __ret;
66791  __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
66792  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66793  return __ret;
66794}
66795#endif
66796
66797#ifdef __LITTLE_ENDIAN__
66798__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66799  int16x4_t __ret;
66800  __ret = __p0 + vabd_s16(__p1, __p2);
66801  return __ret;
66802}
66803#else
66804__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66805  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66806  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66807  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66808  int16x4_t __ret;
66809  __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
66810  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66811  return __ret;
66812}
66813#endif
66814
66815#ifdef __LITTLE_ENDIAN__
66816__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66817  uint16x8_t __ret;
66818  __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
66819  return __ret;
66820}
66821#else
66822__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66823  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66824  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66825  uint16x8_t __ret;
66826  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
66827  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66828  return __ret;
66829}
66830__ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66831  uint16x8_t __ret;
66832  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
66833  return __ret;
66834}
66835#endif
66836
66837#ifdef __LITTLE_ENDIAN__
66838__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66839  uint64x2_t __ret;
66840  __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
66841  return __ret;
66842}
66843#else
66844__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66845  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66846  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66847  uint64x2_t __ret;
66848  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
66849  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66850  return __ret;
66851}
66852__ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66853  uint64x2_t __ret;
66854  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
66855  return __ret;
66856}
66857#endif
66858
66859#ifdef __LITTLE_ENDIAN__
66860__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66861  uint32x4_t __ret;
66862  __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
66863  return __ret;
66864}
66865#else
66866__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66867  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66868  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66869  uint32x4_t __ret;
66870  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
66871  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66872  return __ret;
66873}
66874__ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66875  uint32x4_t __ret;
66876  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
66877  return __ret;
66878}
66879#endif
66880
66881#ifdef __LITTLE_ENDIAN__
66882__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66883  int16x8_t __ret;
66884  __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
66885  return __ret;
66886}
66887#else
66888__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66889  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66890  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66891  int16x8_t __ret;
66892  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
66893  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66894  return __ret;
66895}
66896__ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66897  int16x8_t __ret;
66898  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
66899  return __ret;
66900}
66901#endif
66902
66903#ifdef __LITTLE_ENDIAN__
66904__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66905  int64x2_t __ret;
66906  __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
66907  return __ret;
66908}
66909#else
66910__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66911  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66912  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66913  int64x2_t __ret;
66914  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
66915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66916  return __ret;
66917}
66918__ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66919  int64x2_t __ret;
66920  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
66921  return __ret;
66922}
66923#endif
66924
66925#ifdef __LITTLE_ENDIAN__
66926__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66927  int32x4_t __ret;
66928  __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
66929  return __ret;
66930}
66931#else
66932__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66933  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66934  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66935  int32x4_t __ret;
66936  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
66937  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66938  return __ret;
66939}
66940__ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66941  int32x4_t __ret;
66942  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
66943  return __ret;
66944}
66945#endif
66946
66947#ifdef __LITTLE_ENDIAN__
66948__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66949  uint16x8_t __ret;
66950  __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
66951  return __ret;
66952}
66953#else
66954__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66955  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66956  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66957  uint16x8_t __ret;
66958  __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
66959  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66960  return __ret;
66961}
66962#endif
66963
66964#ifdef __LITTLE_ENDIAN__
66965__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66966  uint64x2_t __ret;
66967  __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
66968  return __ret;
66969}
66970#else
66971__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66972  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66973  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66974  uint64x2_t __ret;
66975  __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
66976  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66977  return __ret;
66978}
66979#endif
66980
66981#ifdef __LITTLE_ENDIAN__
66982__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66983  uint32x4_t __ret;
66984  __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
66985  return __ret;
66986}
66987#else
66988__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66989  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66990  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66991  uint32x4_t __ret;
66992  __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
66993  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66994  return __ret;
66995}
66996#endif
66997
66998#ifdef __LITTLE_ENDIAN__
66999__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67000  int16x8_t __ret;
67001  __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
67002  return __ret;
67003}
67004#else
67005__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67006  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67007  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67008  int16x8_t __ret;
67009  __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
67010  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67011  return __ret;
67012}
67013#endif
67014
67015#ifdef __LITTLE_ENDIAN__
67016__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67017  int64x2_t __ret;
67018  __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
67019  return __ret;
67020}
67021#else
67022__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67023  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67024  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67025  int64x2_t __ret;
67026  __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
67027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67028  return __ret;
67029}
67030#endif
67031
67032#ifdef __LITTLE_ENDIAN__
67033__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67034  int32x4_t __ret;
67035  __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
67036  return __ret;
67037}
67038#else
67039__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67040  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67041  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67042  int32x4_t __ret;
67043  __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
67044  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67045  return __ret;
67046}
67047#endif
67048
67049#ifdef __LITTLE_ENDIAN__
67050__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67051  uint16x8_t __ret;
67052  __ret = __p0 + vmovl_u8(__p1);
67053  return __ret;
67054}
67055#else
67056__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67057  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67058  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67059  uint16x8_t __ret;
67060  __ret = __rev0 + __noswap_vmovl_u8(__rev1);
67061  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67062  return __ret;
67063}
67064#endif
67065
67066#ifdef __LITTLE_ENDIAN__
67067__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67068  uint64x2_t __ret;
67069  __ret = __p0 + vmovl_u32(__p1);
67070  return __ret;
67071}
67072#else
67073__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67074  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67075  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67076  uint64x2_t __ret;
67077  __ret = __rev0 + __noswap_vmovl_u32(__rev1);
67078  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67079  return __ret;
67080}
67081#endif
67082
67083#ifdef __LITTLE_ENDIAN__
67084__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67085  uint32x4_t __ret;
67086  __ret = __p0 + vmovl_u16(__p1);
67087  return __ret;
67088}
67089#else
67090__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67091  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67092  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67093  uint32x4_t __ret;
67094  __ret = __rev0 + __noswap_vmovl_u16(__rev1);
67095  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67096  return __ret;
67097}
67098#endif
67099
67100#ifdef __LITTLE_ENDIAN__
67101__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67102  int16x8_t __ret;
67103  __ret = __p0 + vmovl_s8(__p1);
67104  return __ret;
67105}
67106#else
67107__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67108  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67109  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67110  int16x8_t __ret;
67111  __ret = __rev0 + __noswap_vmovl_s8(__rev1);
67112  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67113  return __ret;
67114}
67115#endif
67116
67117#ifdef __LITTLE_ENDIAN__
67118__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67119  int64x2_t __ret;
67120  __ret = __p0 + vmovl_s32(__p1);
67121  return __ret;
67122}
67123#else
67124__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67125  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67126  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67127  int64x2_t __ret;
67128  __ret = __rev0 + __noswap_vmovl_s32(__rev1);
67129  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67130  return __ret;
67131}
67132#endif
67133
67134#ifdef __LITTLE_ENDIAN__
67135__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67136  int32x4_t __ret;
67137  __ret = __p0 + vmovl_s16(__p1);
67138  return __ret;
67139}
67140#else
67141__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67142  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67143  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67144  int32x4_t __ret;
67145  __ret = __rev0 + __noswap_vmovl_s16(__rev1);
67146  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67147  return __ret;
67148}
67149#endif
67150
67151#ifdef __LITTLE_ENDIAN__
67152#define vget_lane_f16(__p0_242, __p1_242) __extension__ ({ \
67153  float16x4_t __s0_242 = __p0_242; \
67154  float16_t __ret_242; \
67155float16x4_t __reint_242 = __s0_242; \
67156int16_t __reint1_242 = vget_lane_s16(*(int16x4_t *) &__reint_242, __p1_242); \
67157  __ret_242 = *(float16_t *) &__reint1_242; \
67158  __ret_242; \
67159})
67160#else
67161#define vget_lane_f16(__p0_243, __p1_243) __extension__ ({ \
67162  float16x4_t __s0_243 = __p0_243; \
67163  float16x4_t __rev0_243;  __rev0_243 = __builtin_shufflevector(__s0_243, __s0_243, 3, 2, 1, 0); \
67164  float16_t __ret_243; \
67165float16x4_t __reint_243 = __rev0_243; \
67166int16_t __reint1_243 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_243, __p1_243); \
67167  __ret_243 = *(float16_t *) &__reint1_243; \
67168  __ret_243; \
67169})
67170#endif
67171
67172#ifdef __LITTLE_ENDIAN__
67173#define vgetq_lane_f16(__p0_244, __p1_244) __extension__ ({ \
67174  float16x8_t __s0_244 = __p0_244; \
67175  float16_t __ret_244; \
67176float16x8_t __reint_244 = __s0_244; \
67177int16_t __reint1_244 = vgetq_lane_s16(*(int16x8_t *) &__reint_244, __p1_244); \
67178  __ret_244 = *(float16_t *) &__reint1_244; \
67179  __ret_244; \
67180})
67181#else
67182#define vgetq_lane_f16(__p0_245, __p1_245) __extension__ ({ \
67183  float16x8_t __s0_245 = __p0_245; \
67184  float16x8_t __rev0_245;  __rev0_245 = __builtin_shufflevector(__s0_245, __s0_245, 7, 6, 5, 4, 3, 2, 1, 0); \
67185  float16_t __ret_245; \
67186float16x8_t __reint_245 = __rev0_245; \
67187int16_t __reint1_245 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_245, __p1_245); \
67188  __ret_245 = *(float16_t *) &__reint1_245; \
67189  __ret_245; \
67190})
67191#endif
67192
67193#ifdef __LITTLE_ENDIAN__
67194__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67195  uint16x8_t __ret;
67196  __ret = __p0 + vmull_u8(__p1, __p2);
67197  return __ret;
67198}
67199#else
67200__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67201  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67202  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67203  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67204  uint16x8_t __ret;
67205  __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
67206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67207  return __ret;
67208}
67209__ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67210  uint16x8_t __ret;
67211  __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
67212  return __ret;
67213}
67214#endif
67215
67216#ifdef __LITTLE_ENDIAN__
67217__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67218  uint64x2_t __ret;
67219  __ret = __p0 + vmull_u32(__p1, __p2);
67220  return __ret;
67221}
67222#else
67223__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67224  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67225  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67226  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67227  uint64x2_t __ret;
67228  __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
67229  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67230  return __ret;
67231}
67232__ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67233  uint64x2_t __ret;
67234  __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
67235  return __ret;
67236}
67237#endif
67238
67239#ifdef __LITTLE_ENDIAN__
67240__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67241  uint32x4_t __ret;
67242  __ret = __p0 + vmull_u16(__p1, __p2);
67243  return __ret;
67244}
67245#else
67246__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67247  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67248  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67249  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67250  uint32x4_t __ret;
67251  __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
67252  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67253  return __ret;
67254}
67255__ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67256  uint32x4_t __ret;
67257  __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
67258  return __ret;
67259}
67260#endif
67261
67262#ifdef __LITTLE_ENDIAN__
67263__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67264  int16x8_t __ret;
67265  __ret = __p0 + vmull_s8(__p1, __p2);
67266  return __ret;
67267}
67268#else
67269__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67270  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67271  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67272  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67273  int16x8_t __ret;
67274  __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
67275  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67276  return __ret;
67277}
67278__ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67279  int16x8_t __ret;
67280  __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
67281  return __ret;
67282}
67283#endif
67284
67285#ifdef __LITTLE_ENDIAN__
67286__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67287  int64x2_t __ret;
67288  __ret = __p0 + vmull_s32(__p1, __p2);
67289  return __ret;
67290}
67291#else
67292__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67293  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67294  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67295  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67296  int64x2_t __ret;
67297  __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
67298  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67299  return __ret;
67300}
67301__ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67302  int64x2_t __ret;
67303  __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
67304  return __ret;
67305}
67306#endif
67307
67308#ifdef __LITTLE_ENDIAN__
67309__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67310  int32x4_t __ret;
67311  __ret = __p0 + vmull_s16(__p1, __p2);
67312  return __ret;
67313}
67314#else
67315__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67316  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67317  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67318  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67319  int32x4_t __ret;
67320  __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
67321  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67322  return __ret;
67323}
67324__ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67325  int32x4_t __ret;
67326  __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
67327  return __ret;
67328}
67329#endif
67330
67331#ifdef __LITTLE_ENDIAN__
67332#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67333  uint64x2_t __s0 = __p0; \
67334  uint32x2_t __s1 = __p1; \
67335  uint32x2_t __s2 = __p2; \
67336  uint64x2_t __ret; \
67337  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67338  __ret; \
67339})
67340#else
67341#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67342  uint64x2_t __s0 = __p0; \
67343  uint32x2_t __s1 = __p1; \
67344  uint32x2_t __s2 = __p2; \
67345  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67346  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67347  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67348  uint64x2_t __ret; \
67349  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67350  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67351  __ret; \
67352})
67353#endif
67354
67355#ifdef __LITTLE_ENDIAN__
67356#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67357  uint32x4_t __s0 = __p0; \
67358  uint16x4_t __s1 = __p1; \
67359  uint16x4_t __s2 = __p2; \
67360  uint32x4_t __ret; \
67361  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67362  __ret; \
67363})
67364#else
67365#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67366  uint32x4_t __s0 = __p0; \
67367  uint16x4_t __s1 = __p1; \
67368  uint16x4_t __s2 = __p2; \
67369  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67370  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67371  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67372  uint32x4_t __ret; \
67373  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67374  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67375  __ret; \
67376})
67377#endif
67378
67379#ifdef __LITTLE_ENDIAN__
67380#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67381  int64x2_t __s0 = __p0; \
67382  int32x2_t __s1 = __p1; \
67383  int32x2_t __s2 = __p2; \
67384  int64x2_t __ret; \
67385  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67386  __ret; \
67387})
67388#else
67389#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67390  int64x2_t __s0 = __p0; \
67391  int32x2_t __s1 = __p1; \
67392  int32x2_t __s2 = __p2; \
67393  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67394  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67395  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67396  int64x2_t __ret; \
67397  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67399  __ret; \
67400})
67401#endif
67402
67403#ifdef __LITTLE_ENDIAN__
67404#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67405  int32x4_t __s0 = __p0; \
67406  int16x4_t __s1 = __p1; \
67407  int16x4_t __s2 = __p2; \
67408  int32x4_t __ret; \
67409  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67410  __ret; \
67411})
67412#else
67413#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67414  int32x4_t __s0 = __p0; \
67415  int16x4_t __s1 = __p1; \
67416  int16x4_t __s2 = __p2; \
67417  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67418  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67419  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67420  int32x4_t __ret; \
67421  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67422  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67423  __ret; \
67424})
67425#endif
67426
67427#ifdef __LITTLE_ENDIAN__
67428__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67429  uint64x2_t __ret;
67430  __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67431  return __ret;
67432}
67433#else
67434__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67435  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67436  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67437  uint64x2_t __ret;
67438  __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67439  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67440  return __ret;
67441}
67442__ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67443  uint64x2_t __ret;
67444  __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67445  return __ret;
67446}
67447#endif
67448
67449#ifdef __LITTLE_ENDIAN__
67450__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67451  uint32x4_t __ret;
67452  __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67453  return __ret;
67454}
67455#else
67456__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67457  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67458  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67459  uint32x4_t __ret;
67460  __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67461  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67462  return __ret;
67463}
67464__ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67465  uint32x4_t __ret;
67466  __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67467  return __ret;
67468}
67469#endif
67470
67471#ifdef __LITTLE_ENDIAN__
67472__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67473  int64x2_t __ret;
67474  __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67475  return __ret;
67476}
67477#else
67478__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67479  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67480  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67481  int64x2_t __ret;
67482  __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67483  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67484  return __ret;
67485}
67486__ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67487  int64x2_t __ret;
67488  __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67489  return __ret;
67490}
67491#endif
67492
67493#ifdef __LITTLE_ENDIAN__
67494__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67495  int32x4_t __ret;
67496  __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67497  return __ret;
67498}
67499#else
67500__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67501  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67502  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67503  int32x4_t __ret;
67504  __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67505  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67506  return __ret;
67507}
67508__ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67509  int32x4_t __ret;
67510  __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67511  return __ret;
67512}
67513#endif
67514
67515#ifdef __LITTLE_ENDIAN__
67516__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67517  uint16x8_t __ret;
67518  __ret = __p0 - vmull_u8(__p1, __p2);
67519  return __ret;
67520}
67521#else
67522__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67523  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67524  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67525  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67526  uint16x8_t __ret;
67527  __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
67528  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67529  return __ret;
67530}
67531__ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67532  uint16x8_t __ret;
67533  __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
67534  return __ret;
67535}
67536#endif
67537
67538#ifdef __LITTLE_ENDIAN__
67539__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67540  uint64x2_t __ret;
67541  __ret = __p0 - vmull_u32(__p1, __p2);
67542  return __ret;
67543}
67544#else
67545__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67546  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67547  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67548  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67549  uint64x2_t __ret;
67550  __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
67551  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67552  return __ret;
67553}
67554__ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67555  uint64x2_t __ret;
67556  __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
67557  return __ret;
67558}
67559#endif
67560
67561#ifdef __LITTLE_ENDIAN__
67562__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67563  uint32x4_t __ret;
67564  __ret = __p0 - vmull_u16(__p1, __p2);
67565  return __ret;
67566}
67567#else
67568__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67569  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67570  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67571  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67572  uint32x4_t __ret;
67573  __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
67574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67575  return __ret;
67576}
67577__ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67578  uint32x4_t __ret;
67579  __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
67580  return __ret;
67581}
67582#endif
67583
67584#ifdef __LITTLE_ENDIAN__
67585__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67586  int16x8_t __ret;
67587  __ret = __p0 - vmull_s8(__p1, __p2);
67588  return __ret;
67589}
67590#else
67591__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67592  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67593  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67594  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67595  int16x8_t __ret;
67596  __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
67597  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67598  return __ret;
67599}
67600__ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67601  int16x8_t __ret;
67602  __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
67603  return __ret;
67604}
67605#endif
67606
67607#ifdef __LITTLE_ENDIAN__
67608__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67609  int64x2_t __ret;
67610  __ret = __p0 - vmull_s32(__p1, __p2);
67611  return __ret;
67612}
67613#else
67614__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67615  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67616  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67617  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67618  int64x2_t __ret;
67619  __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
67620  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67621  return __ret;
67622}
67623__ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67624  int64x2_t __ret;
67625  __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
67626  return __ret;
67627}
67628#endif
67629
67630#ifdef __LITTLE_ENDIAN__
67631__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67632  int32x4_t __ret;
67633  __ret = __p0 - vmull_s16(__p1, __p2);
67634  return __ret;
67635}
67636#else
67637__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67638  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67639  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67640  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67641  int32x4_t __ret;
67642  __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
67643  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67644  return __ret;
67645}
67646__ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67647  int32x4_t __ret;
67648  __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
67649  return __ret;
67650}
67651#endif
67652
67653#ifdef __LITTLE_ENDIAN__
67654#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67655  uint64x2_t __s0 = __p0; \
67656  uint32x2_t __s1 = __p1; \
67657  uint32x2_t __s2 = __p2; \
67658  uint64x2_t __ret; \
67659  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67660  __ret; \
67661})
67662#else
67663#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67664  uint64x2_t __s0 = __p0; \
67665  uint32x2_t __s1 = __p1; \
67666  uint32x2_t __s2 = __p2; \
67667  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67668  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67669  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67670  uint64x2_t __ret; \
67671  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67672  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67673  __ret; \
67674})
67675#endif
67676
67677#ifdef __LITTLE_ENDIAN__
67678#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67679  uint32x4_t __s0 = __p0; \
67680  uint16x4_t __s1 = __p1; \
67681  uint16x4_t __s2 = __p2; \
67682  uint32x4_t __ret; \
67683  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67684  __ret; \
67685})
67686#else
67687#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67688  uint32x4_t __s0 = __p0; \
67689  uint16x4_t __s1 = __p1; \
67690  uint16x4_t __s2 = __p2; \
67691  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67692  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67693  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67694  uint32x4_t __ret; \
67695  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67696  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67697  __ret; \
67698})
67699#endif
67700
67701#ifdef __LITTLE_ENDIAN__
67702#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67703  int64x2_t __s0 = __p0; \
67704  int32x2_t __s1 = __p1; \
67705  int32x2_t __s2 = __p2; \
67706  int64x2_t __ret; \
67707  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67708  __ret; \
67709})
67710#else
67711#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67712  int64x2_t __s0 = __p0; \
67713  int32x2_t __s1 = __p1; \
67714  int32x2_t __s2 = __p2; \
67715  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67716  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67717  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67718  int64x2_t __ret; \
67719  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67720  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67721  __ret; \
67722})
67723#endif
67724
67725#ifdef __LITTLE_ENDIAN__
67726#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67727  int32x4_t __s0 = __p0; \
67728  int16x4_t __s1 = __p1; \
67729  int16x4_t __s2 = __p2; \
67730  int32x4_t __ret; \
67731  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67732  __ret; \
67733})
67734#else
67735#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67736  int32x4_t __s0 = __p0; \
67737  int16x4_t __s1 = __p1; \
67738  int16x4_t __s2 = __p2; \
67739  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67740  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67741  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67742  int32x4_t __ret; \
67743  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67744  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67745  __ret; \
67746})
67747#endif
67748
67749#ifdef __LITTLE_ENDIAN__
67750__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67751  uint64x2_t __ret;
67752  __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67753  return __ret;
67754}
67755#else
67756__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67757  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67758  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67759  uint64x2_t __ret;
67760  __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67761  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67762  return __ret;
67763}
67764__ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67765  uint64x2_t __ret;
67766  __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67767  return __ret;
67768}
67769#endif
67770
67771#ifdef __LITTLE_ENDIAN__
67772__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67773  uint32x4_t __ret;
67774  __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67775  return __ret;
67776}
67777#else
67778__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67779  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67781  uint32x4_t __ret;
67782  __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67784  return __ret;
67785}
67786__ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67787  uint32x4_t __ret;
67788  __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67789  return __ret;
67790}
67791#endif
67792
67793#ifdef __LITTLE_ENDIAN__
67794__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67795  int64x2_t __ret;
67796  __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67797  return __ret;
67798}
67799#else
67800__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67801  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67802  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67803  int64x2_t __ret;
67804  __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67806  return __ret;
67807}
67808__ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67809  int64x2_t __ret;
67810  __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67811  return __ret;
67812}
67813#endif
67814
67815#ifdef __LITTLE_ENDIAN__
67816__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67817  int32x4_t __ret;
67818  __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67819  return __ret;
67820}
67821#else
67822__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67823  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67824  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67825  int32x4_t __ret;
67826  __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67827  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67828  return __ret;
67829}
67830__ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67831  int32x4_t __ret;
67832  __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67833  return __ret;
67834}
67835#endif
67836
67837#ifdef __LITTLE_ENDIAN__
67838#define vset_lane_f16(__p0_246, __p1_246, __p2_246) __extension__ ({ \
67839  float16_t __s0_246 = __p0_246; \
67840  float16x4_t __s1_246 = __p1_246; \
67841  float16x4_t __ret_246; \
67842float16_t __reint_246 = __s0_246; \
67843float16x4_t __reint1_246 = __s1_246; \
67844int16x4_t __reint2_246 = vset_lane_s16(*(int16_t *) &__reint_246, *(int16x4_t *) &__reint1_246, __p2_246); \
67845  __ret_246 = *(float16x4_t *) &__reint2_246; \
67846  __ret_246; \
67847})
67848#else
67849#define vset_lane_f16(__p0_247, __p1_247, __p2_247) __extension__ ({ \
67850  float16_t __s0_247 = __p0_247; \
67851  float16x4_t __s1_247 = __p1_247; \
67852  float16x4_t __rev1_247;  __rev1_247 = __builtin_shufflevector(__s1_247, __s1_247, 3, 2, 1, 0); \
67853  float16x4_t __ret_247; \
67854float16_t __reint_247 = __s0_247; \
67855float16x4_t __reint1_247 = __rev1_247; \
67856int16x4_t __reint2_247 = __noswap_vset_lane_s16(*(int16_t *) &__reint_247, *(int16x4_t *) &__reint1_247, __p2_247); \
67857  __ret_247 = *(float16x4_t *) &__reint2_247; \
67858  __ret_247 = __builtin_shufflevector(__ret_247, __ret_247, 3, 2, 1, 0); \
67859  __ret_247; \
67860})
67861#endif
67862
67863#ifdef __LITTLE_ENDIAN__
67864#define vsetq_lane_f16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
67865  float16_t __s0_248 = __p0_248; \
67866  float16x8_t __s1_248 = __p1_248; \
67867  float16x8_t __ret_248; \
67868float16_t __reint_248 = __s0_248; \
67869float16x8_t __reint1_248 = __s1_248; \
67870int16x8_t __reint2_248 = vsetq_lane_s16(*(int16_t *) &__reint_248, *(int16x8_t *) &__reint1_248, __p2_248); \
67871  __ret_248 = *(float16x8_t *) &__reint2_248; \
67872  __ret_248; \
67873})
67874#else
67875#define vsetq_lane_f16(__p0_249, __p1_249, __p2_249) __extension__ ({ \
67876  float16_t __s0_249 = __p0_249; \
67877  float16x8_t __s1_249 = __p1_249; \
67878  float16x8_t __rev1_249;  __rev1_249 = __builtin_shufflevector(__s1_249, __s1_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67879  float16x8_t __ret_249; \
67880float16_t __reint_249 = __s0_249; \
67881float16x8_t __reint1_249 = __rev1_249; \
67882int16x8_t __reint2_249 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_249, *(int16x8_t *) &__reint1_249, __p2_249); \
67883  __ret_249 = *(float16x8_t *) &__reint2_249; \
67884  __ret_249 = __builtin_shufflevector(__ret_249, __ret_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67885  __ret_249; \
67886})
67887#endif
67888
67889#if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
67890#ifdef __LITTLE_ENDIAN__
67891__ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67892  int32_t __ret;
67893  __ret = vqadds_s32(__p0, vqrdmulhs_s32(__p1, __p2));
67894  return __ret;
67895}
67896#else
67897__ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67898  int32_t __ret;
67899  __ret = __noswap_vqadds_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
67900  return __ret;
67901}
67902#endif
67903
67904#ifdef __LITTLE_ENDIAN__
67905__ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67906  int16_t __ret;
67907  __ret = vqaddh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
67908  return __ret;
67909}
67910#else
67911__ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67912  int16_t __ret;
67913  __ret = __noswap_vqaddh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
67914  return __ret;
67915}
67916#endif
67917
67918#ifdef __LITTLE_ENDIAN__
67919#define vqrdmlahs_lane_s32(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
67920  int32_t __s0_250 = __p0_250; \
67921  int32_t __s1_250 = __p1_250; \
67922  int32x2_t __s2_250 = __p2_250; \
67923  int32_t __ret_250; \
67924  __ret_250 = vqadds_s32(__s0_250, vqrdmulhs_s32(__s1_250, vget_lane_s32(__s2_250, __p3_250))); \
67925  __ret_250; \
67926})
67927#else
67928#define vqrdmlahs_lane_s32(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
67929  int32_t __s0_251 = __p0_251; \
67930  int32_t __s1_251 = __p1_251; \
67931  int32x2_t __s2_251 = __p2_251; \
67932  int32x2_t __rev2_251;  __rev2_251 = __builtin_shufflevector(__s2_251, __s2_251, 1, 0); \
67933  int32_t __ret_251; \
67934  __ret_251 = __noswap_vqadds_s32(__s0_251, __noswap_vqrdmulhs_s32(__s1_251, __noswap_vget_lane_s32(__rev2_251, __p3_251))); \
67935  __ret_251; \
67936})
67937#endif
67938
67939#ifdef __LITTLE_ENDIAN__
67940#define vqrdmlahh_lane_s16(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
67941  int16_t __s0_252 = __p0_252; \
67942  int16_t __s1_252 = __p1_252; \
67943  int16x4_t __s2_252 = __p2_252; \
67944  int16_t __ret_252; \
67945  __ret_252 = vqaddh_s16(__s0_252, vqrdmulhh_s16(__s1_252, vget_lane_s16(__s2_252, __p3_252))); \
67946  __ret_252; \
67947})
67948#else
67949#define vqrdmlahh_lane_s16(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
67950  int16_t __s0_253 = __p0_253; \
67951  int16_t __s1_253 = __p1_253; \
67952  int16x4_t __s2_253 = __p2_253; \
67953  int16x4_t __rev2_253;  __rev2_253 = __builtin_shufflevector(__s2_253, __s2_253, 3, 2, 1, 0); \
67954  int16_t __ret_253; \
67955  __ret_253 = __noswap_vqaddh_s16(__s0_253, __noswap_vqrdmulhh_s16(__s1_253, __noswap_vget_lane_s16(__rev2_253, __p3_253))); \
67956  __ret_253; \
67957})
67958#endif
67959
67960#ifdef __LITTLE_ENDIAN__
67961#define vqrdmlahs_laneq_s32(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
67962  int32_t __s0_254 = __p0_254; \
67963  int32_t __s1_254 = __p1_254; \
67964  int32x4_t __s2_254 = __p2_254; \
67965  int32_t __ret_254; \
67966  __ret_254 = vqadds_s32(__s0_254, vqrdmulhs_s32(__s1_254, vgetq_lane_s32(__s2_254, __p3_254))); \
67967  __ret_254; \
67968})
67969#else
67970#define vqrdmlahs_laneq_s32(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
67971  int32_t __s0_255 = __p0_255; \
67972  int32_t __s1_255 = __p1_255; \
67973  int32x4_t __s2_255 = __p2_255; \
67974  int32x4_t __rev2_255;  __rev2_255 = __builtin_shufflevector(__s2_255, __s2_255, 3, 2, 1, 0); \
67975  int32_t __ret_255; \
67976  __ret_255 = __noswap_vqadds_s32(__s0_255, __noswap_vqrdmulhs_s32(__s1_255, __noswap_vgetq_lane_s32(__rev2_255, __p3_255))); \
67977  __ret_255; \
67978})
67979#endif
67980
67981#ifdef __LITTLE_ENDIAN__
67982#define vqrdmlahh_laneq_s16(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
67983  int16_t __s0_256 = __p0_256; \
67984  int16_t __s1_256 = __p1_256; \
67985  int16x8_t __s2_256 = __p2_256; \
67986  int16_t __ret_256; \
67987  __ret_256 = vqaddh_s16(__s0_256, vqrdmulhh_s16(__s1_256, vgetq_lane_s16(__s2_256, __p3_256))); \
67988  __ret_256; \
67989})
67990#else
67991#define vqrdmlahh_laneq_s16(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
67992  int16_t __s0_257 = __p0_257; \
67993  int16_t __s1_257 = __p1_257; \
67994  int16x8_t __s2_257 = __p2_257; \
67995  int16x8_t __rev2_257;  __rev2_257 = __builtin_shufflevector(__s2_257, __s2_257, 7, 6, 5, 4, 3, 2, 1, 0); \
67996  int16_t __ret_257; \
67997  __ret_257 = __noswap_vqaddh_s16(__s0_257, __noswap_vqrdmulhh_s16(__s1_257, __noswap_vgetq_lane_s16(__rev2_257, __p3_257))); \
67998  __ret_257; \
67999})
68000#endif
68001
68002#ifdef __LITTLE_ENDIAN__
68003__ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68004  int32_t __ret;
68005  __ret = vqsubs_s32(__p0, vqrdmulhs_s32(__p1, __p2));
68006  return __ret;
68007}
68008#else
68009__ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68010  int32_t __ret;
68011  __ret = __noswap_vqsubs_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
68012  return __ret;
68013}
68014#endif
68015
68016#ifdef __LITTLE_ENDIAN__
68017__ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68018  int16_t __ret;
68019  __ret = vqsubh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
68020  return __ret;
68021}
68022#else
68023__ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68024  int16_t __ret;
68025  __ret = __noswap_vqsubh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
68026  return __ret;
68027}
68028#endif
68029
68030#ifdef __LITTLE_ENDIAN__
68031#define vqrdmlshs_lane_s32(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
68032  int32_t __s0_258 = __p0_258; \
68033  int32_t __s1_258 = __p1_258; \
68034  int32x2_t __s2_258 = __p2_258; \
68035  int32_t __ret_258; \
68036  __ret_258 = vqsubs_s32(__s0_258, vqrdmulhs_s32(__s1_258, vget_lane_s32(__s2_258, __p3_258))); \
68037  __ret_258; \
68038})
68039#else
68040#define vqrdmlshs_lane_s32(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
68041  int32_t __s0_259 = __p0_259; \
68042  int32_t __s1_259 = __p1_259; \
68043  int32x2_t __s2_259 = __p2_259; \
68044  int32x2_t __rev2_259;  __rev2_259 = __builtin_shufflevector(__s2_259, __s2_259, 1, 0); \
68045  int32_t __ret_259; \
68046  __ret_259 = __noswap_vqsubs_s32(__s0_259, __noswap_vqrdmulhs_s32(__s1_259, __noswap_vget_lane_s32(__rev2_259, __p3_259))); \
68047  __ret_259; \
68048})
68049#endif
68050
68051#ifdef __LITTLE_ENDIAN__
68052#define vqrdmlshh_lane_s16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
68053  int16_t __s0_260 = __p0_260; \
68054  int16_t __s1_260 = __p1_260; \
68055  int16x4_t __s2_260 = __p2_260; \
68056  int16_t __ret_260; \
68057  __ret_260 = vqsubh_s16(__s0_260, vqrdmulhh_s16(__s1_260, vget_lane_s16(__s2_260, __p3_260))); \
68058  __ret_260; \
68059})
68060#else
68061#define vqrdmlshh_lane_s16(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
68062  int16_t __s0_261 = __p0_261; \
68063  int16_t __s1_261 = __p1_261; \
68064  int16x4_t __s2_261 = __p2_261; \
68065  int16x4_t __rev2_261;  __rev2_261 = __builtin_shufflevector(__s2_261, __s2_261, 3, 2, 1, 0); \
68066  int16_t __ret_261; \
68067  __ret_261 = __noswap_vqsubh_s16(__s0_261, __noswap_vqrdmulhh_s16(__s1_261, __noswap_vget_lane_s16(__rev2_261, __p3_261))); \
68068  __ret_261; \
68069})
68070#endif
68071
68072#ifdef __LITTLE_ENDIAN__
68073#define vqrdmlshs_laneq_s32(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
68074  int32_t __s0_262 = __p0_262; \
68075  int32_t __s1_262 = __p1_262; \
68076  int32x4_t __s2_262 = __p2_262; \
68077  int32_t __ret_262; \
68078  __ret_262 = vqsubs_s32(__s0_262, vqrdmulhs_s32(__s1_262, vgetq_lane_s32(__s2_262, __p3_262))); \
68079  __ret_262; \
68080})
68081#else
68082#define vqrdmlshs_laneq_s32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
68083  int32_t __s0_263 = __p0_263; \
68084  int32_t __s1_263 = __p1_263; \
68085  int32x4_t __s2_263 = __p2_263; \
68086  int32x4_t __rev2_263;  __rev2_263 = __builtin_shufflevector(__s2_263, __s2_263, 3, 2, 1, 0); \
68087  int32_t __ret_263; \
68088  __ret_263 = __noswap_vqsubs_s32(__s0_263, __noswap_vqrdmulhs_s32(__s1_263, __noswap_vgetq_lane_s32(__rev2_263, __p3_263))); \
68089  __ret_263; \
68090})
68091#endif
68092
68093#ifdef __LITTLE_ENDIAN__
68094#define vqrdmlshh_laneq_s16(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
68095  int16_t __s0_264 = __p0_264; \
68096  int16_t __s1_264 = __p1_264; \
68097  int16x8_t __s2_264 = __p2_264; \
68098  int16_t __ret_264; \
68099  __ret_264 = vqsubh_s16(__s0_264, vqrdmulhh_s16(__s1_264, vgetq_lane_s16(__s2_264, __p3_264))); \
68100  __ret_264; \
68101})
68102#else
68103#define vqrdmlshh_laneq_s16(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
68104  int16_t __s0_265 = __p0_265; \
68105  int16_t __s1_265 = __p1_265; \
68106  int16x8_t __s2_265 = __p2_265; \
68107  int16x8_t __rev2_265;  __rev2_265 = __builtin_shufflevector(__s2_265, __s2_265, 7, 6, 5, 4, 3, 2, 1, 0); \
68108  int16_t __ret_265; \
68109  __ret_265 = __noswap_vqsubh_s16(__s0_265, __noswap_vqrdmulhh_s16(__s1_265, __noswap_vgetq_lane_s16(__rev2_265, __p3_265))); \
68110  __ret_265; \
68111})
68112#endif
68113
68114#endif
68115#if defined(__aarch64__)
68116#ifdef __LITTLE_ENDIAN__
68117__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68118  uint16x8_t __ret;
68119  __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
68120  return __ret;
68121}
68122#else
68123__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68124  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68125  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68126  uint16x8_t __ret;
68127  __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
68128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68129  return __ret;
68130}
68131#endif
68132
68133#ifdef __LITTLE_ENDIAN__
68134__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68135  uint64x2_t __ret;
68136  __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
68137  return __ret;
68138}
68139#else
68140__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68141  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68142  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68143  uint64x2_t __ret;
68144  __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
68145  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68146  return __ret;
68147}
68148#endif
68149
68150#ifdef __LITTLE_ENDIAN__
68151__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68152  uint32x4_t __ret;
68153  __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
68154  return __ret;
68155}
68156#else
68157__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68158  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68159  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68160  uint32x4_t __ret;
68161  __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
68162  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68163  return __ret;
68164}
68165#endif
68166
68167#ifdef __LITTLE_ENDIAN__
68168__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68169  int16x8_t __ret;
68170  __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
68171  return __ret;
68172}
68173#else
68174__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68175  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68176  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68177  int16x8_t __ret;
68178  __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
68179  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68180  return __ret;
68181}
68182#endif
68183
68184#ifdef __LITTLE_ENDIAN__
68185__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68186  int64x2_t __ret;
68187  __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
68188  return __ret;
68189}
68190#else
68191__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68192  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68193  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68194  int64x2_t __ret;
68195  __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
68196  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68197  return __ret;
68198}
68199#endif
68200
68201#ifdef __LITTLE_ENDIAN__
68202__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68203  int32x4_t __ret;
68204  __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
68205  return __ret;
68206}
68207#else
68208__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68209  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68210  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68211  int32x4_t __ret;
68212  __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
68213  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68214  return __ret;
68215}
68216#endif
68217
68218#ifdef __LITTLE_ENDIAN__
68219__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68220  uint16x8_t __ret;
68221  __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
68222  return __ret;
68223}
68224#else
68225__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68226  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68227  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68228  uint16x8_t __ret;
68229  __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
68230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68231  return __ret;
68232}
68233#endif
68234
68235#ifdef __LITTLE_ENDIAN__
68236__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68237  uint64x2_t __ret;
68238  __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
68239  return __ret;
68240}
68241#else
68242__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68243  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68244  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68245  uint64x2_t __ret;
68246  __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
68247  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68248  return __ret;
68249}
68250#endif
68251
68252#ifdef __LITTLE_ENDIAN__
68253__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68254  uint32x4_t __ret;
68255  __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
68256  return __ret;
68257}
68258#else
68259__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68260  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68261  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68262  uint32x4_t __ret;
68263  __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
68264  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68265  return __ret;
68266}
68267#endif
68268
68269#ifdef __LITTLE_ENDIAN__
68270__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68271  int16x8_t __ret;
68272  __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
68273  return __ret;
68274}
68275#else
68276__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68277  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68278  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68279  int16x8_t __ret;
68280  __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
68281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68282  return __ret;
68283}
68284#endif
68285
68286#ifdef __LITTLE_ENDIAN__
68287__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68288  int64x2_t __ret;
68289  __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
68290  return __ret;
68291}
68292#else
68293__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68294  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68295  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68296  int64x2_t __ret;
68297  __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
68298  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68299  return __ret;
68300}
68301#endif
68302
68303#ifdef __LITTLE_ENDIAN__
68304__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68305  int32x4_t __ret;
68306  __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
68307  return __ret;
68308}
68309#else
68310__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68311  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68312  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68313  int32x4_t __ret;
68314  __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
68315  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68316  return __ret;
68317}
68318#endif
68319
68320#ifdef __LITTLE_ENDIAN__
68321__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68322  uint16x8_t __ret;
68323  __ret = __p0 + vmovl_high_u8(__p1);
68324  return __ret;
68325}
68326#else
68327__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68328  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68329  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68330  uint16x8_t __ret;
68331  __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
68332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68333  return __ret;
68334}
68335#endif
68336
68337#ifdef __LITTLE_ENDIAN__
68338__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68339  uint64x2_t __ret;
68340  __ret = __p0 + vmovl_high_u32(__p1);
68341  return __ret;
68342}
68343#else
68344__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68345  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68346  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68347  uint64x2_t __ret;
68348  __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
68349  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68350  return __ret;
68351}
68352#endif
68353
68354#ifdef __LITTLE_ENDIAN__
68355__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68356  uint32x4_t __ret;
68357  __ret = __p0 + vmovl_high_u16(__p1);
68358  return __ret;
68359}
68360#else
68361__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68362  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68363  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68364  uint32x4_t __ret;
68365  __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
68366  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68367  return __ret;
68368}
68369#endif
68370
68371#ifdef __LITTLE_ENDIAN__
68372__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68373  int16x8_t __ret;
68374  __ret = __p0 + vmovl_high_s8(__p1);
68375  return __ret;
68376}
68377#else
68378__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68379  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68380  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68381  int16x8_t __ret;
68382  __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
68383  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68384  return __ret;
68385}
68386#endif
68387
68388#ifdef __LITTLE_ENDIAN__
68389__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68390  int64x2_t __ret;
68391  __ret = __p0 + vmovl_high_s32(__p1);
68392  return __ret;
68393}
68394#else
68395__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68396  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68397  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68398  int64x2_t __ret;
68399  __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
68400  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68401  return __ret;
68402}
68403#endif
68404
68405#ifdef __LITTLE_ENDIAN__
68406__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68407  int32x4_t __ret;
68408  __ret = __p0 + vmovl_high_s16(__p1);
68409  return __ret;
68410}
68411#else
68412__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68413  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68414  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68415  int32x4_t __ret;
68416  __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
68417  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68418  return __ret;
68419}
68420#endif
68421
68422#ifdef __LITTLE_ENDIAN__
68423#define vcopyq_lane_p64(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
68424  poly64x2_t __s0_266 = __p0_266; \
68425  poly64x1_t __s2_266 = __p2_266; \
68426  poly64x2_t __ret_266; \
68427  __ret_266 = vsetq_lane_p64(vget_lane_p64(__s2_266, __p3_266), __s0_266, __p1_266); \
68428  __ret_266; \
68429})
68430#else
68431#define vcopyq_lane_p64(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
68432  poly64x2_t __s0_267 = __p0_267; \
68433  poly64x1_t __s2_267 = __p2_267; \
68434  poly64x2_t __rev0_267;  __rev0_267 = __builtin_shufflevector(__s0_267, __s0_267, 1, 0); \
68435  poly64x2_t __ret_267; \
68436  __ret_267 = __noswap_vsetq_lane_p64(__noswap_vget_lane_p64(__s2_267, __p3_267), __rev0_267, __p1_267); \
68437  __ret_267 = __builtin_shufflevector(__ret_267, __ret_267, 1, 0); \
68438  __ret_267; \
68439})
68440#endif
68441
68442#ifdef __LITTLE_ENDIAN__
68443#define vcopyq_lane_f64(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
68444  float64x2_t __s0_268 = __p0_268; \
68445  float64x1_t __s2_268 = __p2_268; \
68446  float64x2_t __ret_268; \
68447  __ret_268 = vsetq_lane_f64(vget_lane_f64(__s2_268, __p3_268), __s0_268, __p1_268); \
68448  __ret_268; \
68449})
68450#else
68451#define vcopyq_lane_f64(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
68452  float64x2_t __s0_269 = __p0_269; \
68453  float64x1_t __s2_269 = __p2_269; \
68454  float64x2_t __rev0_269;  __rev0_269 = __builtin_shufflevector(__s0_269, __s0_269, 1, 0); \
68455  float64x2_t __ret_269; \
68456  __ret_269 = __noswap_vsetq_lane_f64(__noswap_vget_lane_f64(__s2_269, __p3_269), __rev0_269, __p1_269); \
68457  __ret_269 = __builtin_shufflevector(__ret_269, __ret_269, 1, 0); \
68458  __ret_269; \
68459})
68460#endif
68461
68462#ifdef __LITTLE_ENDIAN__
68463#define vcopy_lane_p64(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
68464  poly64x1_t __s0_270 = __p0_270; \
68465  poly64x1_t __s2_270 = __p2_270; \
68466  poly64x1_t __ret_270; \
68467  __ret_270 = vset_lane_p64(vget_lane_p64(__s2_270, __p3_270), __s0_270, __p1_270); \
68468  __ret_270; \
68469})
68470#else
68471#define vcopy_lane_p64(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
68472  poly64x1_t __s0_271 = __p0_271; \
68473  poly64x1_t __s2_271 = __p2_271; \
68474  poly64x1_t __ret_271; \
68475  __ret_271 = __noswap_vset_lane_p64(__noswap_vget_lane_p64(__s2_271, __p3_271), __s0_271, __p1_271); \
68476  __ret_271; \
68477})
68478#endif
68479
68480#ifdef __LITTLE_ENDIAN__
68481#define vcopy_lane_f64(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
68482  float64x1_t __s0_272 = __p0_272; \
68483  float64x1_t __s2_272 = __p2_272; \
68484  float64x1_t __ret_272; \
68485  __ret_272 = vset_lane_f64(vget_lane_f64(__s2_272, __p3_272), __s0_272, __p1_272); \
68486  __ret_272; \
68487})
68488#else
68489#define vcopy_lane_f64(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
68490  float64x1_t __s0_273 = __p0_273; \
68491  float64x1_t __s2_273 = __p2_273; \
68492  float64x1_t __ret_273; \
68493  __ret_273 = __noswap_vset_lane_f64(__noswap_vget_lane_f64(__s2_273, __p3_273), __s0_273, __p1_273); \
68494  __ret_273; \
68495})
68496#endif
68497
68498#ifdef __LITTLE_ENDIAN__
68499#define vcopyq_laneq_p64(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
68500  poly64x2_t __s0_274 = __p0_274; \
68501  poly64x2_t __s2_274 = __p2_274; \
68502  poly64x2_t __ret_274; \
68503  __ret_274 = vsetq_lane_p64(vgetq_lane_p64(__s2_274, __p3_274), __s0_274, __p1_274); \
68504  __ret_274; \
68505})
68506#else
68507#define vcopyq_laneq_p64(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
68508  poly64x2_t __s0_275 = __p0_275; \
68509  poly64x2_t __s2_275 = __p2_275; \
68510  poly64x2_t __rev0_275;  __rev0_275 = __builtin_shufflevector(__s0_275, __s0_275, 1, 0); \
68511  poly64x2_t __rev2_275;  __rev2_275 = __builtin_shufflevector(__s2_275, __s2_275, 1, 0); \
68512  poly64x2_t __ret_275; \
68513  __ret_275 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_275, __p3_275), __rev0_275, __p1_275); \
68514  __ret_275 = __builtin_shufflevector(__ret_275, __ret_275, 1, 0); \
68515  __ret_275; \
68516})
68517#endif
68518
68519#ifdef __LITTLE_ENDIAN__
68520#define vcopyq_laneq_f64(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
68521  float64x2_t __s0_276 = __p0_276; \
68522  float64x2_t __s2_276 = __p2_276; \
68523  float64x2_t __ret_276; \
68524  __ret_276 = vsetq_lane_f64(vgetq_lane_f64(__s2_276, __p3_276), __s0_276, __p1_276); \
68525  __ret_276; \
68526})
68527#else
68528#define vcopyq_laneq_f64(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
68529  float64x2_t __s0_277 = __p0_277; \
68530  float64x2_t __s2_277 = __p2_277; \
68531  float64x2_t __rev0_277;  __rev0_277 = __builtin_shufflevector(__s0_277, __s0_277, 1, 0); \
68532  float64x2_t __rev2_277;  __rev2_277 = __builtin_shufflevector(__s2_277, __s2_277, 1, 0); \
68533  float64x2_t __ret_277; \
68534  __ret_277 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_277, __p3_277), __rev0_277, __p1_277); \
68535  __ret_277 = __builtin_shufflevector(__ret_277, __ret_277, 1, 0); \
68536  __ret_277; \
68537})
68538#endif
68539
68540#ifdef __LITTLE_ENDIAN__
68541#define vcopy_laneq_p64(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
68542  poly64x1_t __s0_278 = __p0_278; \
68543  poly64x2_t __s2_278 = __p2_278; \
68544  poly64x1_t __ret_278; \
68545  __ret_278 = vset_lane_p64(vgetq_lane_p64(__s2_278, __p3_278), __s0_278, __p1_278); \
68546  __ret_278; \
68547})
68548#else
68549#define vcopy_laneq_p64(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
68550  poly64x1_t __s0_279 = __p0_279; \
68551  poly64x2_t __s2_279 = __p2_279; \
68552  poly64x2_t __rev2_279;  __rev2_279 = __builtin_shufflevector(__s2_279, __s2_279, 1, 0); \
68553  poly64x1_t __ret_279; \
68554  __ret_279 = __noswap_vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_279, __p3_279), __s0_279, __p1_279); \
68555  __ret_279; \
68556})
68557#endif
68558
68559#ifdef __LITTLE_ENDIAN__
68560#define vcopy_laneq_f64(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
68561  float64x1_t __s0_280 = __p0_280; \
68562  float64x2_t __s2_280 = __p2_280; \
68563  float64x1_t __ret_280; \
68564  __ret_280 = vset_lane_f64(vgetq_lane_f64(__s2_280, __p3_280), __s0_280, __p1_280); \
68565  __ret_280; \
68566})
68567#else
68568#define vcopy_laneq_f64(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
68569  float64x1_t __s0_281 = __p0_281; \
68570  float64x2_t __s2_281 = __p2_281; \
68571  float64x2_t __rev2_281;  __rev2_281 = __builtin_shufflevector(__s2_281, __s2_281, 1, 0); \
68572  float64x1_t __ret_281; \
68573  __ret_281 = __noswap_vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_281, __p3_281), __s0_281, __p1_281); \
68574  __ret_281; \
68575})
68576#endif
68577
68578#ifdef __LITTLE_ENDIAN__
68579__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68580  uint16x8_t __ret;
68581  __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68582  return __ret;
68583}
68584#else
68585__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68586  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68587  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68588  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68589  uint16x8_t __ret;
68590  __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68591  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68592  return __ret;
68593}
68594#endif
68595
68596#ifdef __LITTLE_ENDIAN__
68597__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68598  uint64x2_t __ret;
68599  __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68600  return __ret;
68601}
68602#else
68603__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68604  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68605  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68606  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68607  uint64x2_t __ret;
68608  __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68609  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68610  return __ret;
68611}
68612#endif
68613
68614#ifdef __LITTLE_ENDIAN__
68615__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68616  uint32x4_t __ret;
68617  __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68618  return __ret;
68619}
68620#else
68621__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68622  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68623  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68624  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68625  uint32x4_t __ret;
68626  __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68627  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68628  return __ret;
68629}
68630#endif
68631
68632#ifdef __LITTLE_ENDIAN__
68633__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68634  int16x8_t __ret;
68635  __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68636  return __ret;
68637}
68638#else
68639__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68640  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68641  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68642  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68643  int16x8_t __ret;
68644  __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68645  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68646  return __ret;
68647}
68648#endif
68649
68650#ifdef __LITTLE_ENDIAN__
68651__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68652  int64x2_t __ret;
68653  __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68654  return __ret;
68655}
68656#else
68657__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68658  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68659  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68660  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68661  int64x2_t __ret;
68662  __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68663  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68664  return __ret;
68665}
68666#endif
68667
68668#ifdef __LITTLE_ENDIAN__
68669__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68670  int32x4_t __ret;
68671  __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68672  return __ret;
68673}
68674#else
68675__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68676  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68677  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68678  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68679  int32x4_t __ret;
68680  __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68682  return __ret;
68683}
68684#endif
68685
68686#ifdef __LITTLE_ENDIAN__
68687__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68688  uint64x2_t __ret;
68689  __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
68690  return __ret;
68691}
68692#else
68693__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68694  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68695  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68696  uint64x2_t __ret;
68697  __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68699  return __ret;
68700}
68701#endif
68702
68703#ifdef __LITTLE_ENDIAN__
68704__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68705  uint32x4_t __ret;
68706  __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
68707  return __ret;
68708}
68709#else
68710__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68711  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68712  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68713  uint32x4_t __ret;
68714  __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68716  return __ret;
68717}
68718#endif
68719
68720#ifdef __LITTLE_ENDIAN__
68721__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68722  int64x2_t __ret;
68723  __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
68724  return __ret;
68725}
68726#else
68727__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68728  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68729  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68730  int64x2_t __ret;
68731  __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68733  return __ret;
68734}
68735#endif
68736
68737#ifdef __LITTLE_ENDIAN__
68738__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68739  int32x4_t __ret;
68740  __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
68741  return __ret;
68742}
68743#else
68744__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68745  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68746  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68747  int32x4_t __ret;
68748  __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68749  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68750  return __ret;
68751}
68752#endif
68753
68754#ifdef __LITTLE_ENDIAN__
68755__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68756  uint16x8_t __ret;
68757  __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68758  return __ret;
68759}
68760#else
68761__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68762  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68763  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68764  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68765  uint16x8_t __ret;
68766  __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68767  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68768  return __ret;
68769}
68770#endif
68771
68772#ifdef __LITTLE_ENDIAN__
68773__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68774  uint64x2_t __ret;
68775  __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68776  return __ret;
68777}
68778#else
68779__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68780  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68781  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68782  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68783  uint64x2_t __ret;
68784  __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68786  return __ret;
68787}
68788#endif
68789
68790#ifdef __LITTLE_ENDIAN__
68791__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68792  uint32x4_t __ret;
68793  __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68794  return __ret;
68795}
68796#else
68797__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68798  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68799  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68800  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68801  uint32x4_t __ret;
68802  __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68804  return __ret;
68805}
68806#endif
68807
68808#ifdef __LITTLE_ENDIAN__
68809__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68810  int16x8_t __ret;
68811  __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68812  return __ret;
68813}
68814#else
68815__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68816  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68817  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68818  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68819  int16x8_t __ret;
68820  __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68822  return __ret;
68823}
68824#endif
68825
68826#ifdef __LITTLE_ENDIAN__
68827__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68828  int64x2_t __ret;
68829  __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68830  return __ret;
68831}
68832#else
68833__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68834  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68835  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68836  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68837  int64x2_t __ret;
68838  __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68839  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68840  return __ret;
68841}
68842#endif
68843
68844#ifdef __LITTLE_ENDIAN__
68845__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68846  int32x4_t __ret;
68847  __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68848  return __ret;
68849}
68850#else
68851__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68852  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68853  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68854  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68855  int32x4_t __ret;
68856  __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68858  return __ret;
68859}
68860#endif
68861
68862#ifdef __LITTLE_ENDIAN__
68863__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68864  uint64x2_t __ret;
68865  __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
68866  return __ret;
68867}
68868#else
68869__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68870  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68871  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68872  uint64x2_t __ret;
68873  __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68874  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68875  return __ret;
68876}
68877#endif
68878
68879#ifdef __LITTLE_ENDIAN__
68880__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68881  uint32x4_t __ret;
68882  __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
68883  return __ret;
68884}
68885#else
68886__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68887  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68888  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68889  uint32x4_t __ret;
68890  __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68891  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68892  return __ret;
68893}
68894#endif
68895
68896#ifdef __LITTLE_ENDIAN__
68897__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68898  int64x2_t __ret;
68899  __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
68900  return __ret;
68901}
68902#else
68903__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68904  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68905  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68906  int64x2_t __ret;
68907  __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68908  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68909  return __ret;
68910}
68911#endif
68912
68913#ifdef __LITTLE_ENDIAN__
68914__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68915  int32x4_t __ret;
68916  __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
68917  return __ret;
68918}
68919#else
68920__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68921  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68922  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68923  int32x4_t __ret;
68924  __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68926  return __ret;
68927}
68928#endif
68929
68930#ifdef __LITTLE_ENDIAN__
68931#define vmulx_lane_f64(__p0_282, __p1_282, __p2_282) __extension__ ({ \
68932  float64x1_t __s0_282 = __p0_282; \
68933  float64x1_t __s1_282 = __p1_282; \
68934  float64x1_t __ret_282; \
68935  float64_t __x_282 = vget_lane_f64(__s0_282, 0); \
68936  float64_t __y_282 = vget_lane_f64(__s1_282, __p2_282); \
68937  float64_t __z_282 = vmulxd_f64(__x_282, __y_282); \
68938  __ret_282 = vset_lane_f64(__z_282, __s0_282, __p2_282); \
68939  __ret_282; \
68940})
68941#else
68942#define vmulx_lane_f64(__p0_283, __p1_283, __p2_283) __extension__ ({ \
68943  float64x1_t __s0_283 = __p0_283; \
68944  float64x1_t __s1_283 = __p1_283; \
68945  float64x1_t __ret_283; \
68946  float64_t __x_283 = __noswap_vget_lane_f64(__s0_283, 0); \
68947  float64_t __y_283 = __noswap_vget_lane_f64(__s1_283, __p2_283); \
68948  float64_t __z_283 = __noswap_vmulxd_f64(__x_283, __y_283); \
68949  __ret_283 = __noswap_vset_lane_f64(__z_283, __s0_283, __p2_283); \
68950  __ret_283; \
68951})
68952#endif
68953
68954#ifdef __LITTLE_ENDIAN__
68955#define vmulx_laneq_f64(__p0_284, __p1_284, __p2_284) __extension__ ({ \
68956  float64x1_t __s0_284 = __p0_284; \
68957  float64x2_t __s1_284 = __p1_284; \
68958  float64x1_t __ret_284; \
68959  float64_t __x_284 = vget_lane_f64(__s0_284, 0); \
68960  float64_t __y_284 = vgetq_lane_f64(__s1_284, __p2_284); \
68961  float64_t __z_284 = vmulxd_f64(__x_284, __y_284); \
68962  __ret_284 = vset_lane_f64(__z_284, __s0_284, 0); \
68963  __ret_284; \
68964})
68965#else
68966#define vmulx_laneq_f64(__p0_285, __p1_285, __p2_285) __extension__ ({ \
68967  float64x1_t __s0_285 = __p0_285; \
68968  float64x2_t __s1_285 = __p1_285; \
68969  float64x2_t __rev1_285;  __rev1_285 = __builtin_shufflevector(__s1_285, __s1_285, 1, 0); \
68970  float64x1_t __ret_285; \
68971  float64_t __x_285 = __noswap_vget_lane_f64(__s0_285, 0); \
68972  float64_t __y_285 = __noswap_vgetq_lane_f64(__rev1_285, __p2_285); \
68973  float64_t __z_285 = __noswap_vmulxd_f64(__x_285, __y_285); \
68974  __ret_285 = __noswap_vset_lane_f64(__z_285, __s0_285, 0); \
68975  __ret_285; \
68976})
68977#endif
68978
68979#endif
68980#ifdef __LITTLE_ENDIAN__
68981__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68982  uint16x8_t __ret;
68983  __ret = __p0 + vabdl_u8(__p1, __p2);
68984  return __ret;
68985}
68986#else
68987__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68988  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68989  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68990  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68991  uint16x8_t __ret;
68992  __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
68993  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68994  return __ret;
68995}
68996__ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68997  uint16x8_t __ret;
68998  __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
68999  return __ret;
69000}
69001#endif
69002
69003#ifdef __LITTLE_ENDIAN__
69004__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69005  uint64x2_t __ret;
69006  __ret = __p0 + vabdl_u32(__p1, __p2);
69007  return __ret;
69008}
69009#else
69010__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69011  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69012  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69013  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69014  uint64x2_t __ret;
69015  __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
69016  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69017  return __ret;
69018}
69019__ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69020  uint64x2_t __ret;
69021  __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
69022  return __ret;
69023}
69024#endif
69025
69026#ifdef __LITTLE_ENDIAN__
69027__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69028  uint32x4_t __ret;
69029  __ret = __p0 + vabdl_u16(__p1, __p2);
69030  return __ret;
69031}
69032#else
69033__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69034  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69035  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69036  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69037  uint32x4_t __ret;
69038  __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
69039  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69040  return __ret;
69041}
69042__ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69043  uint32x4_t __ret;
69044  __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
69045  return __ret;
69046}
69047#endif
69048
69049#ifdef __LITTLE_ENDIAN__
69050__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69051  int16x8_t __ret;
69052  __ret = __p0 + vabdl_s8(__p1, __p2);
69053  return __ret;
69054}
69055#else
69056__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69057  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69058  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69059  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69060  int16x8_t __ret;
69061  __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
69062  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69063  return __ret;
69064}
69065__ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69066  int16x8_t __ret;
69067  __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
69068  return __ret;
69069}
69070#endif
69071
69072#ifdef __LITTLE_ENDIAN__
69073__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69074  int64x2_t __ret;
69075  __ret = __p0 + vabdl_s32(__p1, __p2);
69076  return __ret;
69077}
69078#else
69079__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69080  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69081  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69082  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69083  int64x2_t __ret;
69084  __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
69085  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69086  return __ret;
69087}
69088__ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69089  int64x2_t __ret;
69090  __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
69091  return __ret;
69092}
69093#endif
69094
69095#ifdef __LITTLE_ENDIAN__
69096__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69097  int32x4_t __ret;
69098  __ret = __p0 + vabdl_s16(__p1, __p2);
69099  return __ret;
69100}
69101#else
69102__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69103  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69104  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69105  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69106  int32x4_t __ret;
69107  __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
69108  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69109  return __ret;
69110}
69111__ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69112  int32x4_t __ret;
69113  __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
69114  return __ret;
69115}
69116#endif
69117
69118#if defined(__aarch64__)
69119#ifdef __LITTLE_ENDIAN__
69120__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69121  uint16x8_t __ret;
69122  __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
69123  return __ret;
69124}
69125#else
69126__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69127  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69128  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69129  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69130  uint16x8_t __ret;
69131  __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
69132  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69133  return __ret;
69134}
69135#endif
69136
69137#ifdef __LITTLE_ENDIAN__
69138__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69139  uint64x2_t __ret;
69140  __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
69141  return __ret;
69142}
69143#else
69144__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69145  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69146  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69147  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69148  uint64x2_t __ret;
69149  __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
69150  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69151  return __ret;
69152}
69153#endif
69154
69155#ifdef __LITTLE_ENDIAN__
69156__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69157  uint32x4_t __ret;
69158  __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
69159  return __ret;
69160}
69161#else
69162__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69163  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69164  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69165  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69166  uint32x4_t __ret;
69167  __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
69168  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69169  return __ret;
69170}
69171#endif
69172
69173#ifdef __LITTLE_ENDIAN__
69174__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69175  int16x8_t __ret;
69176  __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
69177  return __ret;
69178}
69179#else
69180__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69181  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69182  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69183  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69184  int16x8_t __ret;
69185  __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
69186  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69187  return __ret;
69188}
69189#endif
69190
69191#ifdef __LITTLE_ENDIAN__
69192__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69193  int64x2_t __ret;
69194  __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
69195  return __ret;
69196}
69197#else
69198__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69199  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69200  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69201  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69202  int64x2_t __ret;
69203  __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
69204  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69205  return __ret;
69206}
69207#endif
69208
69209#ifdef __LITTLE_ENDIAN__
69210__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69211  int32x4_t __ret;
69212  __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
69213  return __ret;
69214}
69215#else
69216__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69217  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69218  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69219  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69220  int32x4_t __ret;
69221  __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
69222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69223  return __ret;
69224}
69225#endif
69226
69227#endif
69228
69229#undef __ai
69230
69231#endif /* __ARM_NEON_H */
69232