ETISS 0.8.0
Extendable Translating Instruction Set Simulator (version 0.8.0)
arm_neon.h
Go to the documentation of this file.
1 /*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2  *
3  * Permission is hereby granted, free of charge, to any person obtaining a copy
4  * of this software and associated documentation files (the "Software"), to deal
5  * in the Software without restriction, including without limitation the rights
6  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7  * copies of the Software, and to permit persons to whom the Software is
8  * furnished to do so, subject to the following conditions:
9  *
10  * The above copyright notice and this permission notice shall be included in
11  * all copies or substantial portions of the Software.
12  *
13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19  * THE SOFTWARE.
20  *
21  *===-----------------------------------------------------------------------===
22  */
23 
24 #ifndef __ARM_NEON_H
25 #define __ARM_NEON_H
26 
27 #ifndef __ARM_FP
28 #error "NEON intrinsics not available with the soft-float ABI. Please use -mfloat-abi=softfp or -mfloat-abi=hard"
29 #else
30 
31 #if !defined(__ARM_NEON)
32 #error "NEON support not enabled"
33 #else
34 
35 #include <stdint.h>
36 
37 #ifdef __ARM_FEATURE_BF16
38 #include <arm_bf16.h>
39 typedef __bf16 bfloat16_t;
40 #endif
41 
42 typedef float float32_t;
43 typedef __fp16 float16_t;
44 #ifdef __aarch64__
45 typedef double float64_t;
46 #endif
47 
48 #ifdef __aarch64__
49 typedef uint8_t poly8_t;
50 typedef uint16_t poly16_t;
51 typedef uint64_t poly64_t;
52 typedef __uint128_t poly128_t;
53 #else
54 typedef int8_t poly8_t;
55 typedef int16_t poly16_t;
56 typedef int64_t poly64_t;
57 #endif
58 typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
59 typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
60 typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
61 typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
62 typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
63 typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
64 typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
65 typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
66 typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
67 typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
68 typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
69 typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
70 typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
71 typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
72 typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
73 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
74 typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
75 typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
76 typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
77 typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
78 #ifdef __aarch64__
79 typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
80 typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
81 #endif
82 typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
83 typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
84 typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
85 typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
86 typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
87 typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
88 
89 typedef struct int8x8x2_t {
90  int8x8_t val[2];
91 } int8x8x2_t;
92 
93 typedef struct int8x16x2_t {
94  int8x16_t val[2];
95 } int8x16x2_t;
96 
97 typedef struct int16x4x2_t {
98  int16x4_t val[2];
99 } int16x4x2_t;
100 
101 typedef struct int16x8x2_t {
102  int16x8_t val[2];
103 } int16x8x2_t;
104 
105 typedef struct int32x2x2_t {
106  int32x2_t val[2];
107 } int32x2x2_t;
108 
109 typedef struct int32x4x2_t {
110  int32x4_t val[2];
111 } int32x4x2_t;
112 
113 typedef struct int64x1x2_t {
114  int64x1_t val[2];
115 } int64x1x2_t;
116 
117 typedef struct int64x2x2_t {
118  int64x2_t val[2];
119 } int64x2x2_t;
120 
121 typedef struct uint8x8x2_t {
122  uint8x8_t val[2];
123 } uint8x8x2_t;
124 
125 typedef struct uint8x16x2_t {
126  uint8x16_t val[2];
127 } uint8x16x2_t;
128 
129 typedef struct uint16x4x2_t {
130  uint16x4_t val[2];
131 } uint16x4x2_t;
132 
133 typedef struct uint16x8x2_t {
134  uint16x8_t val[2];
135 } uint16x8x2_t;
136 
137 typedef struct uint32x2x2_t {
138  uint32x2_t val[2];
139 } uint32x2x2_t;
140 
141 typedef struct uint32x4x2_t {
142  uint32x4_t val[2];
143 } uint32x4x2_t;
144 
145 typedef struct uint64x1x2_t {
146  uint64x1_t val[2];
147 } uint64x1x2_t;
148 
149 typedef struct uint64x2x2_t {
150  uint64x2_t val[2];
151 } uint64x2x2_t;
152 
153 typedef struct float16x4x2_t {
154  float16x4_t val[2];
155 } float16x4x2_t;
156 
157 typedef struct float16x8x2_t {
158  float16x8_t val[2];
159 } float16x8x2_t;
160 
161 typedef struct float32x2x2_t {
162  float32x2_t val[2];
163 } float32x2x2_t;
164 
165 typedef struct float32x4x2_t {
166  float32x4_t val[2];
167 } float32x4x2_t;
168 
169 #ifdef __aarch64__
170 typedef struct float64x1x2_t {
171  float64x1_t val[2];
172 } float64x1x2_t;
173 
174 typedef struct float64x2x2_t {
175  float64x2_t val[2];
176 } float64x2x2_t;
177 
178 #endif
179 typedef struct poly8x8x2_t {
180  poly8x8_t val[2];
181 } poly8x8x2_t;
182 
183 typedef struct poly8x16x2_t {
184  poly8x16_t val[2];
185 } poly8x16x2_t;
186 
187 typedef struct poly16x4x2_t {
188  poly16x4_t val[2];
189 } poly16x4x2_t;
190 
191 typedef struct poly16x8x2_t {
192  poly16x8_t val[2];
193 } poly16x8x2_t;
194 
195 typedef struct poly64x1x2_t {
196  poly64x1_t val[2];
197 } poly64x1x2_t;
198 
199 typedef struct poly64x2x2_t {
200  poly64x2_t val[2];
201 } poly64x2x2_t;
202 
203 typedef struct int8x8x3_t {
204  int8x8_t val[3];
205 } int8x8x3_t;
206 
207 typedef struct int8x16x3_t {
208  int8x16_t val[3];
209 } int8x16x3_t;
210 
211 typedef struct int16x4x3_t {
212  int16x4_t val[3];
213 } int16x4x3_t;
214 
215 typedef struct int16x8x3_t {
216  int16x8_t val[3];
217 } int16x8x3_t;
218 
219 typedef struct int32x2x3_t {
220  int32x2_t val[3];
221 } int32x2x3_t;
222 
223 typedef struct int32x4x3_t {
224  int32x4_t val[3];
225 } int32x4x3_t;
226 
227 typedef struct int64x1x3_t {
228  int64x1_t val[3];
229 } int64x1x3_t;
230 
231 typedef struct int64x2x3_t {
232  int64x2_t val[3];
233 } int64x2x3_t;
234 
235 typedef struct uint8x8x3_t {
236  uint8x8_t val[3];
237 } uint8x8x3_t;
238 
239 typedef struct uint8x16x3_t {
240  uint8x16_t val[3];
241 } uint8x16x3_t;
242 
243 typedef struct uint16x4x3_t {
244  uint16x4_t val[3];
245 } uint16x4x3_t;
246 
247 typedef struct uint16x8x3_t {
248  uint16x8_t val[3];
249 } uint16x8x3_t;
250 
251 typedef struct uint32x2x3_t {
252  uint32x2_t val[3];
253 } uint32x2x3_t;
254 
255 typedef struct uint32x4x3_t {
256  uint32x4_t val[3];
257 } uint32x4x3_t;
258 
259 typedef struct uint64x1x3_t {
260  uint64x1_t val[3];
261 } uint64x1x3_t;
262 
263 typedef struct uint64x2x3_t {
264  uint64x2_t val[3];
265 } uint64x2x3_t;
266 
267 typedef struct float16x4x3_t {
268  float16x4_t val[3];
269 } float16x4x3_t;
270 
271 typedef struct float16x8x3_t {
272  float16x8_t val[3];
273 } float16x8x3_t;
274 
275 typedef struct float32x2x3_t {
276  float32x2_t val[3];
277 } float32x2x3_t;
278 
279 typedef struct float32x4x3_t {
280  float32x4_t val[3];
281 } float32x4x3_t;
282 
283 #ifdef __aarch64__
284 typedef struct float64x1x3_t {
285  float64x1_t val[3];
286 } float64x1x3_t;
287 
288 typedef struct float64x2x3_t {
289  float64x2_t val[3];
290 } float64x2x3_t;
291 
292 #endif
293 typedef struct poly8x8x3_t {
294  poly8x8_t val[3];
295 } poly8x8x3_t;
296 
297 typedef struct poly8x16x3_t {
298  poly8x16_t val[3];
299 } poly8x16x3_t;
300 
301 typedef struct poly16x4x3_t {
302  poly16x4_t val[3];
303 } poly16x4x3_t;
304 
305 typedef struct poly16x8x3_t {
306  poly16x8_t val[3];
307 } poly16x8x3_t;
308 
309 typedef struct poly64x1x3_t {
310  poly64x1_t val[3];
311 } poly64x1x3_t;
312 
313 typedef struct poly64x2x3_t {
314  poly64x2_t val[3];
315 } poly64x2x3_t;
316 
317 typedef struct int8x8x4_t {
318  int8x8_t val[4];
319 } int8x8x4_t;
320 
321 typedef struct int8x16x4_t {
322  int8x16_t val[4];
323 } int8x16x4_t;
324 
325 typedef struct int16x4x4_t {
326  int16x4_t val[4];
327 } int16x4x4_t;
328 
329 typedef struct int16x8x4_t {
330  int16x8_t val[4];
331 } int16x8x4_t;
332 
333 typedef struct int32x2x4_t {
334  int32x2_t val[4];
335 } int32x2x4_t;
336 
337 typedef struct int32x4x4_t {
338  int32x4_t val[4];
339 } int32x4x4_t;
340 
341 typedef struct int64x1x4_t {
342  int64x1_t val[4];
343 } int64x1x4_t;
344 
345 typedef struct int64x2x4_t {
346  int64x2_t val[4];
347 } int64x2x4_t;
348 
349 typedef struct uint8x8x4_t {
350  uint8x8_t val[4];
351 } uint8x8x4_t;
352 
353 typedef struct uint8x16x4_t {
354  uint8x16_t val[4];
355 } uint8x16x4_t;
356 
357 typedef struct uint16x4x4_t {
358  uint16x4_t val[4];
359 } uint16x4x4_t;
360 
361 typedef struct uint16x8x4_t {
362  uint16x8_t val[4];
363 } uint16x8x4_t;
364 
365 typedef struct uint32x2x4_t {
366  uint32x2_t val[4];
367 } uint32x2x4_t;
368 
369 typedef struct uint32x4x4_t {
370  uint32x4_t val[4];
371 } uint32x4x4_t;
372 
373 typedef struct uint64x1x4_t {
374  uint64x1_t val[4];
375 } uint64x1x4_t;
376 
377 typedef struct uint64x2x4_t {
378  uint64x2_t val[4];
379 } uint64x2x4_t;
380 
381 typedef struct float16x4x4_t {
382  float16x4_t val[4];
383 } float16x4x4_t;
384 
385 typedef struct float16x8x4_t {
386  float16x8_t val[4];
387 } float16x8x4_t;
388 
389 typedef struct float32x2x4_t {
390  float32x2_t val[4];
391 } float32x2x4_t;
392 
393 typedef struct float32x4x4_t {
394  float32x4_t val[4];
395 } float32x4x4_t;
396 
397 #ifdef __aarch64__
398 typedef struct float64x1x4_t {
399  float64x1_t val[4];
400 } float64x1x4_t;
401 
402 typedef struct float64x2x4_t {
403  float64x2_t val[4];
404 } float64x2x4_t;
405 
406 #endif
407 typedef struct poly8x8x4_t {
408  poly8x8_t val[4];
409 } poly8x8x4_t;
410 
411 typedef struct poly8x16x4_t {
412  poly8x16_t val[4];
413 } poly8x16x4_t;
414 
415 typedef struct poly16x4x4_t {
416  poly16x4_t val[4];
417 } poly16x4x4_t;
418 
419 typedef struct poly16x8x4_t {
420  poly16x8_t val[4];
421 } poly16x8x4_t;
422 
423 typedef struct poly64x1x4_t {
424  poly64x1_t val[4];
425 } poly64x1x4_t;
426 
427 typedef struct poly64x2x4_t {
428  poly64x2_t val[4];
429 } poly64x2x4_t;
430 
431 #ifdef __ARM_FEATURE_BF16
432 typedef __attribute__((neon_vector_type(4))) bfloat16_t bfloat16x4_t;
433 typedef __attribute__((neon_vector_type(8))) bfloat16_t bfloat16x8_t;
434 
435 typedef struct bfloat16x4x2_t {
436  bfloat16x4_t val[2];
437 } bfloat16x4x2_t;
438 
439 typedef struct bfloat16x8x2_t {
440  bfloat16x8_t val[2];
441 } bfloat16x8x2_t;
442 
443 typedef struct bfloat16x4x3_t {
444  bfloat16x4_t val[3];
445 } bfloat16x4x3_t;
446 
447 typedef struct bfloat16x8x3_t {
448  bfloat16x8_t val[3];
449 } bfloat16x8x3_t;
450 
451 typedef struct bfloat16x4x4_t {
452  bfloat16x4_t val[4];
453 } bfloat16x4x4_t;
454 
455 typedef struct bfloat16x8x4_t {
456  bfloat16x8_t val[4];
457 } bfloat16x8x4_t;
458 
459 #endif
460 
461 #define __ai static __inline__ __attribute__((__always_inline__, __nodebug__))
462 
463 #ifdef __LITTLE_ENDIAN__
464 #define splat_lane_p8(__p0, __p1) __extension__ ({ \
465  poly8x8_t __s0 = __p0; \
466  poly8x8_t __ret; \
467  __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 4); \
468  __ret; \
469 })
470 #else
471 #define splat_lane_p8(__p0, __p1) __extension__ ({ \
472  poly8x8_t __s0 = __p0; \
473  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
474  poly8x8_t __ret; \
475  __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 4); \
476  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
477  __ret; \
478 })
479 #define __noswap_splat_lane_p8(__p0, __p1) __extension__ ({ \
480  poly8x8_t __s0 = __p0; \
481  poly8x8_t __ret; \
482  __ret = (poly8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 4); \
483  __ret; \
484 })
485 #endif
486 
487 #define splat_lane_p64(__p0, __p1) __extension__ ({ \
488  poly64x1_t __s0 = __p0; \
489  poly64x1_t __ret; \
490  __ret = (poly64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 6); \
491  __ret; \
492 })
493 #ifdef __LITTLE_ENDIAN__
494 #define splat_lane_p16(__p0, __p1) __extension__ ({ \
495  poly16x4_t __s0 = __p0; \
496  poly16x4_t __ret; \
497  __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 5); \
498  __ret; \
499 })
500 #else
501 #define splat_lane_p16(__p0, __p1) __extension__ ({ \
502  poly16x4_t __s0 = __p0; \
503  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
504  poly16x4_t __ret; \
505  __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 5); \
506  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
507  __ret; \
508 })
509 #define __noswap_splat_lane_p16(__p0, __p1) __extension__ ({ \
510  poly16x4_t __s0 = __p0; \
511  poly16x4_t __ret; \
512  __ret = (poly16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 5); \
513  __ret; \
514 })
515 #endif
516 
517 #ifdef __LITTLE_ENDIAN__
518 #define splatq_lane_p8(__p0, __p1) __extension__ ({ \
519  poly8x8_t __s0 = __p0; \
520  poly8x16_t __ret; \
521  __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 4); \
522  __ret; \
523 })
524 #else
525 #define splatq_lane_p8(__p0, __p1) __extension__ ({ \
526  poly8x8_t __s0 = __p0; \
527  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
528  poly8x16_t __ret; \
529  __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 4); \
530  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
531  __ret; \
532 })
533 #define __noswap_splatq_lane_p8(__p0, __p1) __extension__ ({ \
534  poly8x8_t __s0 = __p0; \
535  poly8x16_t __ret; \
536  __ret = (poly8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 4); \
537  __ret; \
538 })
539 #endif
540 
541 #ifdef __LITTLE_ENDIAN__
542 #define splatq_lane_p64(__p0, __p1) __extension__ ({ \
543  poly64x1_t __s0 = __p0; \
544  poly64x2_t __ret; \
545  __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
546  __ret; \
547 })
548 #else
549 #define splatq_lane_p64(__p0, __p1) __extension__ ({ \
550  poly64x1_t __s0 = __p0; \
551  poly64x2_t __ret; \
552  __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
553  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
554  __ret; \
555 })
556 #define __noswap_splatq_lane_p64(__p0, __p1) __extension__ ({ \
557  poly64x1_t __s0 = __p0; \
558  poly64x2_t __ret; \
559  __ret = (poly64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 6); \
560  __ret; \
561 })
562 #endif
563 
564 #ifdef __LITTLE_ENDIAN__
565 #define splatq_lane_p16(__p0, __p1) __extension__ ({ \
566  poly16x4_t __s0 = __p0; \
567  poly16x8_t __ret; \
568  __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 5); \
569  __ret; \
570 })
571 #else
572 #define splatq_lane_p16(__p0, __p1) __extension__ ({ \
573  poly16x4_t __s0 = __p0; \
574  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
575  poly16x8_t __ret; \
576  __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 5); \
577  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
578  __ret; \
579 })
580 #define __noswap_splatq_lane_p16(__p0, __p1) __extension__ ({ \
581  poly16x4_t __s0 = __p0; \
582  poly16x8_t __ret; \
583  __ret = (poly16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 5); \
584  __ret; \
585 })
586 #endif
587 
588 #ifdef __LITTLE_ENDIAN__
589 #define splatq_lane_u8(__p0, __p1) __extension__ ({ \
590  uint8x8_t __s0 = __p0; \
591  uint8x16_t __ret; \
592  __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 16); \
593  __ret; \
594 })
595 #else
596 #define splatq_lane_u8(__p0, __p1) __extension__ ({ \
597  uint8x8_t __s0 = __p0; \
598  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
599  uint8x16_t __ret; \
600  __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 16); \
601  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
602  __ret; \
603 })
604 #define __noswap_splatq_lane_u8(__p0, __p1) __extension__ ({ \
605  uint8x8_t __s0 = __p0; \
606  uint8x16_t __ret; \
607  __ret = (uint8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 16); \
608  __ret; \
609 })
610 #endif
611 
612 #ifdef __LITTLE_ENDIAN__
613 #define splatq_lane_u32(__p0, __p1) __extension__ ({ \
614  uint32x2_t __s0 = __p0; \
615  uint32x4_t __ret; \
616  __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 18); \
617  __ret; \
618 })
619 #else
620 #define splatq_lane_u32(__p0, __p1) __extension__ ({ \
621  uint32x2_t __s0 = __p0; \
622  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
623  uint32x4_t __ret; \
624  __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 18); \
625  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
626  __ret; \
627 })
628 #define __noswap_splatq_lane_u32(__p0, __p1) __extension__ ({ \
629  uint32x2_t __s0 = __p0; \
630  uint32x4_t __ret; \
631  __ret = (uint32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 18); \
632  __ret; \
633 })
634 #endif
635 
636 #ifdef __LITTLE_ENDIAN__
637 #define splatq_lane_u64(__p0, __p1) __extension__ ({ \
638  uint64x1_t __s0 = __p0; \
639  uint64x2_t __ret; \
640  __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
641  __ret; \
642 })
643 #else
644 #define splatq_lane_u64(__p0, __p1) __extension__ ({ \
645  uint64x1_t __s0 = __p0; \
646  uint64x2_t __ret; \
647  __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
648  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
649  __ret; \
650 })
651 #define __noswap_splatq_lane_u64(__p0, __p1) __extension__ ({ \
652  uint64x1_t __s0 = __p0; \
653  uint64x2_t __ret; \
654  __ret = (uint64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 19); \
655  __ret; \
656 })
657 #endif
658 
659 #ifdef __LITTLE_ENDIAN__
660 #define splatq_lane_u16(__p0, __p1) __extension__ ({ \
661  uint16x4_t __s0 = __p0; \
662  uint16x8_t __ret; \
663  __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 17); \
664  __ret; \
665 })
666 #else
667 #define splatq_lane_u16(__p0, __p1) __extension__ ({ \
668  uint16x4_t __s0 = __p0; \
669  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
670  uint16x8_t __ret; \
671  __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 17); \
672  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
673  __ret; \
674 })
675 #define __noswap_splatq_lane_u16(__p0, __p1) __extension__ ({ \
676  uint16x4_t __s0 = __p0; \
677  uint16x8_t __ret; \
678  __ret = (uint16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 17); \
679  __ret; \
680 })
681 #endif
682 
683 #ifdef __LITTLE_ENDIAN__
684 #define splatq_lane_s8(__p0, __p1) __extension__ ({ \
685  int8x8_t __s0 = __p0; \
686  int8x16_t __ret; \
687  __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 0); \
688  __ret; \
689 })
690 #else
691 #define splatq_lane_s8(__p0, __p1) __extension__ ({ \
692  int8x8_t __s0 = __p0; \
693  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
694  int8x16_t __ret; \
695  __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 0); \
696  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
697  __ret; \
698 })
699 #define __noswap_splatq_lane_s8(__p0, __p1) __extension__ ({ \
700  int8x8_t __s0 = __p0; \
701  int8x16_t __ret; \
702  __ret = (int8x16_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 0); \
703  __ret; \
704 })
705 #endif
706 
707 #ifdef __LITTLE_ENDIAN__
708 #define splatq_lane_f64(__p0, __p1) __extension__ ({ \
709  float64x1_t __s0 = __p0; \
710  float64x2_t __ret; \
711  __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
712  __ret; \
713 })
714 #else
715 #define splatq_lane_f64(__p0, __p1) __extension__ ({ \
716  float64x1_t __s0 = __p0; \
717  float64x2_t __ret; \
718  __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
720  __ret; \
721 })
722 #define __noswap_splatq_lane_f64(__p0, __p1) __extension__ ({ \
723  float64x1_t __s0 = __p0; \
724  float64x2_t __ret; \
725  __ret = (float64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 10); \
726  __ret; \
727 })
728 #endif
729 
730 #ifdef __LITTLE_ENDIAN__
731 #define splatq_lane_f32(__p0, __p1) __extension__ ({ \
732  float32x2_t __s0 = __p0; \
733  float32x4_t __ret; \
734  __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 9); \
735  __ret; \
736 })
737 #else
738 #define splatq_lane_f32(__p0, __p1) __extension__ ({ \
739  float32x2_t __s0 = __p0; \
740  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
741  float32x4_t __ret; \
742  __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 9); \
743  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
744  __ret; \
745 })
746 #define __noswap_splatq_lane_f32(__p0, __p1) __extension__ ({ \
747  float32x2_t __s0 = __p0; \
748  float32x4_t __ret; \
749  __ret = (float32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 9); \
750  __ret; \
751 })
752 #endif
753 
754 #ifdef __LITTLE_ENDIAN__
755 #define splatq_lane_f16(__p0, __p1) __extension__ ({ \
756  float16x4_t __s0 = __p0; \
757  float16x8_t __ret; \
758  __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 8); \
759  __ret; \
760 })
761 #else
762 #define splatq_lane_f16(__p0, __p1) __extension__ ({ \
763  float16x4_t __s0 = __p0; \
764  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
765  float16x8_t __ret; \
766  __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 8); \
767  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
768  __ret; \
769 })
770 #define __noswap_splatq_lane_f16(__p0, __p1) __extension__ ({ \
771  float16x4_t __s0 = __p0; \
772  float16x8_t __ret; \
773  __ret = (float16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 8); \
774  __ret; \
775 })
776 #endif
777 
778 #ifdef __LITTLE_ENDIAN__
779 #define splatq_lane_s32(__p0, __p1) __extension__ ({ \
780  int32x2_t __s0 = __p0; \
781  int32x4_t __ret; \
782  __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 2); \
783  __ret; \
784 })
785 #else
786 #define splatq_lane_s32(__p0, __p1) __extension__ ({ \
787  int32x2_t __s0 = __p0; \
788  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
789  int32x4_t __ret; \
790  __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 2); \
791  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
792  __ret; \
793 })
794 #define __noswap_splatq_lane_s32(__p0, __p1) __extension__ ({ \
795  int32x2_t __s0 = __p0; \
796  int32x4_t __ret; \
797  __ret = (int32x4_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 2); \
798  __ret; \
799 })
800 #endif
801 
802 #ifdef __LITTLE_ENDIAN__
803 #define splatq_lane_s64(__p0, __p1) __extension__ ({ \
804  int64x1_t __s0 = __p0; \
805  int64x2_t __ret; \
806  __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
807  __ret; \
808 })
809 #else
810 #define splatq_lane_s64(__p0, __p1) __extension__ ({ \
811  int64x1_t __s0 = __p0; \
812  int64x2_t __ret; \
813  __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
814  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
815  __ret; \
816 })
817 #define __noswap_splatq_lane_s64(__p0, __p1) __extension__ ({ \
818  int64x1_t __s0 = __p0; \
819  int64x2_t __ret; \
820  __ret = (int64x2_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 3); \
821  __ret; \
822 })
823 #endif
824 
825 #ifdef __LITTLE_ENDIAN__
826 #define splatq_lane_s16(__p0, __p1) __extension__ ({ \
827  int16x4_t __s0 = __p0; \
828  int16x8_t __ret; \
829  __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 1); \
830  __ret; \
831 })
832 #else
833 #define splatq_lane_s16(__p0, __p1) __extension__ ({ \
834  int16x4_t __s0 = __p0; \
835  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
836  int16x8_t __ret; \
837  __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 1); \
838  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
839  __ret; \
840 })
841 #define __noswap_splatq_lane_s16(__p0, __p1) __extension__ ({ \
842  int16x4_t __s0 = __p0; \
843  int16x8_t __ret; \
844  __ret = (int16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 1); \
845  __ret; \
846 })
847 #endif
848 
849 #ifdef __LITTLE_ENDIAN__
850 #define splat_lane_u8(__p0, __p1) __extension__ ({ \
851  uint8x8_t __s0 = __p0; \
852  uint8x8_t __ret; \
853  __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 16); \
854  __ret; \
855 })
856 #else
857 #define splat_lane_u8(__p0, __p1) __extension__ ({ \
858  uint8x8_t __s0 = __p0; \
859  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
860  uint8x8_t __ret; \
861  __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 16); \
862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
863  __ret; \
864 })
865 #define __noswap_splat_lane_u8(__p0, __p1) __extension__ ({ \
866  uint8x8_t __s0 = __p0; \
867  uint8x8_t __ret; \
868  __ret = (uint8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 16); \
869  __ret; \
870 })
871 #endif
872 
873 #ifdef __LITTLE_ENDIAN__
874 #define splat_lane_u32(__p0, __p1) __extension__ ({ \
875  uint32x2_t __s0 = __p0; \
876  uint32x2_t __ret; \
877  __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 18); \
878  __ret; \
879 })
880 #else
881 #define splat_lane_u32(__p0, __p1) __extension__ ({ \
882  uint32x2_t __s0 = __p0; \
883  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
884  uint32x2_t __ret; \
885  __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 18); \
886  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
887  __ret; \
888 })
889 #define __noswap_splat_lane_u32(__p0, __p1) __extension__ ({ \
890  uint32x2_t __s0 = __p0; \
891  uint32x2_t __ret; \
892  __ret = (uint32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 18); \
893  __ret; \
894 })
895 #endif
896 
897 #define splat_lane_u64(__p0, __p1) __extension__ ({ \
898  uint64x1_t __s0 = __p0; \
899  uint64x1_t __ret; \
900  __ret = (uint64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 19); \
901  __ret; \
902 })
903 #ifdef __LITTLE_ENDIAN__
904 #define splat_lane_u16(__p0, __p1) __extension__ ({ \
905  uint16x4_t __s0 = __p0; \
906  uint16x4_t __ret; \
907  __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 17); \
908  __ret; \
909 })
910 #else
911 #define splat_lane_u16(__p0, __p1) __extension__ ({ \
912  uint16x4_t __s0 = __p0; \
913  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
914  uint16x4_t __ret; \
915  __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 17); \
916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
917  __ret; \
918 })
919 #define __noswap_splat_lane_u16(__p0, __p1) __extension__ ({ \
920  uint16x4_t __s0 = __p0; \
921  uint16x4_t __ret; \
922  __ret = (uint16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 17); \
923  __ret; \
924 })
925 #endif
926 
927 #ifdef __LITTLE_ENDIAN__
928 #define splat_lane_s8(__p0, __p1) __extension__ ({ \
929  int8x8_t __s0 = __p0; \
930  int8x8_t __ret; \
931  __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 0); \
932  __ret; \
933 })
934 #else
935 #define splat_lane_s8(__p0, __p1) __extension__ ({ \
936  int8x8_t __s0 = __p0; \
937  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
938  int8x8_t __ret; \
939  __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 0); \
940  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
941  __ret; \
942 })
943 #define __noswap_splat_lane_s8(__p0, __p1) __extension__ ({ \
944  int8x8_t __s0 = __p0; \
945  int8x8_t __ret; \
946  __ret = (int8x8_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 0); \
947  __ret; \
948 })
949 #endif
950 
951 #define splat_lane_f64(__p0, __p1) __extension__ ({ \
952  float64x1_t __s0 = __p0; \
953  float64x1_t __ret; \
954  __ret = (float64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 10); \
955  __ret; \
956 })
957 #ifdef __LITTLE_ENDIAN__
958 #define splat_lane_f32(__p0, __p1) __extension__ ({ \
959  float32x2_t __s0 = __p0; \
960  float32x2_t __ret; \
961  __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 9); \
962  __ret; \
963 })
964 #else
965 #define splat_lane_f32(__p0, __p1) __extension__ ({ \
966  float32x2_t __s0 = __p0; \
967  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
968  float32x2_t __ret; \
969  __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 9); \
970  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
971  __ret; \
972 })
973 #define __noswap_splat_lane_f32(__p0, __p1) __extension__ ({ \
974  float32x2_t __s0 = __p0; \
975  float32x2_t __ret; \
976  __ret = (float32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 9); \
977  __ret; \
978 })
979 #endif
980 
981 #ifdef __LITTLE_ENDIAN__
982 #define splat_lane_f16(__p0, __p1) __extension__ ({ \
983  float16x4_t __s0 = __p0; \
984  float16x4_t __ret; \
985  __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 8); \
986  __ret; \
987 })
988 #else
989 #define splat_lane_f16(__p0, __p1) __extension__ ({ \
990  float16x4_t __s0 = __p0; \
991  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
992  float16x4_t __ret; \
993  __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 8); \
994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
995  __ret; \
996 })
997 #define __noswap_splat_lane_f16(__p0, __p1) __extension__ ({ \
998  float16x4_t __s0 = __p0; \
999  float16x4_t __ret; \
1000  __ret = (float16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 8); \
1001  __ret; \
1002 })
1003 #endif
1004 
1005 #ifdef __LITTLE_ENDIAN__
1006 #define splat_lane_s32(__p0, __p1) __extension__ ({ \
1007  int32x2_t __s0 = __p0; \
1008  int32x2_t __ret; \
1009  __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 2); \
1010  __ret; \
1011 })
1012 #else
1013 #define splat_lane_s32(__p0, __p1) __extension__ ({ \
1014  int32x2_t __s0 = __p0; \
1015  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1016  int32x2_t __ret; \
1017  __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 2); \
1018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1019  __ret; \
1020 })
1021 #define __noswap_splat_lane_s32(__p0, __p1) __extension__ ({ \
1022  int32x2_t __s0 = __p0; \
1023  int32x2_t __ret; \
1024  __ret = (int32x2_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 2); \
1025  __ret; \
1026 })
1027 #endif
1028 
1029 #define splat_lane_s64(__p0, __p1) __extension__ ({ \
1030  int64x1_t __s0 = __p0; \
1031  int64x1_t __ret; \
1032  __ret = (int64x1_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 3); \
1033  __ret; \
1034 })
1035 #ifdef __LITTLE_ENDIAN__
1036 #define splat_lane_s16(__p0, __p1) __extension__ ({ \
1037  int16x4_t __s0 = __p0; \
1038  int16x4_t __ret; \
1039  __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 1); \
1040  __ret; \
1041 })
1042 #else
1043 #define splat_lane_s16(__p0, __p1) __extension__ ({ \
1044  int16x4_t __s0 = __p0; \
1045  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1046  int16x4_t __ret; \
1047  __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 1); \
1048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1049  __ret; \
1050 })
1051 #define __noswap_splat_lane_s16(__p0, __p1) __extension__ ({ \
1052  int16x4_t __s0 = __p0; \
1053  int16x4_t __ret; \
1054  __ret = (int16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 1); \
1055  __ret; \
1056 })
1057 #endif
1058 
1059 #ifdef __LITTLE_ENDIAN__
1060 #define splat_laneq_p8(__p0, __p1) __extension__ ({ \
1061  poly8x16_t __s0 = __p0; \
1062  poly8x8_t __ret; \
1063  __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 36); \
1064  __ret; \
1065 })
1066 #else
1067 #define splat_laneq_p8(__p0, __p1) __extension__ ({ \
1068  poly8x16_t __s0 = __p0; \
1069  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1070  poly8x8_t __ret; \
1071  __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 36); \
1072  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1073  __ret; \
1074 })
1075 #define __noswap_splat_laneq_p8(__p0, __p1) __extension__ ({ \
1076  poly8x16_t __s0 = __p0; \
1077  poly8x8_t __ret; \
1078  __ret = (poly8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 36); \
1079  __ret; \
1080 })
1081 #endif
1082 
1083 #ifdef __LITTLE_ENDIAN__
1084 #define splat_laneq_p64(__p0, __p1) __extension__ ({ \
1085  poly64x2_t __s0 = __p0; \
1086  poly64x1_t __ret; \
1087  __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 38); \
1088  __ret; \
1089 })
1090 #else
1091 #define splat_laneq_p64(__p0, __p1) __extension__ ({ \
1092  poly64x2_t __s0 = __p0; \
1093  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1094  poly64x1_t __ret; \
1095  __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 38); \
1096  __ret; \
1097 })
1098 #define __noswap_splat_laneq_p64(__p0, __p1) __extension__ ({ \
1099  poly64x2_t __s0 = __p0; \
1100  poly64x1_t __ret; \
1101  __ret = (poly64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 38); \
1102  __ret; \
1103 })
1104 #endif
1105 
1106 #ifdef __LITTLE_ENDIAN__
1107 #define splat_laneq_p16(__p0, __p1) __extension__ ({ \
1108  poly16x8_t __s0 = __p0; \
1109  poly16x4_t __ret; \
1110  __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 37); \
1111  __ret; \
1112 })
1113 #else
1114 #define splat_laneq_p16(__p0, __p1) __extension__ ({ \
1115  poly16x8_t __s0 = __p0; \
1116  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1117  poly16x4_t __ret; \
1118  __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 37); \
1119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1120  __ret; \
1121 })
1122 #define __noswap_splat_laneq_p16(__p0, __p1) __extension__ ({ \
1123  poly16x8_t __s0 = __p0; \
1124  poly16x4_t __ret; \
1125  __ret = (poly16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 37); \
1126  __ret; \
1127 })
1128 #endif
1129 
1130 #ifdef __LITTLE_ENDIAN__
1131 #define splatq_laneq_p8(__p0, __p1) __extension__ ({ \
1132  poly8x16_t __s0 = __p0; \
1133  poly8x16_t __ret; \
1134  __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 36); \
1135  __ret; \
1136 })
1137 #else
1138 #define splatq_laneq_p8(__p0, __p1) __extension__ ({ \
1139  poly8x16_t __s0 = __p0; \
1140  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1141  poly8x16_t __ret; \
1142  __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 36); \
1143  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1144  __ret; \
1145 })
1146 #define __noswap_splatq_laneq_p8(__p0, __p1) __extension__ ({ \
1147  poly8x16_t __s0 = __p0; \
1148  poly8x16_t __ret; \
1149  __ret = (poly8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 36); \
1150  __ret; \
1151 })
1152 #endif
1153 
1154 #ifdef __LITTLE_ENDIAN__
1155 #define splatq_laneq_p64(__p0, __p1) __extension__ ({ \
1156  poly64x2_t __s0 = __p0; \
1157  poly64x2_t __ret; \
1158  __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 38); \
1159  __ret; \
1160 })
1161 #else
1162 #define splatq_laneq_p64(__p0, __p1) __extension__ ({ \
1163  poly64x2_t __s0 = __p0; \
1164  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1165  poly64x2_t __ret; \
1166  __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 38); \
1167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1168  __ret; \
1169 })
1170 #define __noswap_splatq_laneq_p64(__p0, __p1) __extension__ ({ \
1171  poly64x2_t __s0 = __p0; \
1172  poly64x2_t __ret; \
1173  __ret = (poly64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 38); \
1174  __ret; \
1175 })
1176 #endif
1177 
1178 #ifdef __LITTLE_ENDIAN__
1179 #define splatq_laneq_p16(__p0, __p1) __extension__ ({ \
1180  poly16x8_t __s0 = __p0; \
1181  poly16x8_t __ret; \
1182  __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 37); \
1183  __ret; \
1184 })
1185 #else
1186 #define splatq_laneq_p16(__p0, __p1) __extension__ ({ \
1187  poly16x8_t __s0 = __p0; \
1188  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1189  poly16x8_t __ret; \
1190  __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 37); \
1191  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1192  __ret; \
1193 })
1194 #define __noswap_splatq_laneq_p16(__p0, __p1) __extension__ ({ \
1195  poly16x8_t __s0 = __p0; \
1196  poly16x8_t __ret; \
1197  __ret = (poly16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 37); \
1198  __ret; \
1199 })
1200 #endif
1201 
1202 #ifdef __LITTLE_ENDIAN__
1203 #define splatq_laneq_u8(__p0, __p1) __extension__ ({ \
1204  uint8x16_t __s0 = __p0; \
1205  uint8x16_t __ret; \
1206  __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 48); \
1207  __ret; \
1208 })
1209 #else
1210 #define splatq_laneq_u8(__p0, __p1) __extension__ ({ \
1211  uint8x16_t __s0 = __p0; \
1212  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1213  uint8x16_t __ret; \
1214  __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 48); \
1215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1216  __ret; \
1217 })
1218 #define __noswap_splatq_laneq_u8(__p0, __p1) __extension__ ({ \
1219  uint8x16_t __s0 = __p0; \
1220  uint8x16_t __ret; \
1221  __ret = (uint8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 48); \
1222  __ret; \
1223 })
1224 #endif
1225 
1226 #ifdef __LITTLE_ENDIAN__
1227 #define splatq_laneq_u32(__p0, __p1) __extension__ ({ \
1228  uint32x4_t __s0 = __p0; \
1229  uint32x4_t __ret; \
1230  __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 50); \
1231  __ret; \
1232 })
1233 #else
1234 #define splatq_laneq_u32(__p0, __p1) __extension__ ({ \
1235  uint32x4_t __s0 = __p0; \
1236  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1237  uint32x4_t __ret; \
1238  __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 50); \
1239  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1240  __ret; \
1241 })
1242 #define __noswap_splatq_laneq_u32(__p0, __p1) __extension__ ({ \
1243  uint32x4_t __s0 = __p0; \
1244  uint32x4_t __ret; \
1245  __ret = (uint32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 50); \
1246  __ret; \
1247 })
1248 #endif
1249 
1250 #ifdef __LITTLE_ENDIAN__
1251 #define splatq_laneq_u64(__p0, __p1) __extension__ ({ \
1252  uint64x2_t __s0 = __p0; \
1253  uint64x2_t __ret; \
1254  __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 51); \
1255  __ret; \
1256 })
1257 #else
1258 #define splatq_laneq_u64(__p0, __p1) __extension__ ({ \
1259  uint64x2_t __s0 = __p0; \
1260  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1261  uint64x2_t __ret; \
1262  __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 51); \
1263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1264  __ret; \
1265 })
1266 #define __noswap_splatq_laneq_u64(__p0, __p1) __extension__ ({ \
1267  uint64x2_t __s0 = __p0; \
1268  uint64x2_t __ret; \
1269  __ret = (uint64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 51); \
1270  __ret; \
1271 })
1272 #endif
1273 
1274 #ifdef __LITTLE_ENDIAN__
1275 #define splatq_laneq_u16(__p0, __p1) __extension__ ({ \
1276  uint16x8_t __s0 = __p0; \
1277  uint16x8_t __ret; \
1278  __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 49); \
1279  __ret; \
1280 })
1281 #else
1282 #define splatq_laneq_u16(__p0, __p1) __extension__ ({ \
1283  uint16x8_t __s0 = __p0; \
1284  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1285  uint16x8_t __ret; \
1286  __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 49); \
1287  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1288  __ret; \
1289 })
1290 #define __noswap_splatq_laneq_u16(__p0, __p1) __extension__ ({ \
1291  uint16x8_t __s0 = __p0; \
1292  uint16x8_t __ret; \
1293  __ret = (uint16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 49); \
1294  __ret; \
1295 })
1296 #endif
1297 
1298 #ifdef __LITTLE_ENDIAN__
1299 #define splatq_laneq_s8(__p0, __p1) __extension__ ({ \
1300  int8x16_t __s0 = __p0; \
1301  int8x16_t __ret; \
1302  __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 32); \
1303  __ret; \
1304 })
1305 #else
1306 #define splatq_laneq_s8(__p0, __p1) __extension__ ({ \
1307  int8x16_t __s0 = __p0; \
1308  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1309  int8x16_t __ret; \
1310  __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 32); \
1311  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1312  __ret; \
1313 })
1314 #define __noswap_splatq_laneq_s8(__p0, __p1) __extension__ ({ \
1315  int8x16_t __s0 = __p0; \
1316  int8x16_t __ret; \
1317  __ret = (int8x16_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 32); \
1318  __ret; \
1319 })
1320 #endif
1321 
1322 #ifdef __LITTLE_ENDIAN__
1323 #define splatq_laneq_f64(__p0, __p1) __extension__ ({ \
1324  float64x2_t __s0 = __p0; \
1325  float64x2_t __ret; \
1326  __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 42); \
1327  __ret; \
1328 })
1329 #else
1330 #define splatq_laneq_f64(__p0, __p1) __extension__ ({ \
1331  float64x2_t __s0 = __p0; \
1332  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1333  float64x2_t __ret; \
1334  __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 42); \
1335  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1336  __ret; \
1337 })
1338 #define __noswap_splatq_laneq_f64(__p0, __p1) __extension__ ({ \
1339  float64x2_t __s0 = __p0; \
1340  float64x2_t __ret; \
1341  __ret = (float64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 42); \
1342  __ret; \
1343 })
1344 #endif
1345 
1346 #ifdef __LITTLE_ENDIAN__
1347 #define splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1348  float32x4_t __s0 = __p0; \
1349  float32x4_t __ret; \
1350  __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 41); \
1351  __ret; \
1352 })
1353 #else
1354 #define splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1355  float32x4_t __s0 = __p0; \
1356  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1357  float32x4_t __ret; \
1358  __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 41); \
1359  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1360  __ret; \
1361 })
1362 #define __noswap_splatq_laneq_f32(__p0, __p1) __extension__ ({ \
1363  float32x4_t __s0 = __p0; \
1364  float32x4_t __ret; \
1365  __ret = (float32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 41); \
1366  __ret; \
1367 })
1368 #endif
1369 
1370 #ifdef __LITTLE_ENDIAN__
1371 #define splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1372  float16x8_t __s0 = __p0; \
1373  float16x8_t __ret; \
1374  __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 40); \
1375  __ret; \
1376 })
1377 #else
1378 #define splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1379  float16x8_t __s0 = __p0; \
1380  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1381  float16x8_t __ret; \
1382  __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 40); \
1383  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1384  __ret; \
1385 })
1386 #define __noswap_splatq_laneq_f16(__p0, __p1) __extension__ ({ \
1387  float16x8_t __s0 = __p0; \
1388  float16x8_t __ret; \
1389  __ret = (float16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 40); \
1390  __ret; \
1391 })
1392 #endif
1393 
1394 #ifdef __LITTLE_ENDIAN__
1395 #define splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1396  int32x4_t __s0 = __p0; \
1397  int32x4_t __ret; \
1398  __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 34); \
1399  __ret; \
1400 })
1401 #else
1402 #define splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1403  int32x4_t __s0 = __p0; \
1404  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1405  int32x4_t __ret; \
1406  __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 34); \
1407  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1408  __ret; \
1409 })
1410 #define __noswap_splatq_laneq_s32(__p0, __p1) __extension__ ({ \
1411  int32x4_t __s0 = __p0; \
1412  int32x4_t __ret; \
1413  __ret = (int32x4_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 34); \
1414  __ret; \
1415 })
1416 #endif
1417 
1418 #ifdef __LITTLE_ENDIAN__
1419 #define splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1420  int64x2_t __s0 = __p0; \
1421  int64x2_t __ret; \
1422  __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 35); \
1423  __ret; \
1424 })
1425 #else
1426 #define splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1427  int64x2_t __s0 = __p0; \
1428  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1429  int64x2_t __ret; \
1430  __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 35); \
1431  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1432  __ret; \
1433 })
1434 #define __noswap_splatq_laneq_s64(__p0, __p1) __extension__ ({ \
1435  int64x2_t __s0 = __p0; \
1436  int64x2_t __ret; \
1437  __ret = (int64x2_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 35); \
1438  __ret; \
1439 })
1440 #endif
1441 
1442 #ifdef __LITTLE_ENDIAN__
1443 #define splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1444  int16x8_t __s0 = __p0; \
1445  int16x8_t __ret; \
1446  __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 33); \
1447  __ret; \
1448 })
1449 #else
1450 #define splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1451  int16x8_t __s0 = __p0; \
1452  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1453  int16x8_t __ret; \
1454  __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 33); \
1455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1456  __ret; \
1457 })
1458 #define __noswap_splatq_laneq_s16(__p0, __p1) __extension__ ({ \
1459  int16x8_t __s0 = __p0; \
1460  int16x8_t __ret; \
1461  __ret = (int16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 33); \
1462  __ret; \
1463 })
1464 #endif
1465 
1466 #ifdef __LITTLE_ENDIAN__
1467 #define splat_laneq_u8(__p0, __p1) __extension__ ({ \
1468  uint8x16_t __s0 = __p0; \
1469  uint8x8_t __ret; \
1470  __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 48); \
1471  __ret; \
1472 })
1473 #else
1474 #define splat_laneq_u8(__p0, __p1) __extension__ ({ \
1475  uint8x16_t __s0 = __p0; \
1476  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1477  uint8x8_t __ret; \
1478  __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 48); \
1479  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1480  __ret; \
1481 })
1482 #define __noswap_splat_laneq_u8(__p0, __p1) __extension__ ({ \
1483  uint8x16_t __s0 = __p0; \
1484  uint8x8_t __ret; \
1485  __ret = (uint8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 48); \
1486  __ret; \
1487 })
1488 #endif
1489 
1490 #ifdef __LITTLE_ENDIAN__
1491 #define splat_laneq_u32(__p0, __p1) __extension__ ({ \
1492  uint32x4_t __s0 = __p0; \
1493  uint32x2_t __ret; \
1494  __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 50); \
1495  __ret; \
1496 })
1497 #else
1498 #define splat_laneq_u32(__p0, __p1) __extension__ ({ \
1499  uint32x4_t __s0 = __p0; \
1500  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1501  uint32x2_t __ret; \
1502  __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 50); \
1503  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1504  __ret; \
1505 })
1506 #define __noswap_splat_laneq_u32(__p0, __p1) __extension__ ({ \
1507  uint32x4_t __s0 = __p0; \
1508  uint32x2_t __ret; \
1509  __ret = (uint32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 50); \
1510  __ret; \
1511 })
1512 #endif
1513 
1514 #ifdef __LITTLE_ENDIAN__
1515 #define splat_laneq_u64(__p0, __p1) __extension__ ({ \
1516  uint64x2_t __s0 = __p0; \
1517  uint64x1_t __ret; \
1518  __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 51); \
1519  __ret; \
1520 })
1521 #else
1522 #define splat_laneq_u64(__p0, __p1) __extension__ ({ \
1523  uint64x2_t __s0 = __p0; \
1524  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1525  uint64x1_t __ret; \
1526  __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 51); \
1527  __ret; \
1528 })
1529 #define __noswap_splat_laneq_u64(__p0, __p1) __extension__ ({ \
1530  uint64x2_t __s0 = __p0; \
1531  uint64x1_t __ret; \
1532  __ret = (uint64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 51); \
1533  __ret; \
1534 })
1535 #endif
1536 
1537 #ifdef __LITTLE_ENDIAN__
1538 #define splat_laneq_u16(__p0, __p1) __extension__ ({ \
1539  uint16x8_t __s0 = __p0; \
1540  uint16x4_t __ret; \
1541  __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 49); \
1542  __ret; \
1543 })
1544 #else
1545 #define splat_laneq_u16(__p0, __p1) __extension__ ({ \
1546  uint16x8_t __s0 = __p0; \
1547  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1548  uint16x4_t __ret; \
1549  __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 49); \
1550  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1551  __ret; \
1552 })
1553 #define __noswap_splat_laneq_u16(__p0, __p1) __extension__ ({ \
1554  uint16x8_t __s0 = __p0; \
1555  uint16x4_t __ret; \
1556  __ret = (uint16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 49); \
1557  __ret; \
1558 })
1559 #endif
1560 
1561 #ifdef __LITTLE_ENDIAN__
1562 #define splat_laneq_s8(__p0, __p1) __extension__ ({ \
1563  int8x16_t __s0 = __p0; \
1564  int8x8_t __ret; \
1565  __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 32); \
1566  __ret; \
1567 })
1568 #else
1569 #define splat_laneq_s8(__p0, __p1) __extension__ ({ \
1570  int8x16_t __s0 = __p0; \
1571  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
1572  int8x8_t __ret; \
1573  __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 32); \
1574  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
1575  __ret; \
1576 })
1577 #define __noswap_splat_laneq_s8(__p0, __p1) __extension__ ({ \
1578  int8x16_t __s0 = __p0; \
1579  int8x8_t __ret; \
1580  __ret = (int8x8_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 32); \
1581  __ret; \
1582 })
1583 #endif
1584 
1585 #ifdef __LITTLE_ENDIAN__
1586 #define splat_laneq_f64(__p0, __p1) __extension__ ({ \
1587  float64x2_t __s0 = __p0; \
1588  float64x1_t __ret; \
1589  __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 42); \
1590  __ret; \
1591 })
1592 #else
1593 #define splat_laneq_f64(__p0, __p1) __extension__ ({ \
1594  float64x2_t __s0 = __p0; \
1595  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1596  float64x1_t __ret; \
1597  __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 42); \
1598  __ret; \
1599 })
1600 #define __noswap_splat_laneq_f64(__p0, __p1) __extension__ ({ \
1601  float64x2_t __s0 = __p0; \
1602  float64x1_t __ret; \
1603  __ret = (float64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 42); \
1604  __ret; \
1605 })
1606 #endif
1607 
1608 #ifdef __LITTLE_ENDIAN__
1609 #define splat_laneq_f32(__p0, __p1) __extension__ ({ \
1610  float32x4_t __s0 = __p0; \
1611  float32x2_t __ret; \
1612  __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 41); \
1613  __ret; \
1614 })
1615 #else
1616 #define splat_laneq_f32(__p0, __p1) __extension__ ({ \
1617  float32x4_t __s0 = __p0; \
1618  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1619  float32x2_t __ret; \
1620  __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 41); \
1621  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1622  __ret; \
1623 })
1624 #define __noswap_splat_laneq_f32(__p0, __p1) __extension__ ({ \
1625  float32x4_t __s0 = __p0; \
1626  float32x2_t __ret; \
1627  __ret = (float32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 41); \
1628  __ret; \
1629 })
1630 #endif
1631 
1632 #ifdef __LITTLE_ENDIAN__
1633 #define splat_laneq_f16(__p0, __p1) __extension__ ({ \
1634  float16x8_t __s0 = __p0; \
1635  float16x4_t __ret; \
1636  __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 40); \
1637  __ret; \
1638 })
1639 #else
1640 #define splat_laneq_f16(__p0, __p1) __extension__ ({ \
1641  float16x8_t __s0 = __p0; \
1642  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1643  float16x4_t __ret; \
1644  __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 40); \
1645  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1646  __ret; \
1647 })
1648 #define __noswap_splat_laneq_f16(__p0, __p1) __extension__ ({ \
1649  float16x8_t __s0 = __p0; \
1650  float16x4_t __ret; \
1651  __ret = (float16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 40); \
1652  __ret; \
1653 })
1654 #endif
1655 
1656 #ifdef __LITTLE_ENDIAN__
1657 #define splat_laneq_s32(__p0, __p1) __extension__ ({ \
1658  int32x4_t __s0 = __p0; \
1659  int32x2_t __ret; \
1660  __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 34); \
1661  __ret; \
1662 })
1663 #else
1664 #define splat_laneq_s32(__p0, __p1) __extension__ ({ \
1665  int32x4_t __s0 = __p0; \
1666  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
1667  int32x2_t __ret; \
1668  __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 34); \
1669  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
1670  __ret; \
1671 })
1672 #define __noswap_splat_laneq_s32(__p0, __p1) __extension__ ({ \
1673  int32x4_t __s0 = __p0; \
1674  int32x2_t __ret; \
1675  __ret = (int32x2_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 34); \
1676  __ret; \
1677 })
1678 #endif
1679 
1680 #ifdef __LITTLE_ENDIAN__
1681 #define splat_laneq_s64(__p0, __p1) __extension__ ({ \
1682  int64x2_t __s0 = __p0; \
1683  int64x1_t __ret; \
1684  __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 35); \
1685  __ret; \
1686 })
1687 #else
1688 #define splat_laneq_s64(__p0, __p1) __extension__ ({ \
1689  int64x2_t __s0 = __p0; \
1690  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
1691  int64x1_t __ret; \
1692  __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 35); \
1693  __ret; \
1694 })
1695 #define __noswap_splat_laneq_s64(__p0, __p1) __extension__ ({ \
1696  int64x2_t __s0 = __p0; \
1697  int64x1_t __ret; \
1698  __ret = (int64x1_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 35); \
1699  __ret; \
1700 })
1701 #endif
1702 
1703 #ifdef __LITTLE_ENDIAN__
1704 #define splat_laneq_s16(__p0, __p1) __extension__ ({ \
1705  int16x8_t __s0 = __p0; \
1706  int16x4_t __ret; \
1707  __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 33); \
1708  __ret; \
1709 })
1710 #else
1711 #define splat_laneq_s16(__p0, __p1) __extension__ ({ \
1712  int16x8_t __s0 = __p0; \
1713  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
1714  int16x4_t __ret; \
1715  __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 33); \
1716  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
1717  __ret; \
1718 })
1719 #define __noswap_splat_laneq_s16(__p0, __p1) __extension__ ({ \
1720  int16x8_t __s0 = __p0; \
1721  int16x4_t __ret; \
1722  __ret = (int16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 33); \
1723  __ret; \
1724 })
1725 #endif
1726 
1727 #ifdef __LITTLE_ENDIAN__
1728 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1729  uint8x16_t __ret;
1730  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
1731  return __ret;
1732 }
1733 #else
1734 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1735  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1736  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1737  uint8x16_t __ret;
1738  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
1739  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1740  return __ret;
1741 }
1742 __ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1743  uint8x16_t __ret;
1744  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
1745  return __ret;
1746 }
1747 #endif
1748 
1749 #ifdef __LITTLE_ENDIAN__
1750 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1751  uint32x4_t __ret;
1752  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
1753  return __ret;
1754 }
1755 #else
1756 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1757  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1758  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1759  uint32x4_t __ret;
1760  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
1761  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1762  return __ret;
1763 }
1764 __ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1765  uint32x4_t __ret;
1766  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
1767  return __ret;
1768 }
1769 #endif
1770 
1771 #ifdef __LITTLE_ENDIAN__
1772 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1773  uint16x8_t __ret;
1774  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
1775  return __ret;
1776 }
1777 #else
1778 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1779  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1780  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1781  uint16x8_t __ret;
1782  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
1783  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1784  return __ret;
1785 }
1786 __ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1787  uint16x8_t __ret;
1788  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
1789  return __ret;
1790 }
1791 #endif
1792 
1793 #ifdef __LITTLE_ENDIAN__
1794 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1795  int8x16_t __ret;
1796  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
1797  return __ret;
1798 }
1799 #else
1800 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1801  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1802  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1803  int8x16_t __ret;
1804  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
1805  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1806  return __ret;
1807 }
1808 __ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
1809  int8x16_t __ret;
1810  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
1811  return __ret;
1812 }
1813 #endif
1814 
1815 #ifdef __LITTLE_ENDIAN__
1816 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
1817  float32x4_t __ret;
1818  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
1819  return __ret;
1820 }
1821 #else
1822 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
1823  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1824  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1825  float32x4_t __ret;
1826  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
1827  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1828  return __ret;
1829 }
1830 #endif
1831 
1832 #ifdef __LITTLE_ENDIAN__
1833 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1834  int32x4_t __ret;
1835  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
1836  return __ret;
1837 }
1838 #else
1839 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1840  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1841  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1842  int32x4_t __ret;
1843  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
1844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1845  return __ret;
1846 }
1847 __ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
1848  int32x4_t __ret;
1849  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
1850  return __ret;
1851 }
1852 #endif
1853 
1854 #ifdef __LITTLE_ENDIAN__
1855 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1856  int16x8_t __ret;
1857  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
1858  return __ret;
1859 }
1860 #else
1861 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1862  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1863  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1864  int16x8_t __ret;
1865  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
1866  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1867  return __ret;
1868 }
1869 __ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
1870  int16x8_t __ret;
1871  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
1872  return __ret;
1873 }
1874 #endif
1875 
1876 #ifdef __LITTLE_ENDIAN__
1877 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1878  uint8x8_t __ret;
1879  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
1880  return __ret;
1881 }
1882 #else
1883 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1884  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1885  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1886  uint8x8_t __ret;
1887  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
1888  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1889  return __ret;
1890 }
1891 __ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1892  uint8x8_t __ret;
1893  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
1894  return __ret;
1895 }
1896 #endif
1897 
1898 #ifdef __LITTLE_ENDIAN__
1899 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1900  uint32x2_t __ret;
1901  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
1902  return __ret;
1903 }
1904 #else
1905 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1906  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1907  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1908  uint32x2_t __ret;
1909  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
1910  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1911  return __ret;
1912 }
1913 __ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1914  uint32x2_t __ret;
1915  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
1916  return __ret;
1917 }
1918 #endif
1919 
1920 #ifdef __LITTLE_ENDIAN__
1921 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1922  uint16x4_t __ret;
1923  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
1924  return __ret;
1925 }
1926 #else
1927 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1928  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1929  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1930  uint16x4_t __ret;
1931  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
1932  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1933  return __ret;
1934 }
1935 __ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1936  uint16x4_t __ret;
1937  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
1938  return __ret;
1939 }
1940 #endif
1941 
1942 #ifdef __LITTLE_ENDIAN__
1943 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1944  int8x8_t __ret;
1945  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
1946  return __ret;
1947 }
1948 #else
1949 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1950  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1951  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1952  int8x8_t __ret;
1953  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
1954  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1955  return __ret;
1956 }
1957 __ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
1958  int8x8_t __ret;
1959  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
1960  return __ret;
1961 }
1962 #endif
1963 
1964 #ifdef __LITTLE_ENDIAN__
1965 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
1966  float32x2_t __ret;
1967  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
1968  return __ret;
1969 }
1970 #else
1971 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
1972  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1973  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1974  float32x2_t __ret;
1975  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
1976  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1977  return __ret;
1978 }
1979 #endif
1980 
1981 #ifdef __LITTLE_ENDIAN__
1982 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1983  int32x2_t __ret;
1984  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
1985  return __ret;
1986 }
1987 #else
1988 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1989  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1990  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1991  int32x2_t __ret;
1992  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
1993  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1994  return __ret;
1995 }
1996 __ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
1997  int32x2_t __ret;
1998  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
1999  return __ret;
2000 }
2001 #endif
2002 
2003 #ifdef __LITTLE_ENDIAN__
2004 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
2005  int16x4_t __ret;
2006  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
2007  return __ret;
2008 }
2009 #else
2010 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
2011  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2012  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2013  int16x4_t __ret;
2014  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
2015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2016  return __ret;
2017 }
2018 __ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
2019  int16x4_t __ret;
2020  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
2021  return __ret;
2022 }
2023 #endif
2024 
2025 #ifdef __LITTLE_ENDIAN__
2026 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
2027  int8x16_t __ret;
2028  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
2029  return __ret;
2030 }
2031 #else
2032 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
2033  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2034  int8x16_t __ret;
2035  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
2036  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2037  return __ret;
2038 }
2039 #endif
2040 
2041 #ifdef __LITTLE_ENDIAN__
2042 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
2043  float32x4_t __ret;
2044  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
2045  return __ret;
2046 }
2047 #else
2048 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
2049  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2050  float32x4_t __ret;
2051  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
2052  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2053  return __ret;
2054 }
2055 #endif
2056 
2057 #ifdef __LITTLE_ENDIAN__
2058 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
2059  int32x4_t __ret;
2060  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
2061  return __ret;
2062 }
2063 #else
2064 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
2065  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2066  int32x4_t __ret;
2067  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
2068  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2069  return __ret;
2070 }
2071 #endif
2072 
2073 #ifdef __LITTLE_ENDIAN__
2074 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
2075  int16x8_t __ret;
2076  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
2077  return __ret;
2078 }
2079 #else
2080 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
2081  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2082  int16x8_t __ret;
2083  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
2084  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2085  return __ret;
2086 }
2087 #endif
2088 
2089 #ifdef __LITTLE_ENDIAN__
2090 __ai int8x8_t vabs_s8(int8x8_t __p0) {
2091  int8x8_t __ret;
2092  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
2093  return __ret;
2094 }
2095 #else
2096 __ai int8x8_t vabs_s8(int8x8_t __p0) {
2097  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2098  int8x8_t __ret;
2099  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
2100  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2101  return __ret;
2102 }
2103 #endif
2104 
2105 #ifdef __LITTLE_ENDIAN__
2106 __ai float32x2_t vabs_f32(float32x2_t __p0) {
2107  float32x2_t __ret;
2108  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
2109  return __ret;
2110 }
2111 #else
2112 __ai float32x2_t vabs_f32(float32x2_t __p0) {
2113  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2114  float32x2_t __ret;
2115  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
2116  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2117  return __ret;
2118 }
2119 #endif
2120 
2121 #ifdef __LITTLE_ENDIAN__
2122 __ai int32x2_t vabs_s32(int32x2_t __p0) {
2123  int32x2_t __ret;
2124  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
2125  return __ret;
2126 }
2127 #else
2128 __ai int32x2_t vabs_s32(int32x2_t __p0) {
2129  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2130  int32x2_t __ret;
2131  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
2132  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2133  return __ret;
2134 }
2135 #endif
2136 
2137 #ifdef __LITTLE_ENDIAN__
2138 __ai int16x4_t vabs_s16(int16x4_t __p0) {
2139  int16x4_t __ret;
2140  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
2141  return __ret;
2142 }
2143 #else
2144 __ai int16x4_t vabs_s16(int16x4_t __p0) {
2145  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2146  int16x4_t __ret;
2147  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
2148  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2149  return __ret;
2150 }
2151 #endif
2152 
2153 #ifdef __LITTLE_ENDIAN__
2154 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2155  uint8x16_t __ret;
2156  __ret = __p0 + __p1;
2157  return __ret;
2158 }
2159 #else
2160 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2161  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2162  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2163  uint8x16_t __ret;
2164  __ret = __rev0 + __rev1;
2165  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2166  return __ret;
2167 }
2168 #endif
2169 
2170 #ifdef __LITTLE_ENDIAN__
2171 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2172  uint32x4_t __ret;
2173  __ret = __p0 + __p1;
2174  return __ret;
2175 }
2176 #else
2177 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2178  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2179  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2180  uint32x4_t __ret;
2181  __ret = __rev0 + __rev1;
2182  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2183  return __ret;
2184 }
2185 #endif
2186 
2187 #ifdef __LITTLE_ENDIAN__
2188 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2189  uint64x2_t __ret;
2190  __ret = __p0 + __p1;
2191  return __ret;
2192 }
2193 #else
2194 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2195  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2196  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2197  uint64x2_t __ret;
2198  __ret = __rev0 + __rev1;
2199  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2200  return __ret;
2201 }
2202 #endif
2203 
2204 #ifdef __LITTLE_ENDIAN__
2205 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2206  uint16x8_t __ret;
2207  __ret = __p0 + __p1;
2208  return __ret;
2209 }
2210 #else
2211 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2212  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2213  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2214  uint16x8_t __ret;
2215  __ret = __rev0 + __rev1;
2216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2217  return __ret;
2218 }
2219 #endif
2220 
2221 #ifdef __LITTLE_ENDIAN__
2222 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
2223  int8x16_t __ret;
2224  __ret = __p0 + __p1;
2225  return __ret;
2226 }
2227 #else
2228 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
2229  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2230  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2231  int8x16_t __ret;
2232  __ret = __rev0 + __rev1;
2233  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2234  return __ret;
2235 }
2236 #endif
2237 
2238 #ifdef __LITTLE_ENDIAN__
2239 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
2240  float32x4_t __ret;
2241  __ret = __p0 + __p1;
2242  return __ret;
2243 }
2244 #else
2245 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
2246  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2247  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2248  float32x4_t __ret;
2249  __ret = __rev0 + __rev1;
2250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2251  return __ret;
2252 }
2253 #endif
2254 
2255 #ifdef __LITTLE_ENDIAN__
2256 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
2257  int32x4_t __ret;
2258  __ret = __p0 + __p1;
2259  return __ret;
2260 }
2261 #else
2262 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
2263  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2264  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2265  int32x4_t __ret;
2266  __ret = __rev0 + __rev1;
2267  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2268  return __ret;
2269 }
2270 #endif
2271 
2272 #ifdef __LITTLE_ENDIAN__
2273 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
2274  int64x2_t __ret;
2275  __ret = __p0 + __p1;
2276  return __ret;
2277 }
2278 #else
2279 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
2280  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2281  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2282  int64x2_t __ret;
2283  __ret = __rev0 + __rev1;
2284  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2285  return __ret;
2286 }
2287 #endif
2288 
2289 #ifdef __LITTLE_ENDIAN__
2290 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
2291  int16x8_t __ret;
2292  __ret = __p0 + __p1;
2293  return __ret;
2294 }
2295 #else
2296 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
2297  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2298  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2299  int16x8_t __ret;
2300  __ret = __rev0 + __rev1;
2301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2302  return __ret;
2303 }
2304 #endif
2305 
2306 #ifdef __LITTLE_ENDIAN__
2307 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
2308  uint8x8_t __ret;
2309  __ret = __p0 + __p1;
2310  return __ret;
2311 }
2312 #else
2313 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
2314  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2315  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2316  uint8x8_t __ret;
2317  __ret = __rev0 + __rev1;
2318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2319  return __ret;
2320 }
2321 #endif
2322 
2323 #ifdef __LITTLE_ENDIAN__
2324 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
2325  uint32x2_t __ret;
2326  __ret = __p0 + __p1;
2327  return __ret;
2328 }
2329 #else
2330 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
2331  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2332  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2333  uint32x2_t __ret;
2334  __ret = __rev0 + __rev1;
2335  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2336  return __ret;
2337 }
2338 #endif
2339 
2340 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
2341  uint64x1_t __ret;
2342  __ret = __p0 + __p1;
2343  return __ret;
2344 }
2345 #ifdef __LITTLE_ENDIAN__
2346 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
2347  uint16x4_t __ret;
2348  __ret = __p0 + __p1;
2349  return __ret;
2350 }
2351 #else
2352 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
2353  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2354  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2355  uint16x4_t __ret;
2356  __ret = __rev0 + __rev1;
2357  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2358  return __ret;
2359 }
2360 #endif
2361 
2362 #ifdef __LITTLE_ENDIAN__
2363 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
2364  int8x8_t __ret;
2365  __ret = __p0 + __p1;
2366  return __ret;
2367 }
2368 #else
2369 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
2370  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2371  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2372  int8x8_t __ret;
2373  __ret = __rev0 + __rev1;
2374  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2375  return __ret;
2376 }
2377 #endif
2378 
2379 #ifdef __LITTLE_ENDIAN__
2380 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
2381  float32x2_t __ret;
2382  __ret = __p0 + __p1;
2383  return __ret;
2384 }
2385 #else
2386 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
2387  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2388  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2389  float32x2_t __ret;
2390  __ret = __rev0 + __rev1;
2391  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2392  return __ret;
2393 }
2394 #endif
2395 
2396 #ifdef __LITTLE_ENDIAN__
2397 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
2398  int32x2_t __ret;
2399  __ret = __p0 + __p1;
2400  return __ret;
2401 }
2402 #else
2403 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
2404  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2405  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2406  int32x2_t __ret;
2407  __ret = __rev0 + __rev1;
2408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2409  return __ret;
2410 }
2411 #endif
2412 
2413 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
2414  int64x1_t __ret;
2415  __ret = __p0 + __p1;
2416  return __ret;
2417 }
2418 #ifdef __LITTLE_ENDIAN__
2419 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
2420  int16x4_t __ret;
2421  __ret = __p0 + __p1;
2422  return __ret;
2423 }
2424 #else
2425 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
2426  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2427  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2428  int16x4_t __ret;
2429  __ret = __rev0 + __rev1;
2430  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2431  return __ret;
2432 }
2433 #endif
2434 
2435 #ifdef __LITTLE_ENDIAN__
2436 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2437  uint16x4_t __ret;
2438  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
2439  return __ret;
2440 }
2441 #else
2442 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2443  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2444  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2445  uint16x4_t __ret;
2446  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
2447  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2448  return __ret;
2449 }
2450 __ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
2451  uint16x4_t __ret;
2452  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
2453  return __ret;
2454 }
2455 #endif
2456 
2457 #ifdef __LITTLE_ENDIAN__
2458 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2459  uint32x2_t __ret;
2460  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
2461  return __ret;
2462 }
2463 #else
2464 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2465  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2466  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2467  uint32x2_t __ret;
2468  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
2469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2470  return __ret;
2471 }
2472 __ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
2473  uint32x2_t __ret;
2474  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
2475  return __ret;
2476 }
2477 #endif
2478 
2479 #ifdef __LITTLE_ENDIAN__
2480 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2481  uint8x8_t __ret;
2482  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
2483  return __ret;
2484 }
2485 #else
2486 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2487  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2488  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2489  uint8x8_t __ret;
2490  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
2491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2492  return __ret;
2493 }
2494 __ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
2495  uint8x8_t __ret;
2496  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
2497  return __ret;
2498 }
2499 #endif
2500 
2501 #ifdef __LITTLE_ENDIAN__
2502 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2503  int16x4_t __ret;
2504  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
2505  return __ret;
2506 }
2507 #else
2508 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2509  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2510  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2511  int16x4_t __ret;
2512  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
2513  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2514  return __ret;
2515 }
2516 __ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
2517  int16x4_t __ret;
2518  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
2519  return __ret;
2520 }
2521 #endif
2522 
2523 #ifdef __LITTLE_ENDIAN__
2524 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2525  int32x2_t __ret;
2526  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
2527  return __ret;
2528 }
2529 #else
2530 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2531  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2532  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2533  int32x2_t __ret;
2534  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
2535  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2536  return __ret;
2537 }
2538 __ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
2539  int32x2_t __ret;
2540  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
2541  return __ret;
2542 }
2543 #endif
2544 
2545 #ifdef __LITTLE_ENDIAN__
2546 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2547  int8x8_t __ret;
2548  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
2549  return __ret;
2550 }
2551 #else
2552 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2553  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2554  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2555  int8x8_t __ret;
2556  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
2557  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2558  return __ret;
2559 }
2560 __ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
2561  int8x8_t __ret;
2562  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
2563  return __ret;
2564 }
2565 #endif
2566 
2567 #ifdef __LITTLE_ENDIAN__
2568 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2569  uint8x16_t __ret;
2570  __ret = __p0 & __p1;
2571  return __ret;
2572 }
2573 #else
2574 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2575  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2576  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2577  uint8x16_t __ret;
2578  __ret = __rev0 & __rev1;
2579  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2580  return __ret;
2581 }
2582 #endif
2583 
2584 #ifdef __LITTLE_ENDIAN__
2585 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2586  uint32x4_t __ret;
2587  __ret = __p0 & __p1;
2588  return __ret;
2589 }
2590 #else
2591 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2592  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2593  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2594  uint32x4_t __ret;
2595  __ret = __rev0 & __rev1;
2596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2597  return __ret;
2598 }
2599 #endif
2600 
2601 #ifdef __LITTLE_ENDIAN__
2602 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2603  uint64x2_t __ret;
2604  __ret = __p0 & __p1;
2605  return __ret;
2606 }
2607 #else
2608 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2609  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2610  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2611  uint64x2_t __ret;
2612  __ret = __rev0 & __rev1;
2613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2614  return __ret;
2615 }
2616 #endif
2617 
2618 #ifdef __LITTLE_ENDIAN__
2619 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2620  uint16x8_t __ret;
2621  __ret = __p0 & __p1;
2622  return __ret;
2623 }
2624 #else
2625 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2626  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2627  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2628  uint16x8_t __ret;
2629  __ret = __rev0 & __rev1;
2630  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2631  return __ret;
2632 }
2633 #endif
2634 
2635 #ifdef __LITTLE_ENDIAN__
2636 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
2637  int8x16_t __ret;
2638  __ret = __p0 & __p1;
2639  return __ret;
2640 }
2641 #else
2642 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
2643  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2644  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2645  int8x16_t __ret;
2646  __ret = __rev0 & __rev1;
2647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2648  return __ret;
2649 }
2650 #endif
2651 
2652 #ifdef __LITTLE_ENDIAN__
2653 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
2654  int32x4_t __ret;
2655  __ret = __p0 & __p1;
2656  return __ret;
2657 }
2658 #else
2659 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
2660  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2661  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2662  int32x4_t __ret;
2663  __ret = __rev0 & __rev1;
2664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2665  return __ret;
2666 }
2667 #endif
2668 
2669 #ifdef __LITTLE_ENDIAN__
2670 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
2671  int64x2_t __ret;
2672  __ret = __p0 & __p1;
2673  return __ret;
2674 }
2675 #else
2676 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
2677  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2678  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2679  int64x2_t __ret;
2680  __ret = __rev0 & __rev1;
2681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2682  return __ret;
2683 }
2684 #endif
2685 
2686 #ifdef __LITTLE_ENDIAN__
2687 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
2688  int16x8_t __ret;
2689  __ret = __p0 & __p1;
2690  return __ret;
2691 }
2692 #else
2693 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
2694  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2695  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2696  int16x8_t __ret;
2697  __ret = __rev0 & __rev1;
2698  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2699  return __ret;
2700 }
2701 #endif
2702 
2703 #ifdef __LITTLE_ENDIAN__
2704 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
2705  uint8x8_t __ret;
2706  __ret = __p0 & __p1;
2707  return __ret;
2708 }
2709 #else
2710 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
2711  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2712  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2713  uint8x8_t __ret;
2714  __ret = __rev0 & __rev1;
2715  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2716  return __ret;
2717 }
2718 #endif
2719 
2720 #ifdef __LITTLE_ENDIAN__
2721 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
2722  uint32x2_t __ret;
2723  __ret = __p0 & __p1;
2724  return __ret;
2725 }
2726 #else
2727 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
2728  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2729  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2730  uint32x2_t __ret;
2731  __ret = __rev0 & __rev1;
2732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2733  return __ret;
2734 }
2735 #endif
2736 
2737 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
2738  uint64x1_t __ret;
2739  __ret = __p0 & __p1;
2740  return __ret;
2741 }
2742 #ifdef __LITTLE_ENDIAN__
2743 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
2744  uint16x4_t __ret;
2745  __ret = __p0 & __p1;
2746  return __ret;
2747 }
2748 #else
2749 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
2750  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2751  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2752  uint16x4_t __ret;
2753  __ret = __rev0 & __rev1;
2754  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2755  return __ret;
2756 }
2757 #endif
2758 
2759 #ifdef __LITTLE_ENDIAN__
2760 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
2761  int8x8_t __ret;
2762  __ret = __p0 & __p1;
2763  return __ret;
2764 }
2765 #else
2766 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
2767  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2768  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2769  int8x8_t __ret;
2770  __ret = __rev0 & __rev1;
2771  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2772  return __ret;
2773 }
2774 #endif
2775 
2776 #ifdef __LITTLE_ENDIAN__
2777 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
2778  int32x2_t __ret;
2779  __ret = __p0 & __p1;
2780  return __ret;
2781 }
2782 #else
2783 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
2784  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2785  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2786  int32x2_t __ret;
2787  __ret = __rev0 & __rev1;
2788  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2789  return __ret;
2790 }
2791 #endif
2792 
2793 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
2794  int64x1_t __ret;
2795  __ret = __p0 & __p1;
2796  return __ret;
2797 }
2798 #ifdef __LITTLE_ENDIAN__
2799 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
2800  int16x4_t __ret;
2801  __ret = __p0 & __p1;
2802  return __ret;
2803 }
2804 #else
2805 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
2806  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2807  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2808  int16x4_t __ret;
2809  __ret = __rev0 & __rev1;
2810  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2811  return __ret;
2812 }
2813 #endif
2814 
2815 #ifdef __LITTLE_ENDIAN__
2816 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2817  uint8x16_t __ret;
2818  __ret = __p0 & ~__p1;
2819  return __ret;
2820 }
2821 #else
2822 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2823  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2824  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2825  uint8x16_t __ret;
2826  __ret = __rev0 & ~__rev1;
2827  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2828  return __ret;
2829 }
2830 #endif
2831 
2832 #ifdef __LITTLE_ENDIAN__
2833 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2834  uint32x4_t __ret;
2835  __ret = __p0 & ~__p1;
2836  return __ret;
2837 }
2838 #else
2839 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2840  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2841  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2842  uint32x4_t __ret;
2843  __ret = __rev0 & ~__rev1;
2844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2845  return __ret;
2846 }
2847 #endif
2848 
2849 #ifdef __LITTLE_ENDIAN__
2850 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2851  uint64x2_t __ret;
2852  __ret = __p0 & ~__p1;
2853  return __ret;
2854 }
2855 #else
2856 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
2857  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2858  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2859  uint64x2_t __ret;
2860  __ret = __rev0 & ~__rev1;
2861  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2862  return __ret;
2863 }
2864 #endif
2865 
2866 #ifdef __LITTLE_ENDIAN__
2867 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2868  uint16x8_t __ret;
2869  __ret = __p0 & ~__p1;
2870  return __ret;
2871 }
2872 #else
2873 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2874  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2875  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2876  uint16x8_t __ret;
2877  __ret = __rev0 & ~__rev1;
2878  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2879  return __ret;
2880 }
2881 #endif
2882 
2883 #ifdef __LITTLE_ENDIAN__
2884 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
2885  int8x16_t __ret;
2886  __ret = __p0 & ~__p1;
2887  return __ret;
2888 }
2889 #else
2890 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
2891  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2892  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2893  int8x16_t __ret;
2894  __ret = __rev0 & ~__rev1;
2895  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2896  return __ret;
2897 }
2898 #endif
2899 
2900 #ifdef __LITTLE_ENDIAN__
2901 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
2902  int32x4_t __ret;
2903  __ret = __p0 & ~__p1;
2904  return __ret;
2905 }
2906 #else
2907 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
2908  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2909  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2910  int32x4_t __ret;
2911  __ret = __rev0 & ~__rev1;
2912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2913  return __ret;
2914 }
2915 #endif
2916 
2917 #ifdef __LITTLE_ENDIAN__
2918 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
2919  int64x2_t __ret;
2920  __ret = __p0 & ~__p1;
2921  return __ret;
2922 }
2923 #else
2924 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
2925  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2926  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2927  int64x2_t __ret;
2928  __ret = __rev0 & ~__rev1;
2929  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2930  return __ret;
2931 }
2932 #endif
2933 
2934 #ifdef __LITTLE_ENDIAN__
2935 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
2936  int16x8_t __ret;
2937  __ret = __p0 & ~__p1;
2938  return __ret;
2939 }
2940 #else
2941 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
2942  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2943  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2944  int16x8_t __ret;
2945  __ret = __rev0 & ~__rev1;
2946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2947  return __ret;
2948 }
2949 #endif
2950 
2951 #ifdef __LITTLE_ENDIAN__
2952 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
2953  uint8x8_t __ret;
2954  __ret = __p0 & ~__p1;
2955  return __ret;
2956 }
2957 #else
2958 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
2959  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2960  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2961  uint8x8_t __ret;
2962  __ret = __rev0 & ~__rev1;
2963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2964  return __ret;
2965 }
2966 #endif
2967 
2968 #ifdef __LITTLE_ENDIAN__
2969 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
2970  uint32x2_t __ret;
2971  __ret = __p0 & ~__p1;
2972  return __ret;
2973 }
2974 #else
2975 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
2976  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2977  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2978  uint32x2_t __ret;
2979  __ret = __rev0 & ~__rev1;
2980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2981  return __ret;
2982 }
2983 #endif
2984 
2985 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
2986  uint64x1_t __ret;
2987  __ret = __p0 & ~__p1;
2988  return __ret;
2989 }
2990 #ifdef __LITTLE_ENDIAN__
2991 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
2992  uint16x4_t __ret;
2993  __ret = __p0 & ~__p1;
2994  return __ret;
2995 }
2996 #else
2997 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
2998  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2999  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3000  uint16x4_t __ret;
3001  __ret = __rev0 & ~__rev1;
3002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3003  return __ret;
3004 }
3005 #endif
3006 
3007 #ifdef __LITTLE_ENDIAN__
3008 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
3009  int8x8_t __ret;
3010  __ret = __p0 & ~__p1;
3011  return __ret;
3012 }
3013 #else
3014 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
3015  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3016  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3017  int8x8_t __ret;
3018  __ret = __rev0 & ~__rev1;
3019  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3020  return __ret;
3021 }
3022 #endif
3023 
3024 #ifdef __LITTLE_ENDIAN__
3025 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
3026  int32x2_t __ret;
3027  __ret = __p0 & ~__p1;
3028  return __ret;
3029 }
3030 #else
3031 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
3032  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3033  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3034  int32x2_t __ret;
3035  __ret = __rev0 & ~__rev1;
3036  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3037  return __ret;
3038 }
3039 #endif
3040 
3041 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
3042  int64x1_t __ret;
3043  __ret = __p0 & ~__p1;
3044  return __ret;
3045 }
3046 #ifdef __LITTLE_ENDIAN__
3047 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
3048  int16x4_t __ret;
3049  __ret = __p0 & ~__p1;
3050  return __ret;
3051 }
3052 #else
3053 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
3054  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3055  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3056  int16x4_t __ret;
3057  __ret = __rev0 & ~__rev1;
3058  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3059  return __ret;
3060 }
3061 #endif
3062 
3063 #ifdef __LITTLE_ENDIAN__
3064 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
3065  poly8x8_t __ret;
3066  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
3067  return __ret;
3068 }
3069 #else
3070 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
3071  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3072  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3073  poly8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3074  poly8x8_t __ret;
3075  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
3076  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3077  return __ret;
3078 }
3079 #endif
3080 
3081 #ifdef __LITTLE_ENDIAN__
3082 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
3083  poly16x4_t __ret;
3084  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
3085  return __ret;
3086 }
3087 #else
3088 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
3089  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3090  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3091  poly16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3092  poly16x4_t __ret;
3093  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
3094  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3095  return __ret;
3096 }
3097 #endif
3098 
3099 #ifdef __LITTLE_ENDIAN__
3100 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
3101  poly8x16_t __ret;
3102  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
3103  return __ret;
3104 }
3105 #else
3106 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
3107  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3108  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3109  poly8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3110  poly8x16_t __ret;
3111  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
3112  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3113  return __ret;
3114 }
3115 #endif
3116 
3117 #ifdef __LITTLE_ENDIAN__
3118 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
3119  poly16x8_t __ret;
3120  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
3121  return __ret;
3122 }
3123 #else
3124 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
3125  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3126  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3127  poly16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3128  poly16x8_t __ret;
3129  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
3130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3131  return __ret;
3132 }
3133 #endif
3134 
3135 #ifdef __LITTLE_ENDIAN__
3136 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
3137  uint8x16_t __ret;
3138  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
3139  return __ret;
3140 }
3141 #else
3142 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
3143  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3144  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3145  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3146  uint8x16_t __ret;
3147  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
3148  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3149  return __ret;
3150 }
3151 #endif
3152 
3153 #ifdef __LITTLE_ENDIAN__
3154 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
3155  uint32x4_t __ret;
3156  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
3157  return __ret;
3158 }
3159 #else
3160 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
3161  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3162  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3163  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3164  uint32x4_t __ret;
3165  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
3166  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3167  return __ret;
3168 }
3169 #endif
3170 
3171 #ifdef __LITTLE_ENDIAN__
3172 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
3173  uint64x2_t __ret;
3174  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
3175  return __ret;
3176 }
3177 #else
3178 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
3179  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3180  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3181  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3182  uint64x2_t __ret;
3183  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
3184  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3185  return __ret;
3186 }
3187 #endif
3188 
3189 #ifdef __LITTLE_ENDIAN__
3190 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
3191  uint16x8_t __ret;
3192  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
3193  return __ret;
3194 }
3195 #else
3196 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
3197  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3198  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3199  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3200  uint16x8_t __ret;
3201  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
3202  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3203  return __ret;
3204 }
3205 #endif
3206 
3207 #ifdef __LITTLE_ENDIAN__
3208 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
3209  int8x16_t __ret;
3210  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
3211  return __ret;
3212 }
3213 #else
3214 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
3215  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3216  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3217  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3218  int8x16_t __ret;
3219  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
3220  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3221  return __ret;
3222 }
3223 #endif
3224 
3225 #ifdef __LITTLE_ENDIAN__
3226 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
3227  float32x4_t __ret;
3228  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
3229  return __ret;
3230 }
3231 #else
3232 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
3233  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3234  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3235  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3236  float32x4_t __ret;
3237  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
3238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3239  return __ret;
3240 }
3241 #endif
3242 
3243 #ifdef __LITTLE_ENDIAN__
3244 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
3245  int32x4_t __ret;
3246  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
3247  return __ret;
3248 }
3249 #else
3250 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
3251  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3252  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3253  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3254  int32x4_t __ret;
3255  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
3256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3257  return __ret;
3258 }
3259 #endif
3260 
3261 #ifdef __LITTLE_ENDIAN__
3262 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
3263  int64x2_t __ret;
3264  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
3265  return __ret;
3266 }
3267 #else
3268 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
3269  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3270  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3271  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3272  int64x2_t __ret;
3273  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
3274  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3275  return __ret;
3276 }
3277 #endif
3278 
3279 #ifdef __LITTLE_ENDIAN__
3280 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
3281  int16x8_t __ret;
3282  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
3283  return __ret;
3284 }
3285 #else
3286 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
3287  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3288  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3289  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3290  int16x8_t __ret;
3291  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
3292  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3293  return __ret;
3294 }
3295 #endif
3296 
3297 #ifdef __LITTLE_ENDIAN__
3298 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
3299  uint8x8_t __ret;
3300  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
3301  return __ret;
3302 }
3303 #else
3304 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
3305  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3306  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3307  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3308  uint8x8_t __ret;
3309  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
3310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3311  return __ret;
3312 }
3313 #endif
3314 
3315 #ifdef __LITTLE_ENDIAN__
3316 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
3317  uint32x2_t __ret;
3318  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
3319  return __ret;
3320 }
3321 #else
3322 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
3323  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3324  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3325  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3326  uint32x2_t __ret;
3327  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
3328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3329  return __ret;
3330 }
3331 #endif
3332 
3333 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
3334  uint64x1_t __ret;
3335  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
3336  return __ret;
3337 }
3338 #ifdef __LITTLE_ENDIAN__
3339 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
3340  uint16x4_t __ret;
3341  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
3342  return __ret;
3343 }
3344 #else
3345 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
3346  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3347  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3348  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3349  uint16x4_t __ret;
3350  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
3351  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3352  return __ret;
3353 }
3354 #endif
3355 
3356 #ifdef __LITTLE_ENDIAN__
3357 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
3358  int8x8_t __ret;
3359  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
3360  return __ret;
3361 }
3362 #else
3363 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
3364  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3365  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3366  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
3367  int8x8_t __ret;
3368  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
3369  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3370  return __ret;
3371 }
3372 #endif
3373 
3374 #ifdef __LITTLE_ENDIAN__
3375 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
3376  float32x2_t __ret;
3377  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
3378  return __ret;
3379 }
3380 #else
3381 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
3382  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3383  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3384  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3385  float32x2_t __ret;
3386  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
3387  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3388  return __ret;
3389 }
3390 #endif
3391 
3392 #ifdef __LITTLE_ENDIAN__
3393 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
3394  int32x2_t __ret;
3395  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
3396  return __ret;
3397 }
3398 #else
3399 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
3400  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3401  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3402  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
3403  int32x2_t __ret;
3404  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
3405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3406  return __ret;
3407 }
3408 #endif
3409 
3410 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
3411  int64x1_t __ret;
3412  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
3413  return __ret;
3414 }
3415 #ifdef __LITTLE_ENDIAN__
3416 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
3417  int16x4_t __ret;
3418  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
3419  return __ret;
3420 }
3421 #else
3422 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
3423  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3424  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3425  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
3426  int16x4_t __ret;
3427  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
3428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3429  return __ret;
3430 }
3431 #endif
3432 
3433 #ifdef __LITTLE_ENDIAN__
3434 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
3435  uint32x4_t __ret;
3436  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3437  return __ret;
3438 }
3439 #else
3440 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
3441  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3442  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3443  uint32x4_t __ret;
3444  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3445  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3446  return __ret;
3447 }
3448 #endif
3449 
3450 #ifdef __LITTLE_ENDIAN__
3451 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
3452  uint32x2_t __ret;
3453  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3454  return __ret;
3455 }
3456 #else
3457 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
3458  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3459  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3460  uint32x2_t __ret;
3461  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3462  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3463  return __ret;
3464 }
3465 #endif
3466 
3467 #ifdef __LITTLE_ENDIAN__
3468 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
3469  uint32x4_t __ret;
3470  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3471  return __ret;
3472 }
3473 #else
3474 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
3475  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3476  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3477  uint32x4_t __ret;
3478  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3479  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3480  return __ret;
3481 }
3482 #endif
3483 
3484 #ifdef __LITTLE_ENDIAN__
3485 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
3486  uint32x2_t __ret;
3487  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3488  return __ret;
3489 }
3490 #else
3491 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
3492  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3493  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3494  uint32x2_t __ret;
3495  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3496  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3497  return __ret;
3498 }
3499 #endif
3500 
3501 #ifdef __LITTLE_ENDIAN__
3502 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
3503  uint32x4_t __ret;
3504  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3505  return __ret;
3506 }
3507 #else
3508 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
3509  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3510  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3511  uint32x4_t __ret;
3512  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3513  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3514  return __ret;
3515 }
3516 #endif
3517 
3518 #ifdef __LITTLE_ENDIAN__
3519 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
3520  uint32x2_t __ret;
3521  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3522  return __ret;
3523 }
3524 #else
3525 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
3526  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3527  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3528  uint32x2_t __ret;
3529  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3530  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3531  return __ret;
3532 }
3533 #endif
3534 
3535 #ifdef __LITTLE_ENDIAN__
3536 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
3537  uint32x4_t __ret;
3538  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
3539  return __ret;
3540 }
3541 #else
3542 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
3543  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3544  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3545  uint32x4_t __ret;
3546  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
3547  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3548  return __ret;
3549 }
3550 #endif
3551 
3552 #ifdef __LITTLE_ENDIAN__
3553 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
3554  uint32x2_t __ret;
3555  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
3556  return __ret;
3557 }
3558 #else
3559 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
3560  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3561  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3562  uint32x2_t __ret;
3563  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
3564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3565  return __ret;
3566 }
3567 #endif
3568 
3569 #ifdef __LITTLE_ENDIAN__
3570 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
3571  uint8x8_t __ret;
3572  __ret = (uint8x8_t)(__p0 == __p1);
3573  return __ret;
3574 }
3575 #else
3576 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
3577  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3578  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3579  uint8x8_t __ret;
3580  __ret = (uint8x8_t)(__rev0 == __rev1);
3581  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3582  return __ret;
3583 }
3584 #endif
3585 
3586 #ifdef __LITTLE_ENDIAN__
3587 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
3588  uint8x16_t __ret;
3589  __ret = (uint8x16_t)(__p0 == __p1);
3590  return __ret;
3591 }
3592 #else
3593 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
3594  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3595  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3596  uint8x16_t __ret;
3597  __ret = (uint8x16_t)(__rev0 == __rev1);
3598  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3599  return __ret;
3600 }
3601 #endif
3602 
3603 #ifdef __LITTLE_ENDIAN__
3604 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3605  uint8x16_t __ret;
3606  __ret = (uint8x16_t)(__p0 == __p1);
3607  return __ret;
3608 }
3609 #else
3610 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3611  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3612  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3613  uint8x16_t __ret;
3614  __ret = (uint8x16_t)(__rev0 == __rev1);
3615  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3616  return __ret;
3617 }
3618 #endif
3619 
3620 #ifdef __LITTLE_ENDIAN__
3621 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3622  uint32x4_t __ret;
3623  __ret = (uint32x4_t)(__p0 == __p1);
3624  return __ret;
3625 }
3626 #else
3627 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3628  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3629  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3630  uint32x4_t __ret;
3631  __ret = (uint32x4_t)(__rev0 == __rev1);
3632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3633  return __ret;
3634 }
3635 #endif
3636 
3637 #ifdef __LITTLE_ENDIAN__
3638 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3639  uint16x8_t __ret;
3640  __ret = (uint16x8_t)(__p0 == __p1);
3641  return __ret;
3642 }
3643 #else
3644 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3645  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3646  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3647  uint16x8_t __ret;
3648  __ret = (uint16x8_t)(__rev0 == __rev1);
3649  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3650  return __ret;
3651 }
3652 #endif
3653 
3654 #ifdef __LITTLE_ENDIAN__
3655 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
3656  uint8x16_t __ret;
3657  __ret = (uint8x16_t)(__p0 == __p1);
3658  return __ret;
3659 }
3660 #else
3661 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
3662  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3663  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3664  uint8x16_t __ret;
3665  __ret = (uint8x16_t)(__rev0 == __rev1);
3666  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3667  return __ret;
3668 }
3669 #endif
3670 
3671 #ifdef __LITTLE_ENDIAN__
3672 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
3673  uint32x4_t __ret;
3674  __ret = (uint32x4_t)(__p0 == __p1);
3675  return __ret;
3676 }
3677 #else
3678 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
3679  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3680  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3681  uint32x4_t __ret;
3682  __ret = (uint32x4_t)(__rev0 == __rev1);
3683  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3684  return __ret;
3685 }
3686 #endif
3687 
3688 #ifdef __LITTLE_ENDIAN__
3689 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
3690  uint32x4_t __ret;
3691  __ret = (uint32x4_t)(__p0 == __p1);
3692  return __ret;
3693 }
3694 #else
3695 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
3696  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3697  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3698  uint32x4_t __ret;
3699  __ret = (uint32x4_t)(__rev0 == __rev1);
3700  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3701  return __ret;
3702 }
3703 #endif
3704 
3705 #ifdef __LITTLE_ENDIAN__
3706 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
3707  uint16x8_t __ret;
3708  __ret = (uint16x8_t)(__p0 == __p1);
3709  return __ret;
3710 }
3711 #else
3712 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
3713  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3714  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3715  uint16x8_t __ret;
3716  __ret = (uint16x8_t)(__rev0 == __rev1);
3717  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3718  return __ret;
3719 }
3720 #endif
3721 
3722 #ifdef __LITTLE_ENDIAN__
3723 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
3724  uint8x8_t __ret;
3725  __ret = (uint8x8_t)(__p0 == __p1);
3726  return __ret;
3727 }
3728 #else
3729 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
3730  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3731  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3732  uint8x8_t __ret;
3733  __ret = (uint8x8_t)(__rev0 == __rev1);
3734  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3735  return __ret;
3736 }
3737 #endif
3738 
3739 #ifdef __LITTLE_ENDIAN__
3740 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
3741  uint32x2_t __ret;
3742  __ret = (uint32x2_t)(__p0 == __p1);
3743  return __ret;
3744 }
3745 #else
3746 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
3747  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3748  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3749  uint32x2_t __ret;
3750  __ret = (uint32x2_t)(__rev0 == __rev1);
3751  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3752  return __ret;
3753 }
3754 #endif
3755 
3756 #ifdef __LITTLE_ENDIAN__
3757 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
3758  uint16x4_t __ret;
3759  __ret = (uint16x4_t)(__p0 == __p1);
3760  return __ret;
3761 }
3762 #else
3763 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
3764  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3765  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3766  uint16x4_t __ret;
3767  __ret = (uint16x4_t)(__rev0 == __rev1);
3768  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3769  return __ret;
3770 }
3771 #endif
3772 
3773 #ifdef __LITTLE_ENDIAN__
3774 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
3775  uint8x8_t __ret;
3776  __ret = (uint8x8_t)(__p0 == __p1);
3777  return __ret;
3778 }
3779 #else
3780 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
3781  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3782  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3783  uint8x8_t __ret;
3784  __ret = (uint8x8_t)(__rev0 == __rev1);
3785  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3786  return __ret;
3787 }
3788 #endif
3789 
3790 #ifdef __LITTLE_ENDIAN__
3791 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
3792  uint32x2_t __ret;
3793  __ret = (uint32x2_t)(__p0 == __p1);
3794  return __ret;
3795 }
3796 #else
3797 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
3798  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3799  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3800  uint32x2_t __ret;
3801  __ret = (uint32x2_t)(__rev0 == __rev1);
3802  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3803  return __ret;
3804 }
3805 #endif
3806 
3807 #ifdef __LITTLE_ENDIAN__
3808 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
3809  uint32x2_t __ret;
3810  __ret = (uint32x2_t)(__p0 == __p1);
3811  return __ret;
3812 }
3813 #else
3814 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
3815  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3816  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3817  uint32x2_t __ret;
3818  __ret = (uint32x2_t)(__rev0 == __rev1);
3819  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3820  return __ret;
3821 }
3822 #endif
3823 
3824 #ifdef __LITTLE_ENDIAN__
3825 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
3826  uint16x4_t __ret;
3827  __ret = (uint16x4_t)(__p0 == __p1);
3828  return __ret;
3829 }
3830 #else
3831 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
3832  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3833  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3834  uint16x4_t __ret;
3835  __ret = (uint16x4_t)(__rev0 == __rev1);
3836  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3837  return __ret;
3838 }
3839 #endif
3840 
3841 #ifdef __LITTLE_ENDIAN__
3842 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3843  uint8x16_t __ret;
3844  __ret = (uint8x16_t)(__p0 >= __p1);
3845  return __ret;
3846 }
3847 #else
3848 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3849  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3850  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3851  uint8x16_t __ret;
3852  __ret = (uint8x16_t)(__rev0 >= __rev1);
3853  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3854  return __ret;
3855 }
3856 #endif
3857 
3858 #ifdef __LITTLE_ENDIAN__
3859 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3860  uint32x4_t __ret;
3861  __ret = (uint32x4_t)(__p0 >= __p1);
3862  return __ret;
3863 }
3864 #else
3865 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3866  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3867  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3868  uint32x4_t __ret;
3869  __ret = (uint32x4_t)(__rev0 >= __rev1);
3870  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3871  return __ret;
3872 }
3873 #endif
3874 
3875 #ifdef __LITTLE_ENDIAN__
3876 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3877  uint16x8_t __ret;
3878  __ret = (uint16x8_t)(__p0 >= __p1);
3879  return __ret;
3880 }
3881 #else
3882 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3883  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3884  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3885  uint16x8_t __ret;
3886  __ret = (uint16x8_t)(__rev0 >= __rev1);
3887  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3888  return __ret;
3889 }
3890 #endif
3891 
3892 #ifdef __LITTLE_ENDIAN__
3893 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
3894  uint8x16_t __ret;
3895  __ret = (uint8x16_t)(__p0 >= __p1);
3896  return __ret;
3897 }
3898 #else
3899 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
3900  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3901  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3902  uint8x16_t __ret;
3903  __ret = (uint8x16_t)(__rev0 >= __rev1);
3904  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3905  return __ret;
3906 }
3907 #endif
3908 
3909 #ifdef __LITTLE_ENDIAN__
3910 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
3911  uint32x4_t __ret;
3912  __ret = (uint32x4_t)(__p0 >= __p1);
3913  return __ret;
3914 }
3915 #else
3916 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
3917  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3918  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3919  uint32x4_t __ret;
3920  __ret = (uint32x4_t)(__rev0 >= __rev1);
3921  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3922  return __ret;
3923 }
3924 #endif
3925 
3926 #ifdef __LITTLE_ENDIAN__
3927 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
3928  uint32x4_t __ret;
3929  __ret = (uint32x4_t)(__p0 >= __p1);
3930  return __ret;
3931 }
3932 #else
3933 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
3934  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3935  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3936  uint32x4_t __ret;
3937  __ret = (uint32x4_t)(__rev0 >= __rev1);
3938  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3939  return __ret;
3940 }
3941 #endif
3942 
3943 #ifdef __LITTLE_ENDIAN__
3944 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
3945  uint16x8_t __ret;
3946  __ret = (uint16x8_t)(__p0 >= __p1);
3947  return __ret;
3948 }
3949 #else
3950 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
3951  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3952  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3953  uint16x8_t __ret;
3954  __ret = (uint16x8_t)(__rev0 >= __rev1);
3955  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3956  return __ret;
3957 }
3958 #endif
3959 
3960 #ifdef __LITTLE_ENDIAN__
3961 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
3962  uint8x8_t __ret;
3963  __ret = (uint8x8_t)(__p0 >= __p1);
3964  return __ret;
3965 }
3966 #else
3967 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
3968  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3969  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3970  uint8x8_t __ret;
3971  __ret = (uint8x8_t)(__rev0 >= __rev1);
3972  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3973  return __ret;
3974 }
3975 #endif
3976 
3977 #ifdef __LITTLE_ENDIAN__
3978 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
3979  uint32x2_t __ret;
3980  __ret = (uint32x2_t)(__p0 >= __p1);
3981  return __ret;
3982 }
3983 #else
3984 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
3985  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3986  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3987  uint32x2_t __ret;
3988  __ret = (uint32x2_t)(__rev0 >= __rev1);
3989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3990  return __ret;
3991 }
3992 #endif
3993 
3994 #ifdef __LITTLE_ENDIAN__
3995 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
3996  uint16x4_t __ret;
3997  __ret = (uint16x4_t)(__p0 >= __p1);
3998  return __ret;
3999 }
4000 #else
4001 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
4002  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4003  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4004  uint16x4_t __ret;
4005  __ret = (uint16x4_t)(__rev0 >= __rev1);
4006  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4007  return __ret;
4008 }
4009 #endif
4010 
4011 #ifdef __LITTLE_ENDIAN__
4012 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
4013  uint8x8_t __ret;
4014  __ret = (uint8x8_t)(__p0 >= __p1);
4015  return __ret;
4016 }
4017 #else
4018 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
4019  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4020  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4021  uint8x8_t __ret;
4022  __ret = (uint8x8_t)(__rev0 >= __rev1);
4023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4024  return __ret;
4025 }
4026 #endif
4027 
4028 #ifdef __LITTLE_ENDIAN__
4029 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
4030  uint32x2_t __ret;
4031  __ret = (uint32x2_t)(__p0 >= __p1);
4032  return __ret;
4033 }
4034 #else
4035 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
4036  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4037  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4038  uint32x2_t __ret;
4039  __ret = (uint32x2_t)(__rev0 >= __rev1);
4040  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4041  return __ret;
4042 }
4043 #endif
4044 
4045 #ifdef __LITTLE_ENDIAN__
4046 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
4047  uint32x2_t __ret;
4048  __ret = (uint32x2_t)(__p0 >= __p1);
4049  return __ret;
4050 }
4051 #else
4052 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
4053  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4054  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4055  uint32x2_t __ret;
4056  __ret = (uint32x2_t)(__rev0 >= __rev1);
4057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4058  return __ret;
4059 }
4060 #endif
4061 
4062 #ifdef __LITTLE_ENDIAN__
4063 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
4064  uint16x4_t __ret;
4065  __ret = (uint16x4_t)(__p0 >= __p1);
4066  return __ret;
4067 }
4068 #else
4069 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
4070  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4071  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4072  uint16x4_t __ret;
4073  __ret = (uint16x4_t)(__rev0 >= __rev1);
4074  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4075  return __ret;
4076 }
4077 #endif
4078 
4079 #ifdef __LITTLE_ENDIAN__
4080 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4081  uint8x16_t __ret;
4082  __ret = (uint8x16_t)(__p0 > __p1);
4083  return __ret;
4084 }
4085 #else
4086 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4087  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4088  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4089  uint8x16_t __ret;
4090  __ret = (uint8x16_t)(__rev0 > __rev1);
4091  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4092  return __ret;
4093 }
4094 #endif
4095 
4096 #ifdef __LITTLE_ENDIAN__
4097 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4098  uint32x4_t __ret;
4099  __ret = (uint32x4_t)(__p0 > __p1);
4100  return __ret;
4101 }
4102 #else
4103 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4104  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4105  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4106  uint32x4_t __ret;
4107  __ret = (uint32x4_t)(__rev0 > __rev1);
4108  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4109  return __ret;
4110 }
4111 #endif
4112 
4113 #ifdef __LITTLE_ENDIAN__
4114 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4115  uint16x8_t __ret;
4116  __ret = (uint16x8_t)(__p0 > __p1);
4117  return __ret;
4118 }
4119 #else
4120 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4121  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4122  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4123  uint16x8_t __ret;
4124  __ret = (uint16x8_t)(__rev0 > __rev1);
4125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4126  return __ret;
4127 }
4128 #endif
4129 
4130 #ifdef __LITTLE_ENDIAN__
4131 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
4132  uint8x16_t __ret;
4133  __ret = (uint8x16_t)(__p0 > __p1);
4134  return __ret;
4135 }
4136 #else
4137 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
4138  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4139  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4140  uint8x16_t __ret;
4141  __ret = (uint8x16_t)(__rev0 > __rev1);
4142  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4143  return __ret;
4144 }
4145 #endif
4146 
4147 #ifdef __LITTLE_ENDIAN__
4148 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
4149  uint32x4_t __ret;
4150  __ret = (uint32x4_t)(__p0 > __p1);
4151  return __ret;
4152 }
4153 #else
4154 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
4155  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4156  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4157  uint32x4_t __ret;
4158  __ret = (uint32x4_t)(__rev0 > __rev1);
4159  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4160  return __ret;
4161 }
4162 #endif
4163 
4164 #ifdef __LITTLE_ENDIAN__
4165 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
4166  uint32x4_t __ret;
4167  __ret = (uint32x4_t)(__p0 > __p1);
4168  return __ret;
4169 }
4170 #else
4171 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
4172  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4173  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4174  uint32x4_t __ret;
4175  __ret = (uint32x4_t)(__rev0 > __rev1);
4176  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4177  return __ret;
4178 }
4179 #endif
4180 
4181 #ifdef __LITTLE_ENDIAN__
4182 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
4183  uint16x8_t __ret;
4184  __ret = (uint16x8_t)(__p0 > __p1);
4185  return __ret;
4186 }
4187 #else
4188 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
4189  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4190  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4191  uint16x8_t __ret;
4192  __ret = (uint16x8_t)(__rev0 > __rev1);
4193  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4194  return __ret;
4195 }
4196 #endif
4197 
4198 #ifdef __LITTLE_ENDIAN__
4199 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4200  uint8x8_t __ret;
4201  __ret = (uint8x8_t)(__p0 > __p1);
4202  return __ret;
4203 }
4204 #else
4205 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4206  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4207  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4208  uint8x8_t __ret;
4209  __ret = (uint8x8_t)(__rev0 > __rev1);
4210  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4211  return __ret;
4212 }
4213 #endif
4214 
4215 #ifdef __LITTLE_ENDIAN__
4216 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4217  uint32x2_t __ret;
4218  __ret = (uint32x2_t)(__p0 > __p1);
4219  return __ret;
4220 }
4221 #else
4222 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4223  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4224  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4225  uint32x2_t __ret;
4226  __ret = (uint32x2_t)(__rev0 > __rev1);
4227  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4228  return __ret;
4229 }
4230 #endif
4231 
4232 #ifdef __LITTLE_ENDIAN__
4233 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4234  uint16x4_t __ret;
4235  __ret = (uint16x4_t)(__p0 > __p1);
4236  return __ret;
4237 }
4238 #else
4239 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4240  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4241  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4242  uint16x4_t __ret;
4243  __ret = (uint16x4_t)(__rev0 > __rev1);
4244  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4245  return __ret;
4246 }
4247 #endif
4248 
4249 #ifdef __LITTLE_ENDIAN__
4250 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
4251  uint8x8_t __ret;
4252  __ret = (uint8x8_t)(__p0 > __p1);
4253  return __ret;
4254 }
4255 #else
4256 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
4257  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4258  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4259  uint8x8_t __ret;
4260  __ret = (uint8x8_t)(__rev0 > __rev1);
4261  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4262  return __ret;
4263 }
4264 #endif
4265 
4266 #ifdef __LITTLE_ENDIAN__
4267 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
4268  uint32x2_t __ret;
4269  __ret = (uint32x2_t)(__p0 > __p1);
4270  return __ret;
4271 }
4272 #else
4273 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
4274  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4275  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4276  uint32x2_t __ret;
4277  __ret = (uint32x2_t)(__rev0 > __rev1);
4278  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4279  return __ret;
4280 }
4281 #endif
4282 
4283 #ifdef __LITTLE_ENDIAN__
4284 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
4285  uint32x2_t __ret;
4286  __ret = (uint32x2_t)(__p0 > __p1);
4287  return __ret;
4288 }
4289 #else
4290 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
4291  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4292  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4293  uint32x2_t __ret;
4294  __ret = (uint32x2_t)(__rev0 > __rev1);
4295  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4296  return __ret;
4297 }
4298 #endif
4299 
4300 #ifdef __LITTLE_ENDIAN__
4301 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
4302  uint16x4_t __ret;
4303  __ret = (uint16x4_t)(__p0 > __p1);
4304  return __ret;
4305 }
4306 #else
4307 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
4308  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4309  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4310  uint16x4_t __ret;
4311  __ret = (uint16x4_t)(__rev0 > __rev1);
4312  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4313  return __ret;
4314 }
4315 #endif
4316 
4317 #ifdef __LITTLE_ENDIAN__
4318 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4319  uint8x16_t __ret;
4320  __ret = (uint8x16_t)(__p0 <= __p1);
4321  return __ret;
4322 }
4323 #else
4324 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4325  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4326  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4327  uint8x16_t __ret;
4328  __ret = (uint8x16_t)(__rev0 <= __rev1);
4329  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4330  return __ret;
4331 }
4332 #endif
4333 
4334 #ifdef __LITTLE_ENDIAN__
4335 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4336  uint32x4_t __ret;
4337  __ret = (uint32x4_t)(__p0 <= __p1);
4338  return __ret;
4339 }
4340 #else
4341 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4342  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4343  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4344  uint32x4_t __ret;
4345  __ret = (uint32x4_t)(__rev0 <= __rev1);
4346  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4347  return __ret;
4348 }
4349 #endif
4350 
4351 #ifdef __LITTLE_ENDIAN__
4352 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4353  uint16x8_t __ret;
4354  __ret = (uint16x8_t)(__p0 <= __p1);
4355  return __ret;
4356 }
4357 #else
4358 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4359  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4360  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4361  uint16x8_t __ret;
4362  __ret = (uint16x8_t)(__rev0 <= __rev1);
4363  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4364  return __ret;
4365 }
4366 #endif
4367 
4368 #ifdef __LITTLE_ENDIAN__
4369 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
4370  uint8x16_t __ret;
4371  __ret = (uint8x16_t)(__p0 <= __p1);
4372  return __ret;
4373 }
4374 #else
4375 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
4376  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4377  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4378  uint8x16_t __ret;
4379  __ret = (uint8x16_t)(__rev0 <= __rev1);
4380  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4381  return __ret;
4382 }
4383 #endif
4384 
4385 #ifdef __LITTLE_ENDIAN__
4386 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
4387  uint32x4_t __ret;
4388  __ret = (uint32x4_t)(__p0 <= __p1);
4389  return __ret;
4390 }
4391 #else
4392 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
4393  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4394  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4395  uint32x4_t __ret;
4396  __ret = (uint32x4_t)(__rev0 <= __rev1);
4397  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4398  return __ret;
4399 }
4400 #endif
4401 
4402 #ifdef __LITTLE_ENDIAN__
4403 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
4404  uint32x4_t __ret;
4405  __ret = (uint32x4_t)(__p0 <= __p1);
4406  return __ret;
4407 }
4408 #else
4409 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
4410  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4411  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4412  uint32x4_t __ret;
4413  __ret = (uint32x4_t)(__rev0 <= __rev1);
4414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4415  return __ret;
4416 }
4417 #endif
4418 
4419 #ifdef __LITTLE_ENDIAN__
4420 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
4421  uint16x8_t __ret;
4422  __ret = (uint16x8_t)(__p0 <= __p1);
4423  return __ret;
4424 }
4425 #else
4426 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
4427  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4428  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4429  uint16x8_t __ret;
4430  __ret = (uint16x8_t)(__rev0 <= __rev1);
4431  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4432  return __ret;
4433 }
4434 #endif
4435 
4436 #ifdef __LITTLE_ENDIAN__
4437 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
4438  uint8x8_t __ret;
4439  __ret = (uint8x8_t)(__p0 <= __p1);
4440  return __ret;
4441 }
4442 #else
4443 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
4444  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4445  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4446  uint8x8_t __ret;
4447  __ret = (uint8x8_t)(__rev0 <= __rev1);
4448  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4449  return __ret;
4450 }
4451 #endif
4452 
4453 #ifdef __LITTLE_ENDIAN__
4454 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
4455  uint32x2_t __ret;
4456  __ret = (uint32x2_t)(__p0 <= __p1);
4457  return __ret;
4458 }
4459 #else
4460 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
4461  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4462  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4463  uint32x2_t __ret;
4464  __ret = (uint32x2_t)(__rev0 <= __rev1);
4465  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4466  return __ret;
4467 }
4468 #endif
4469 
4470 #ifdef __LITTLE_ENDIAN__
4471 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
4472  uint16x4_t __ret;
4473  __ret = (uint16x4_t)(__p0 <= __p1);
4474  return __ret;
4475 }
4476 #else
4477 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
4478  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4479  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4480  uint16x4_t __ret;
4481  __ret = (uint16x4_t)(__rev0 <= __rev1);
4482  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4483  return __ret;
4484 }
4485 #endif
4486 
4487 #ifdef __LITTLE_ENDIAN__
4488 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
4489  uint8x8_t __ret;
4490  __ret = (uint8x8_t)(__p0 <= __p1);
4491  return __ret;
4492 }
4493 #else
4494 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
4495  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4496  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4497  uint8x8_t __ret;
4498  __ret = (uint8x8_t)(__rev0 <= __rev1);
4499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4500  return __ret;
4501 }
4502 #endif
4503 
4504 #ifdef __LITTLE_ENDIAN__
4505 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
4506  uint32x2_t __ret;
4507  __ret = (uint32x2_t)(__p0 <= __p1);
4508  return __ret;
4509 }
4510 #else
4511 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
4512  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4513  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4514  uint32x2_t __ret;
4515  __ret = (uint32x2_t)(__rev0 <= __rev1);
4516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4517  return __ret;
4518 }
4519 #endif
4520 
4521 #ifdef __LITTLE_ENDIAN__
4522 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
4523  uint32x2_t __ret;
4524  __ret = (uint32x2_t)(__p0 <= __p1);
4525  return __ret;
4526 }
4527 #else
4528 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
4529  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4530  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4531  uint32x2_t __ret;
4532  __ret = (uint32x2_t)(__rev0 <= __rev1);
4533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4534  return __ret;
4535 }
4536 #endif
4537 
4538 #ifdef __LITTLE_ENDIAN__
4539 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
4540  uint16x4_t __ret;
4541  __ret = (uint16x4_t)(__p0 <= __p1);
4542  return __ret;
4543 }
4544 #else
4545 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
4546  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4547  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4548  uint16x4_t __ret;
4549  __ret = (uint16x4_t)(__rev0 <= __rev1);
4550  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4551  return __ret;
4552 }
4553 #endif
4554 
4555 #ifdef __LITTLE_ENDIAN__
4556 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
4557  int8x16_t __ret;
4558  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
4559  return __ret;
4560 }
4561 #else
4562 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
4563  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4564  int8x16_t __ret;
4565  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
4566  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4567  return __ret;
4568 }
4569 #endif
4570 
4571 #ifdef __LITTLE_ENDIAN__
4572 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
4573  int32x4_t __ret;
4574  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
4575  return __ret;
4576 }
4577 #else
4578 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
4579  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4580  int32x4_t __ret;
4581  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
4582  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4583  return __ret;
4584 }
4585 #endif
4586 
4587 #ifdef __LITTLE_ENDIAN__
4588 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
4589  int16x8_t __ret;
4590  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
4591  return __ret;
4592 }
4593 #else
4594 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
4595  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4596  int16x8_t __ret;
4597  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
4598  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4599  return __ret;
4600 }
4601 #endif
4602 
4603 #ifdef __LITTLE_ENDIAN__
4604 __ai int8x8_t vcls_s8(int8x8_t __p0) {
4605  int8x8_t __ret;
4606  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
4607  return __ret;
4608 }
4609 #else
4610 __ai int8x8_t vcls_s8(int8x8_t __p0) {
4611  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4612  int8x8_t __ret;
4613  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
4614  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4615  return __ret;
4616 }
4617 #endif
4618 
4619 #ifdef __LITTLE_ENDIAN__
4620 __ai int32x2_t vcls_s32(int32x2_t __p0) {
4621  int32x2_t __ret;
4622  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
4623  return __ret;
4624 }
4625 #else
4626 __ai int32x2_t vcls_s32(int32x2_t __p0) {
4627  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4628  int32x2_t __ret;
4629  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
4630  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4631  return __ret;
4632 }
4633 #endif
4634 
4635 #ifdef __LITTLE_ENDIAN__
4636 __ai int16x4_t vcls_s16(int16x4_t __p0) {
4637  int16x4_t __ret;
4638  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
4639  return __ret;
4640 }
4641 #else
4642 __ai int16x4_t vcls_s16(int16x4_t __p0) {
4643  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4644  int16x4_t __ret;
4645  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
4646  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4647  return __ret;
4648 }
4649 #endif
4650 
4651 #ifdef __LITTLE_ENDIAN__
4652 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4653  uint8x16_t __ret;
4654  __ret = (uint8x16_t)(__p0 < __p1);
4655  return __ret;
4656 }
4657 #else
4658 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
4659  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4660  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4661  uint8x16_t __ret;
4662  __ret = (uint8x16_t)(__rev0 < __rev1);
4663  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4664  return __ret;
4665 }
4666 #endif
4667 
4668 #ifdef __LITTLE_ENDIAN__
4669 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4670  uint32x4_t __ret;
4671  __ret = (uint32x4_t)(__p0 < __p1);
4672  return __ret;
4673 }
4674 #else
4675 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
4676  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4677  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4678  uint32x4_t __ret;
4679  __ret = (uint32x4_t)(__rev0 < __rev1);
4680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4681  return __ret;
4682 }
4683 #endif
4684 
4685 #ifdef __LITTLE_ENDIAN__
4686 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4687  uint16x8_t __ret;
4688  __ret = (uint16x8_t)(__p0 < __p1);
4689  return __ret;
4690 }
4691 #else
4692 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
4693  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4694  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4695  uint16x8_t __ret;
4696  __ret = (uint16x8_t)(__rev0 < __rev1);
4697  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4698  return __ret;
4699 }
4700 #endif
4701 
4702 #ifdef __LITTLE_ENDIAN__
4703 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
4704  uint8x16_t __ret;
4705  __ret = (uint8x16_t)(__p0 < __p1);
4706  return __ret;
4707 }
4708 #else
4709 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
4710  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4711  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4712  uint8x16_t __ret;
4713  __ret = (uint8x16_t)(__rev0 < __rev1);
4714  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4715  return __ret;
4716 }
4717 #endif
4718 
4719 #ifdef __LITTLE_ENDIAN__
4720 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
4721  uint32x4_t __ret;
4722  __ret = (uint32x4_t)(__p0 < __p1);
4723  return __ret;
4724 }
4725 #else
4726 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
4727  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4728  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4729  uint32x4_t __ret;
4730  __ret = (uint32x4_t)(__rev0 < __rev1);
4731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4732  return __ret;
4733 }
4734 #endif
4735 
4736 #ifdef __LITTLE_ENDIAN__
4737 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
4738  uint32x4_t __ret;
4739  __ret = (uint32x4_t)(__p0 < __p1);
4740  return __ret;
4741 }
4742 #else
4743 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
4744  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4745  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4746  uint32x4_t __ret;
4747  __ret = (uint32x4_t)(__rev0 < __rev1);
4748  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4749  return __ret;
4750 }
4751 #endif
4752 
4753 #ifdef __LITTLE_ENDIAN__
4754 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
4755  uint16x8_t __ret;
4756  __ret = (uint16x8_t)(__p0 < __p1);
4757  return __ret;
4758 }
4759 #else
4760 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
4761  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4762  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4763  uint16x8_t __ret;
4764  __ret = (uint16x8_t)(__rev0 < __rev1);
4765  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4766  return __ret;
4767 }
4768 #endif
4769 
4770 #ifdef __LITTLE_ENDIAN__
4771 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4772  uint8x8_t __ret;
4773  __ret = (uint8x8_t)(__p0 < __p1);
4774  return __ret;
4775 }
4776 #else
4777 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
4778  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4779  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4780  uint8x8_t __ret;
4781  __ret = (uint8x8_t)(__rev0 < __rev1);
4782  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4783  return __ret;
4784 }
4785 #endif
4786 
4787 #ifdef __LITTLE_ENDIAN__
4788 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4789  uint32x2_t __ret;
4790  __ret = (uint32x2_t)(__p0 < __p1);
4791  return __ret;
4792 }
4793 #else
4794 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
4795  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4796  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4797  uint32x2_t __ret;
4798  __ret = (uint32x2_t)(__rev0 < __rev1);
4799  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4800  return __ret;
4801 }
4802 #endif
4803 
4804 #ifdef __LITTLE_ENDIAN__
4805 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4806  uint16x4_t __ret;
4807  __ret = (uint16x4_t)(__p0 < __p1);
4808  return __ret;
4809 }
4810 #else
4811 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
4812  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4813  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4814  uint16x4_t __ret;
4815  __ret = (uint16x4_t)(__rev0 < __rev1);
4816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4817  return __ret;
4818 }
4819 #endif
4820 
4821 #ifdef __LITTLE_ENDIAN__
4822 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
4823  uint8x8_t __ret;
4824  __ret = (uint8x8_t)(__p0 < __p1);
4825  return __ret;
4826 }
4827 #else
4828 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
4829  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4830  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4831  uint8x8_t __ret;
4832  __ret = (uint8x8_t)(__rev0 < __rev1);
4833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4834  return __ret;
4835 }
4836 #endif
4837 
4838 #ifdef __LITTLE_ENDIAN__
4839 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
4840  uint32x2_t __ret;
4841  __ret = (uint32x2_t)(__p0 < __p1);
4842  return __ret;
4843 }
4844 #else
4845 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
4846  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4847  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4848  uint32x2_t __ret;
4849  __ret = (uint32x2_t)(__rev0 < __rev1);
4850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4851  return __ret;
4852 }
4853 #endif
4854 
4855 #ifdef __LITTLE_ENDIAN__
4856 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
4857  uint32x2_t __ret;
4858  __ret = (uint32x2_t)(__p0 < __p1);
4859  return __ret;
4860 }
4861 #else
4862 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
4863  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4864  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4865  uint32x2_t __ret;
4866  __ret = (uint32x2_t)(__rev0 < __rev1);
4867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4868  return __ret;
4869 }
4870 #endif
4871 
4872 #ifdef __LITTLE_ENDIAN__
4873 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
4874  uint16x4_t __ret;
4875  __ret = (uint16x4_t)(__p0 < __p1);
4876  return __ret;
4877 }
4878 #else
4879 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
4880  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4881  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4882  uint16x4_t __ret;
4883  __ret = (uint16x4_t)(__rev0 < __rev1);
4884  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4885  return __ret;
4886 }
4887 #endif
4888 
4889 #ifdef __LITTLE_ENDIAN__
4890 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
4891  uint8x16_t __ret;
4892  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
4893  return __ret;
4894 }
4895 #else
4896 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
4897  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4898  uint8x16_t __ret;
4899  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
4900  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4901  return __ret;
4902 }
4903 #endif
4904 
4905 #ifdef __LITTLE_ENDIAN__
4906 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
4907  uint32x4_t __ret;
4908  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
4909  return __ret;
4910 }
4911 #else
4912 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
4913  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4914  uint32x4_t __ret;
4915  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
4916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4917  return __ret;
4918 }
4919 #endif
4920 
4921 #ifdef __LITTLE_ENDIAN__
4922 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
4923  uint16x8_t __ret;
4924  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
4925  return __ret;
4926 }
4927 #else
4928 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
4929  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4930  uint16x8_t __ret;
4931  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
4932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4933  return __ret;
4934 }
4935 #endif
4936 
4937 #ifdef __LITTLE_ENDIAN__
4938 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
4939  int8x16_t __ret;
4940  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
4941  return __ret;
4942 }
4943 #else
4944 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
4945  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4946  int8x16_t __ret;
4947  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
4948  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4949  return __ret;
4950 }
4951 #endif
4952 
4953 #ifdef __LITTLE_ENDIAN__
4954 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
4955  int32x4_t __ret;
4956  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
4957  return __ret;
4958 }
4959 #else
4960 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
4961  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4962  int32x4_t __ret;
4963  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
4964  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4965  return __ret;
4966 }
4967 #endif
4968 
4969 #ifdef __LITTLE_ENDIAN__
4970 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
4971  int16x8_t __ret;
4972  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
4973  return __ret;
4974 }
4975 #else
4976 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
4977  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4978  int16x8_t __ret;
4979  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
4980  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4981  return __ret;
4982 }
4983 #endif
4984 
4985 #ifdef __LITTLE_ENDIAN__
4986 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
4987  uint8x8_t __ret;
4988  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
4989  return __ret;
4990 }
4991 #else
4992 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
4993  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4994  uint8x8_t __ret;
4995  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
4996  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4997  return __ret;
4998 }
4999 #endif
5000 
5001 #ifdef __LITTLE_ENDIAN__
5002 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
5003  uint32x2_t __ret;
5004  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
5005  return __ret;
5006 }
5007 #else
5008 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
5009  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5010  uint32x2_t __ret;
5011  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
5012  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5013  return __ret;
5014 }
5015 #endif
5016 
5017 #ifdef __LITTLE_ENDIAN__
5018 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
5019  uint16x4_t __ret;
5020  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
5021  return __ret;
5022 }
5023 #else
5024 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
5025  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5026  uint16x4_t __ret;
5027  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
5028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5029  return __ret;
5030 }
5031 #endif
5032 
5033 #ifdef __LITTLE_ENDIAN__
5034 __ai int8x8_t vclz_s8(int8x8_t __p0) {
5035  int8x8_t __ret;
5036  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
5037  return __ret;
5038 }
5039 #else
5040 __ai int8x8_t vclz_s8(int8x8_t __p0) {
5041  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5042  int8x8_t __ret;
5043  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
5044  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5045  return __ret;
5046 }
5047 #endif
5048 
5049 #ifdef __LITTLE_ENDIAN__
5050 __ai int32x2_t vclz_s32(int32x2_t __p0) {
5051  int32x2_t __ret;
5052  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
5053  return __ret;
5054 }
5055 #else
5056 __ai int32x2_t vclz_s32(int32x2_t __p0) {
5057  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5058  int32x2_t __ret;
5059  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
5060  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5061  return __ret;
5062 }
5063 #endif
5064 
5065 #ifdef __LITTLE_ENDIAN__
5066 __ai int16x4_t vclz_s16(int16x4_t __p0) {
5067  int16x4_t __ret;
5068  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
5069  return __ret;
5070 }
5071 #else
5072 __ai int16x4_t vclz_s16(int16x4_t __p0) {
5073  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5074  int16x4_t __ret;
5075  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
5076  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5077  return __ret;
5078 }
5079 #endif
5080 
5081 #ifdef __LITTLE_ENDIAN__
5082 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
5083  poly8x8_t __ret;
5084  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
5085  return __ret;
5086 }
5087 #else
5088 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
5089  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5090  poly8x8_t __ret;
5091  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
5092  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5093  return __ret;
5094 }
5095 #endif
5096 
5097 #ifdef __LITTLE_ENDIAN__
5098 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
5099  poly8x16_t __ret;
5100  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
5101  return __ret;
5102 }
5103 #else
5104 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
5105  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5106  poly8x16_t __ret;
5107  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
5108  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5109  return __ret;
5110 }
5111 #endif
5112 
5113 #ifdef __LITTLE_ENDIAN__
5114 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
5115  uint8x16_t __ret;
5116  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
5117  return __ret;
5118 }
5119 #else
5120 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
5121  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5122  uint8x16_t __ret;
5123  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
5124  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5125  return __ret;
5126 }
5127 #endif
5128 
5129 #ifdef __LITTLE_ENDIAN__
5130 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
5131  int8x16_t __ret;
5132  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
5133  return __ret;
5134 }
5135 #else
5136 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
5137  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5138  int8x16_t __ret;
5139  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
5140  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5141  return __ret;
5142 }
5143 #endif
5144 
5145 #ifdef __LITTLE_ENDIAN__
5146 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
5147  uint8x8_t __ret;
5148  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
5149  return __ret;
5150 }
5151 #else
5152 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
5153  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5154  uint8x8_t __ret;
5155  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
5156  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5157  return __ret;
5158 }
5159 #endif
5160 
5161 #ifdef __LITTLE_ENDIAN__
5162 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
5163  int8x8_t __ret;
5164  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
5165  return __ret;
5166 }
5167 #else
5168 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
5169  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5170  int8x8_t __ret;
5171  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
5172  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5173  return __ret;
5174 }
5175 #endif
5176 
5177 #ifdef __LITTLE_ENDIAN__
5178 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
5179  poly8x16_t __ret;
5180  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5181  return __ret;
5182 }
5183 #else
5184 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
5185  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5186  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5187  poly8x16_t __ret;
5188  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5189  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5190  return __ret;
5191 }
5192 #endif
5193 
5194 #ifdef __LITTLE_ENDIAN__
5195 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
5196  poly16x8_t __ret;
5197  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5198  return __ret;
5199 }
5200 #else
5201 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
5202  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5203  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5204  poly16x8_t __ret;
5205  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5207  return __ret;
5208 }
5209 #endif
5210 
5211 #ifdef __LITTLE_ENDIAN__
5212 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5213  uint8x16_t __ret;
5214  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5215  return __ret;
5216 }
5217 #else
5218 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5219  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5220  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5221  uint8x16_t __ret;
5222  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5223  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5224  return __ret;
5225 }
5226 __ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
5227  uint8x16_t __ret;
5228  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5229  return __ret;
5230 }
5231 #endif
5232 
5233 #ifdef __LITTLE_ENDIAN__
5234 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5235  uint32x4_t __ret;
5236  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5237  return __ret;
5238 }
5239 #else
5240 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5241  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5242  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5243  uint32x4_t __ret;
5244  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5245  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5246  return __ret;
5247 }
5248 __ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
5249  uint32x4_t __ret;
5250  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5251  return __ret;
5252 }
5253 #endif
5254 
5255 #ifdef __LITTLE_ENDIAN__
5256 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
5257  uint64x2_t __ret;
5258  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5259  return __ret;
5260 }
5261 #else
5262 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
5263  uint64x2_t __ret;
5264  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5266  return __ret;
5267 }
5268 #endif
5269 
5270 #ifdef __LITTLE_ENDIAN__
5271 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5272  uint16x8_t __ret;
5273  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5274  return __ret;
5275 }
5276 #else
5277 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5278  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5279  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5280  uint16x8_t __ret;
5281  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5282  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5283  return __ret;
5284 }
5285 __ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
5286  uint16x8_t __ret;
5287  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5288  return __ret;
5289 }
5290 #endif
5291 
5292 #ifdef __LITTLE_ENDIAN__
5293 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5294  int8x16_t __ret;
5295  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5296  return __ret;
5297 }
5298 #else
5299 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5300  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5301  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5302  int8x16_t __ret;
5303  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5304  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5305  return __ret;
5306 }
5307 __ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
5308  int8x16_t __ret;
5309  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
5310  return __ret;
5311 }
5312 #endif
5313 
5314 #ifdef __LITTLE_ENDIAN__
5315 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5316  float32x4_t __ret;
5317  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5318  return __ret;
5319 }
5320 #else
5321 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5322  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5323  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5324  float32x4_t __ret;
5325  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5326  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5327  return __ret;
5328 }
5329 __ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
5330  float32x4_t __ret;
5331  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5332  return __ret;
5333 }
5334 #endif
5335 
5336 #ifdef __LITTLE_ENDIAN__
5337 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5338  float16x8_t __ret;
5339  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5340  return __ret;
5341 }
5342 #else
5343 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5344  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5345  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5346  float16x8_t __ret;
5347  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5348  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5349  return __ret;
5350 }
5351 __ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
5352  float16x8_t __ret;
5353  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5354  return __ret;
5355 }
5356 #endif
5357 
5358 #ifdef __LITTLE_ENDIAN__
5359 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5360  int32x4_t __ret;
5361  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5362  return __ret;
5363 }
5364 #else
5365 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5366  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5367  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5368  int32x4_t __ret;
5369  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
5370  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5371  return __ret;
5372 }
5373 __ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
5374  int32x4_t __ret;
5375  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
5376  return __ret;
5377 }
5378 #endif
5379 
5380 #ifdef __LITTLE_ENDIAN__
5381 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
5382  int64x2_t __ret;
5383  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5384  return __ret;
5385 }
5386 #else
5387 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
5388  int64x2_t __ret;
5389  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
5390  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5391  return __ret;
5392 }
5393 #endif
5394 
5395 #ifdef __LITTLE_ENDIAN__
5396 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5397  int16x8_t __ret;
5398  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5399  return __ret;
5400 }
5401 #else
5402 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5403  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5404  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5405  int16x8_t __ret;
5406  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
5407  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5408  return __ret;
5409 }
5410 __ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
5411  int16x8_t __ret;
5412  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
5413  return __ret;
5414 }
5415 #endif
5416 
5417 #define vcreate_p8(__p0) __extension__ ({ \
5418  poly8x8_t __ret; \
5419  uint64_t __promote = __p0; \
5420  __ret = (poly8x8_t)(__promote); \
5421  __ret; \
5422 })
5423 #define vcreate_p16(__p0) __extension__ ({ \
5424  poly16x4_t __ret; \
5425  uint64_t __promote = __p0; \
5426  __ret = (poly16x4_t)(__promote); \
5427  __ret; \
5428 })
5429 #define vcreate_u8(__p0) __extension__ ({ \
5430  uint8x8_t __ret; \
5431  uint64_t __promote = __p0; \
5432  __ret = (uint8x8_t)(__promote); \
5433  __ret; \
5434 })
5435 #define vcreate_u32(__p0) __extension__ ({ \
5436  uint32x2_t __ret; \
5437  uint64_t __promote = __p0; \
5438  __ret = (uint32x2_t)(__promote); \
5439  __ret; \
5440 })
5441 #define vcreate_u64(__p0) __extension__ ({ \
5442  uint64x1_t __ret; \
5443  uint64_t __promote = __p0; \
5444  __ret = (uint64x1_t)(__promote); \
5445  __ret; \
5446 })
5447 #define vcreate_u16(__p0) __extension__ ({ \
5448  uint16x4_t __ret; \
5449  uint64_t __promote = __p0; \
5450  __ret = (uint16x4_t)(__promote); \
5451  __ret; \
5452 })
5453 #define vcreate_s8(__p0) __extension__ ({ \
5454  int8x8_t __ret; \
5455  uint64_t __promote = __p0; \
5456  __ret = (int8x8_t)(__promote); \
5457  __ret; \
5458 })
5459 #define vcreate_f32(__p0) __extension__ ({ \
5460  float32x2_t __ret; \
5461  uint64_t __promote = __p0; \
5462  __ret = (float32x2_t)(__promote); \
5463  __ret; \
5464 })
5465 #define vcreate_f16(__p0) __extension__ ({ \
5466  float16x4_t __ret; \
5467  uint64_t __promote = __p0; \
5468  __ret = (float16x4_t)(__promote); \
5469  __ret; \
5470 })
5471 #define vcreate_s32(__p0) __extension__ ({ \
5472  int32x2_t __ret; \
5473  uint64_t __promote = __p0; \
5474  __ret = (int32x2_t)(__promote); \
5475  __ret; \
5476 })
5477 #define vcreate_s64(__p0) __extension__ ({ \
5478  int64x1_t __ret; \
5479  uint64_t __promote = __p0; \
5480  __ret = (int64x1_t)(__promote); \
5481  __ret; \
5482 })
5483 #define vcreate_s16(__p0) __extension__ ({ \
5484  int16x4_t __ret; \
5485  uint64_t __promote = __p0; \
5486  __ret = (int16x4_t)(__promote); \
5487  __ret; \
5488 })
5489 #ifdef __LITTLE_ENDIAN__
5490 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
5491  float32x4_t __ret;
5492  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
5493  return __ret;
5494 }
5495 #else
5496 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
5497  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5498  float32x4_t __ret;
5499  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
5500  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5501  return __ret;
5502 }
5503 #endif
5504 
5505 #ifdef __LITTLE_ENDIAN__
5506 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
5507  float32x4_t __ret;
5508  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
5509  return __ret;
5510 }
5511 #else
5512 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
5513  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5514  float32x4_t __ret;
5515  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
5516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5517  return __ret;
5518 }
5519 #endif
5520 
5521 #ifdef __LITTLE_ENDIAN__
5522 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
5523  float32x2_t __ret;
5524  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
5525  return __ret;
5526 }
5527 #else
5528 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
5529  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5530  float32x2_t __ret;
5531  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
5532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5533  return __ret;
5534 }
5535 #endif
5536 
5537 #ifdef __LITTLE_ENDIAN__
5538 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
5539  float32x2_t __ret;
5540  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
5541  return __ret;
5542 }
5543 #else
5544 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
5545  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5546  float32x2_t __ret;
5547  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
5548  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5549  return __ret;
5550 }
5551 #endif
5552 
5553 #ifdef __LITTLE_ENDIAN__
5554 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
5555  uint32x4_t __s0 = __p0; \
5556  float32x4_t __ret; \
5557  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
5558  __ret; \
5559 })
5560 #else
5561 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
5562  uint32x4_t __s0 = __p0; \
5563  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5564  float32x4_t __ret; \
5565  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
5566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5567  __ret; \
5568 })
5569 #endif
5570 
5571 #ifdef __LITTLE_ENDIAN__
5572 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
5573  int32x4_t __s0 = __p0; \
5574  float32x4_t __ret; \
5575  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
5576  __ret; \
5577 })
5578 #else
5579 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
5580  int32x4_t __s0 = __p0; \
5581  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5582  float32x4_t __ret; \
5583  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
5584  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5585  __ret; \
5586 })
5587 #endif
5588 
5589 #ifdef __LITTLE_ENDIAN__
5590 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
5591  uint32x2_t __s0 = __p0; \
5592  float32x2_t __ret; \
5593  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
5594  __ret; \
5595 })
5596 #else
5597 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
5598  uint32x2_t __s0 = __p0; \
5599  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5600  float32x2_t __ret; \
5601  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
5602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5603  __ret; \
5604 })
5605 #endif
5606 
5607 #ifdef __LITTLE_ENDIAN__
5608 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
5609  int32x2_t __s0 = __p0; \
5610  float32x2_t __ret; \
5611  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
5612  __ret; \
5613 })
5614 #else
5615 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
5616  int32x2_t __s0 = __p0; \
5617  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5618  float32x2_t __ret; \
5619  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
5620  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5621  __ret; \
5622 })
5623 #endif
5624 
5625 #ifdef __LITTLE_ENDIAN__
5626 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
5627  float32x4_t __s0 = __p0; \
5628  int32x4_t __ret; \
5629  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
5630  __ret; \
5631 })
5632 #else
5633 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
5634  float32x4_t __s0 = __p0; \
5635  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5636  int32x4_t __ret; \
5637  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
5638  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5639  __ret; \
5640 })
5641 #endif
5642 
5643 #ifdef __LITTLE_ENDIAN__
5644 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
5645  float32x2_t __s0 = __p0; \
5646  int32x2_t __ret; \
5647  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
5648  __ret; \
5649 })
5650 #else
5651 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
5652  float32x2_t __s0 = __p0; \
5653  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5654  int32x2_t __ret; \
5655  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
5656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5657  __ret; \
5658 })
5659 #endif
5660 
5661 #ifdef __LITTLE_ENDIAN__
5662 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
5663  float32x4_t __s0 = __p0; \
5664  uint32x4_t __ret; \
5665  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
5666  __ret; \
5667 })
5668 #else
5669 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
5670  float32x4_t __s0 = __p0; \
5671  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5672  uint32x4_t __ret; \
5673  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
5674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5675  __ret; \
5676 })
5677 #endif
5678 
5679 #ifdef __LITTLE_ENDIAN__
5680 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
5681  float32x2_t __s0 = __p0; \
5682  uint32x2_t __ret; \
5683  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
5684  __ret; \
5685 })
5686 #else
5687 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
5688  float32x2_t __s0 = __p0; \
5689  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5690  uint32x2_t __ret; \
5691  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
5692  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5693  __ret; \
5694 })
5695 #endif
5696 
5697 #ifdef __LITTLE_ENDIAN__
5698 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
5699  int32x4_t __ret;
5700  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
5701  return __ret;
5702 }
5703 #else
5704 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
5705  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5706  int32x4_t __ret;
5707  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
5708  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5709  return __ret;
5710 }
5711 #endif
5712 
5713 #ifdef __LITTLE_ENDIAN__
5714 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
5715  int32x2_t __ret;
5716  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
5717  return __ret;
5718 }
5719 #else
5720 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
5721  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5722  int32x2_t __ret;
5723  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
5724  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5725  return __ret;
5726 }
5727 #endif
5728 
5729 #ifdef __LITTLE_ENDIAN__
5730 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
5731  uint32x4_t __ret;
5732  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
5733  return __ret;
5734 }
5735 #else
5736 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
5737  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5738  uint32x4_t __ret;
5739  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
5740  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5741  return __ret;
5742 }
5743 #endif
5744 
5745 #ifdef __LITTLE_ENDIAN__
5746 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
5747  uint32x2_t __ret;
5748  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
5749  return __ret;
5750 }
5751 #else
5752 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
5753  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5754  uint32x2_t __ret;
5755  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
5756  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5757  return __ret;
5758 }
5759 #endif
5760 
5761 #ifdef __LITTLE_ENDIAN__
5762 #define vdup_lane_p8(__p0_0, __p1_0) __extension__ ({ \
5763  poly8x8_t __s0_0 = __p0_0; \
5764  poly8x8_t __ret_0; \
5765  __ret_0 = splat_lane_p8(__s0_0, __p1_0); \
5766  __ret_0; \
5767 })
5768 #else
5769 #define vdup_lane_p8(__p0_1, __p1_1) __extension__ ({ \
5770  poly8x8_t __s0_1 = __p0_1; \
5771  poly8x8_t __rev0_1; __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 7, 6, 5, 4, 3, 2, 1, 0); \
5772  poly8x8_t __ret_1; \
5773  __ret_1 = __noswap_splat_lane_p8(__rev0_1, __p1_1); \
5774  __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 7, 6, 5, 4, 3, 2, 1, 0); \
5775  __ret_1; \
5776 })
5777 #endif
5778 
5779 #ifdef __LITTLE_ENDIAN__
5780 #define vdup_lane_p16(__p0_2, __p1_2) __extension__ ({ \
5781  poly16x4_t __s0_2 = __p0_2; \
5782  poly16x4_t __ret_2; \
5783  __ret_2 = splat_lane_p16(__s0_2, __p1_2); \
5784  __ret_2; \
5785 })
5786 #else
5787 #define vdup_lane_p16(__p0_3, __p1_3) __extension__ ({ \
5788  poly16x4_t __s0_3 = __p0_3; \
5789  poly16x4_t __rev0_3; __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 3, 2, 1, 0); \
5790  poly16x4_t __ret_3; \
5791  __ret_3 = __noswap_splat_lane_p16(__rev0_3, __p1_3); \
5792  __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 3, 2, 1, 0); \
5793  __ret_3; \
5794 })
5795 #endif
5796 
5797 #ifdef __LITTLE_ENDIAN__
5798 #define vdupq_lane_p8(__p0_4, __p1_4) __extension__ ({ \
5799  poly8x8_t __s0_4 = __p0_4; \
5800  poly8x16_t __ret_4; \
5801  __ret_4 = splatq_lane_p8(__s0_4, __p1_4); \
5802  __ret_4; \
5803 })
5804 #else
5805 #define vdupq_lane_p8(__p0_5, __p1_5) __extension__ ({ \
5806  poly8x8_t __s0_5 = __p0_5; \
5807  poly8x8_t __rev0_5; __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 7, 6, 5, 4, 3, 2, 1, 0); \
5808  poly8x16_t __ret_5; \
5809  __ret_5 = __noswap_splatq_lane_p8(__rev0_5, __p1_5); \
5810  __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5811  __ret_5; \
5812 })
5813 #endif
5814 
5815 #ifdef __LITTLE_ENDIAN__
5816 #define vdupq_lane_p16(__p0_6, __p1_6) __extension__ ({ \
5817  poly16x4_t __s0_6 = __p0_6; \
5818  poly16x8_t __ret_6; \
5819  __ret_6 = splatq_lane_p16(__s0_6, __p1_6); \
5820  __ret_6; \
5821 })
5822 #else
5823 #define vdupq_lane_p16(__p0_7, __p1_7) __extension__ ({ \
5824  poly16x4_t __s0_7 = __p0_7; \
5825  poly16x4_t __rev0_7; __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
5826  poly16x8_t __ret_7; \
5827  __ret_7 = __noswap_splatq_lane_p16(__rev0_7, __p1_7); \
5828  __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 7, 6, 5, 4, 3, 2, 1, 0); \
5829  __ret_7; \
5830 })
5831 #endif
5832 
5833 #ifdef __LITTLE_ENDIAN__
5834 #define vdupq_lane_u8(__p0_8, __p1_8) __extension__ ({ \
5835  uint8x8_t __s0_8 = __p0_8; \
5836  uint8x16_t __ret_8; \
5837  __ret_8 = splatq_lane_u8(__s0_8, __p1_8); \
5838  __ret_8; \
5839 })
5840 #else
5841 #define vdupq_lane_u8(__p0_9, __p1_9) __extension__ ({ \
5842  uint8x8_t __s0_9 = __p0_9; \
5843  uint8x8_t __rev0_9; __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 7, 6, 5, 4, 3, 2, 1, 0); \
5844  uint8x16_t __ret_9; \
5845  __ret_9 = __noswap_splatq_lane_u8(__rev0_9, __p1_9); \
5846  __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5847  __ret_9; \
5848 })
5849 #endif
5850 
5851 #ifdef __LITTLE_ENDIAN__
5852 #define vdupq_lane_u32(__p0_10, __p1_10) __extension__ ({ \
5853  uint32x2_t __s0_10 = __p0_10; \
5854  uint32x4_t __ret_10; \
5855  __ret_10 = splatq_lane_u32(__s0_10, __p1_10); \
5856  __ret_10; \
5857 })
5858 #else
5859 #define vdupq_lane_u32(__p0_11, __p1_11) __extension__ ({ \
5860  uint32x2_t __s0_11 = __p0_11; \
5861  uint32x2_t __rev0_11; __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 1, 0); \
5862  uint32x4_t __ret_11; \
5863  __ret_11 = __noswap_splatq_lane_u32(__rev0_11, __p1_11); \
5864  __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 3, 2, 1, 0); \
5865  __ret_11; \
5866 })
5867 #endif
5868 
5869 #ifdef __LITTLE_ENDIAN__
5870 #define vdupq_lane_u64(__p0_12, __p1_12) __extension__ ({ \
5871  uint64x1_t __s0_12 = __p0_12; \
5872  uint64x2_t __ret_12; \
5873  __ret_12 = splatq_lane_u64(__s0_12, __p1_12); \
5874  __ret_12; \
5875 })
5876 #else
5877 #define vdupq_lane_u64(__p0_13, __p1_13) __extension__ ({ \
5878  uint64x1_t __s0_13 = __p0_13; \
5879  uint64x2_t __ret_13; \
5880  __ret_13 = __noswap_splatq_lane_u64(__s0_13, __p1_13); \
5881  __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 1, 0); \
5882  __ret_13; \
5883 })
5884 #endif
5885 
5886 #ifdef __LITTLE_ENDIAN__
5887 #define vdupq_lane_u16(__p0_14, __p1_14) __extension__ ({ \
5888  uint16x4_t __s0_14 = __p0_14; \
5889  uint16x8_t __ret_14; \
5890  __ret_14 = splatq_lane_u16(__s0_14, __p1_14); \
5891  __ret_14; \
5892 })
5893 #else
5894 #define vdupq_lane_u16(__p0_15, __p1_15) __extension__ ({ \
5895  uint16x4_t __s0_15 = __p0_15; \
5896  uint16x4_t __rev0_15; __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
5897  uint16x8_t __ret_15; \
5898  __ret_15 = __noswap_splatq_lane_u16(__rev0_15, __p1_15); \
5899  __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 7, 6, 5, 4, 3, 2, 1, 0); \
5900  __ret_15; \
5901 })
5902 #endif
5903 
5904 #ifdef __LITTLE_ENDIAN__
5905 #define vdupq_lane_s8(__p0_16, __p1_16) __extension__ ({ \
5906  int8x8_t __s0_16 = __p0_16; \
5907  int8x16_t __ret_16; \
5908  __ret_16 = splatq_lane_s8(__s0_16, __p1_16); \
5909  __ret_16; \
5910 })
5911 #else
5912 #define vdupq_lane_s8(__p0_17, __p1_17) __extension__ ({ \
5913  int8x8_t __s0_17 = __p0_17; \
5914  int8x8_t __rev0_17; __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 7, 6, 5, 4, 3, 2, 1, 0); \
5915  int8x16_t __ret_17; \
5916  __ret_17 = __noswap_splatq_lane_s8(__rev0_17, __p1_17); \
5917  __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5918  __ret_17; \
5919 })
5920 #endif
5921 
5922 #ifdef __LITTLE_ENDIAN__
5923 #define vdupq_lane_f32(__p0_18, __p1_18) __extension__ ({ \
5924  float32x2_t __s0_18 = __p0_18; \
5925  float32x4_t __ret_18; \
5926  __ret_18 = splatq_lane_f32(__s0_18, __p1_18); \
5927  __ret_18; \
5928 })
5929 #else
5930 #define vdupq_lane_f32(__p0_19, __p1_19) __extension__ ({ \
5931  float32x2_t __s0_19 = __p0_19; \
5932  float32x2_t __rev0_19; __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
5933  float32x4_t __ret_19; \
5934  __ret_19 = __noswap_splatq_lane_f32(__rev0_19, __p1_19); \
5935  __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 3, 2, 1, 0); \
5936  __ret_19; \
5937 })
5938 #endif
5939 
5940 #ifdef __LITTLE_ENDIAN__
5941 #define vdupq_lane_s32(__p0_20, __p1_20) __extension__ ({ \
5942  int32x2_t __s0_20 = __p0_20; \
5943  int32x4_t __ret_20; \
5944  __ret_20 = splatq_lane_s32(__s0_20, __p1_20); \
5945  __ret_20; \
5946 })
5947 #else
5948 #define vdupq_lane_s32(__p0_21, __p1_21) __extension__ ({ \
5949  int32x2_t __s0_21 = __p0_21; \
5950  int32x2_t __rev0_21; __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 1, 0); \
5951  int32x4_t __ret_21; \
5952  __ret_21 = __noswap_splatq_lane_s32(__rev0_21, __p1_21); \
5953  __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 3, 2, 1, 0); \
5954  __ret_21; \
5955 })
5956 #endif
5957 
5958 #ifdef __LITTLE_ENDIAN__
5959 #define vdupq_lane_s64(__p0_22, __p1_22) __extension__ ({ \
5960  int64x1_t __s0_22 = __p0_22; \
5961  int64x2_t __ret_22; \
5962  __ret_22 = splatq_lane_s64(__s0_22, __p1_22); \
5963  __ret_22; \
5964 })
5965 #else
5966 #define vdupq_lane_s64(__p0_23, __p1_23) __extension__ ({ \
5967  int64x1_t __s0_23 = __p0_23; \
5968  int64x2_t __ret_23; \
5969  __ret_23 = __noswap_splatq_lane_s64(__s0_23, __p1_23); \
5970  __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 1, 0); \
5971  __ret_23; \
5972 })
5973 #endif
5974 
5975 #ifdef __LITTLE_ENDIAN__
5976 #define vdupq_lane_s16(__p0_24, __p1_24) __extension__ ({ \
5977  int16x4_t __s0_24 = __p0_24; \
5978  int16x8_t __ret_24; \
5979  __ret_24 = splatq_lane_s16(__s0_24, __p1_24); \
5980  __ret_24; \
5981 })
5982 #else
5983 #define vdupq_lane_s16(__p0_25, __p1_25) __extension__ ({ \
5984  int16x4_t __s0_25 = __p0_25; \
5985  int16x4_t __rev0_25; __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
5986  int16x8_t __ret_25; \
5987  __ret_25 = __noswap_splatq_lane_s16(__rev0_25, __p1_25); \
5988  __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 7, 6, 5, 4, 3, 2, 1, 0); \
5989  __ret_25; \
5990 })
5991 #endif
5992 
5993 #ifdef __LITTLE_ENDIAN__
5994 #define vdup_lane_u8(__p0_26, __p1_26) __extension__ ({ \
5995  uint8x8_t __s0_26 = __p0_26; \
5996  uint8x8_t __ret_26; \
5997  __ret_26 = splat_lane_u8(__s0_26, __p1_26); \
5998  __ret_26; \
5999 })
6000 #else
6001 #define vdup_lane_u8(__p0_27, __p1_27) __extension__ ({ \
6002  uint8x8_t __s0_27 = __p0_27; \
6003  uint8x8_t __rev0_27; __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
6004  uint8x8_t __ret_27; \
6005  __ret_27 = __noswap_splat_lane_u8(__rev0_27, __p1_27); \
6006  __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
6007  __ret_27; \
6008 })
6009 #endif
6010 
6011 #ifdef __LITTLE_ENDIAN__
6012 #define vdup_lane_u32(__p0_28, __p1_28) __extension__ ({ \
6013  uint32x2_t __s0_28 = __p0_28; \
6014  uint32x2_t __ret_28; \
6015  __ret_28 = splat_lane_u32(__s0_28, __p1_28); \
6016  __ret_28; \
6017 })
6018 #else
6019 #define vdup_lane_u32(__p0_29, __p1_29) __extension__ ({ \
6020  uint32x2_t __s0_29 = __p0_29; \
6021  uint32x2_t __rev0_29; __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
6022  uint32x2_t __ret_29; \
6023  __ret_29 = __noswap_splat_lane_u32(__rev0_29, __p1_29); \
6024  __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
6025  __ret_29; \
6026 })
6027 #endif
6028 
6029 #define vdup_lane_u64(__p0_30, __p1_30) __extension__ ({ \
6030  uint64x1_t __s0_30 = __p0_30; \
6031  uint64x1_t __ret_30; \
6032  __ret_30 = splat_lane_u64(__s0_30, __p1_30); \
6033  __ret_30; \
6034 })
6035 #ifdef __LITTLE_ENDIAN__
6036 #define vdup_lane_u16(__p0_31, __p1_31) __extension__ ({ \
6037  uint16x4_t __s0_31 = __p0_31; \
6038  uint16x4_t __ret_31; \
6039  __ret_31 = splat_lane_u16(__s0_31, __p1_31); \
6040  __ret_31; \
6041 })
6042 #else
6043 #define vdup_lane_u16(__p0_32, __p1_32) __extension__ ({ \
6044  uint16x4_t __s0_32 = __p0_32; \
6045  uint16x4_t __rev0_32; __rev0_32 = __builtin_shufflevector(__s0_32, __s0_32, 3, 2, 1, 0); \
6046  uint16x4_t __ret_32; \
6047  __ret_32 = __noswap_splat_lane_u16(__rev0_32, __p1_32); \
6048  __ret_32 = __builtin_shufflevector(__ret_32, __ret_32, 3, 2, 1, 0); \
6049  __ret_32; \
6050 })
6051 #endif
6052 
6053 #ifdef __LITTLE_ENDIAN__
6054 #define vdup_lane_s8(__p0_33, __p1_33) __extension__ ({ \
6055  int8x8_t __s0_33 = __p0_33; \
6056  int8x8_t __ret_33; \
6057  __ret_33 = splat_lane_s8(__s0_33, __p1_33); \
6058  __ret_33; \
6059 })
6060 #else
6061 #define vdup_lane_s8(__p0_34, __p1_34) __extension__ ({ \
6062  int8x8_t __s0_34 = __p0_34; \
6063  int8x8_t __rev0_34; __rev0_34 = __builtin_shufflevector(__s0_34, __s0_34, 7, 6, 5, 4, 3, 2, 1, 0); \
6064  int8x8_t __ret_34; \
6065  __ret_34 = __noswap_splat_lane_s8(__rev0_34, __p1_34); \
6066  __ret_34 = __builtin_shufflevector(__ret_34, __ret_34, 7, 6, 5, 4, 3, 2, 1, 0); \
6067  __ret_34; \
6068 })
6069 #endif
6070 
6071 #ifdef __LITTLE_ENDIAN__
6072 #define vdup_lane_f32(__p0_35, __p1_35) __extension__ ({ \
6073  float32x2_t __s0_35 = __p0_35; \
6074  float32x2_t __ret_35; \
6075  __ret_35 = splat_lane_f32(__s0_35, __p1_35); \
6076  __ret_35; \
6077 })
6078 #else
6079 #define vdup_lane_f32(__p0_36, __p1_36) __extension__ ({ \
6080  float32x2_t __s0_36 = __p0_36; \
6081  float32x2_t __rev0_36; __rev0_36 = __builtin_shufflevector(__s0_36, __s0_36, 1, 0); \
6082  float32x2_t __ret_36; \
6083  __ret_36 = __noswap_splat_lane_f32(__rev0_36, __p1_36); \
6084  __ret_36 = __builtin_shufflevector(__ret_36, __ret_36, 1, 0); \
6085  __ret_36; \
6086 })
6087 #endif
6088 
6089 #ifdef __LITTLE_ENDIAN__
6090 #define vdup_lane_s32(__p0_37, __p1_37) __extension__ ({ \
6091  int32x2_t __s0_37 = __p0_37; \
6092  int32x2_t __ret_37; \
6093  __ret_37 = splat_lane_s32(__s0_37, __p1_37); \
6094  __ret_37; \
6095 })
6096 #else
6097 #define vdup_lane_s32(__p0_38, __p1_38) __extension__ ({ \
6098  int32x2_t __s0_38 = __p0_38; \
6099  int32x2_t __rev0_38; __rev0_38 = __builtin_shufflevector(__s0_38, __s0_38, 1, 0); \
6100  int32x2_t __ret_38; \
6101  __ret_38 = __noswap_splat_lane_s32(__rev0_38, __p1_38); \
6102  __ret_38 = __builtin_shufflevector(__ret_38, __ret_38, 1, 0); \
6103  __ret_38; \
6104 })
6105 #endif
6106 
6107 #define vdup_lane_s64(__p0_39, __p1_39) __extension__ ({ \
6108  int64x1_t __s0_39 = __p0_39; \
6109  int64x1_t __ret_39; \
6110  __ret_39 = splat_lane_s64(__s0_39, __p1_39); \
6111  __ret_39; \
6112 })
6113 #ifdef __LITTLE_ENDIAN__
6114 #define vdup_lane_s16(__p0_40, __p1_40) __extension__ ({ \
6115  int16x4_t __s0_40 = __p0_40; \
6116  int16x4_t __ret_40; \
6117  __ret_40 = splat_lane_s16(__s0_40, __p1_40); \
6118  __ret_40; \
6119 })
6120 #else
6121 #define vdup_lane_s16(__p0_41, __p1_41) __extension__ ({ \
6122  int16x4_t __s0_41 = __p0_41; \
6123  int16x4_t __rev0_41; __rev0_41 = __builtin_shufflevector(__s0_41, __s0_41, 3, 2, 1, 0); \
6124  int16x4_t __ret_41; \
6125  __ret_41 = __noswap_splat_lane_s16(__rev0_41, __p1_41); \
6126  __ret_41 = __builtin_shufflevector(__ret_41, __ret_41, 3, 2, 1, 0); \
6127  __ret_41; \
6128 })
6129 #endif
6130 
6131 #ifdef __LITTLE_ENDIAN__
6132 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
6133  poly8x8_t __ret;
6134  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6135  return __ret;
6136 }
6137 #else
6138 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
6139  poly8x8_t __ret;
6140  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6141  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6142  return __ret;
6143 }
6144 #endif
6145 
6146 #ifdef __LITTLE_ENDIAN__
6147 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
6148  poly16x4_t __ret;
6149  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
6150  return __ret;
6151 }
6152 #else
6153 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
6154  poly16x4_t __ret;
6155  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
6156  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6157  return __ret;
6158 }
6159 #endif
6160 
6161 #ifdef __LITTLE_ENDIAN__
6162 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
6163  poly8x16_t __ret;
6164  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6165  return __ret;
6166 }
6167 #else
6168 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
6169  poly8x16_t __ret;
6170  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6171  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6172  return __ret;
6173 }
6174 #endif
6175 
6176 #ifdef __LITTLE_ENDIAN__
6177 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
6178  poly16x8_t __ret;
6179  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6180  return __ret;
6181 }
6182 #else
6183 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
6184  poly16x8_t __ret;
6185  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6186  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6187  return __ret;
6188 }
6189 #endif
6190 
6191 #ifdef __LITTLE_ENDIAN__
6192 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
6193  uint8x16_t __ret;
6194  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6195  return __ret;
6196 }
6197 #else
6198 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
6199  uint8x16_t __ret;
6200  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6201  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6202  return __ret;
6203 }
6204 #endif
6205 
6206 #ifdef __LITTLE_ENDIAN__
6207 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
6208  uint32x4_t __ret;
6209  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
6210  return __ret;
6211 }
6212 #else
6213 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
6214  uint32x4_t __ret;
6215  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
6216  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6217  return __ret;
6218 }
6219 #endif
6220 
6221 #ifdef __LITTLE_ENDIAN__
6222 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
6223  uint64x2_t __ret;
6224  __ret = (uint64x2_t) {__p0, __p0};
6225  return __ret;
6226 }
6227 #else
6228 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
6229  uint64x2_t __ret;
6230  __ret = (uint64x2_t) {__p0, __p0};
6231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6232  return __ret;
6233 }
6234 #endif
6235 
6236 #ifdef __LITTLE_ENDIAN__
6237 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
6238  uint16x8_t __ret;
6239  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6240  return __ret;
6241 }
6242 #else
6243 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
6244  uint16x8_t __ret;
6245  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6246  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6247  return __ret;
6248 }
6249 #endif
6250 
6251 #ifdef __LITTLE_ENDIAN__
6252 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
6253  int8x16_t __ret;
6254  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6255  return __ret;
6256 }
6257 #else
6258 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
6259  int8x16_t __ret;
6260  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6261  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6262  return __ret;
6263 }
6264 #endif
6265 
6266 #ifdef __LITTLE_ENDIAN__
6267 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
6268  float32x4_t __ret;
6269  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
6270  return __ret;
6271 }
6272 #else
6273 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
6274  float32x4_t __ret;
6275  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
6276  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6277  return __ret;
6278 }
6279 #endif
6280 
6281 #ifdef __LITTLE_ENDIAN__
6282 #define vdupq_n_f16(__p0) __extension__ ({ \
6283  float16_t __s0 = __p0; \
6284  float16x8_t __ret; \
6285  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
6286  __ret; \
6287 })
6288 #else
6289 #define vdupq_n_f16(__p0) __extension__ ({ \
6290  float16_t __s0 = __p0; \
6291  float16x8_t __ret; \
6292  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
6293  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6294  __ret; \
6295 })
6296 #endif
6297 
6298 #ifdef __LITTLE_ENDIAN__
6299 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
6300  int32x4_t __ret;
6301  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
6302  return __ret;
6303 }
6304 #else
6305 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
6306  int32x4_t __ret;
6307  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
6308  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6309  return __ret;
6310 }
6311 #endif
6312 
6313 #ifdef __LITTLE_ENDIAN__
6314 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
6315  int64x2_t __ret;
6316  __ret = (int64x2_t) {__p0, __p0};
6317  return __ret;
6318 }
6319 #else
6320 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
6321  int64x2_t __ret;
6322  __ret = (int64x2_t) {__p0, __p0};
6323  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6324  return __ret;
6325 }
6326 #endif
6327 
6328 #ifdef __LITTLE_ENDIAN__
6329 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
6330  int16x8_t __ret;
6331  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6332  return __ret;
6333 }
6334 #else
6335 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
6336  int16x8_t __ret;
6337  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6338  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6339  return __ret;
6340 }
6341 #endif
6342 
6343 #ifdef __LITTLE_ENDIAN__
6344 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
6345  uint8x8_t __ret;
6346  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6347  return __ret;
6348 }
6349 #else
6350 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
6351  uint8x8_t __ret;
6352  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6354  return __ret;
6355 }
6356 #endif
6357 
6358 #ifdef __LITTLE_ENDIAN__
6359 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
6360  uint32x2_t __ret;
6361  __ret = (uint32x2_t) {__p0, __p0};
6362  return __ret;
6363 }
6364 #else
6365 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
6366  uint32x2_t __ret;
6367  __ret = (uint32x2_t) {__p0, __p0};
6368  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6369  return __ret;
6370 }
6371 #endif
6372 
6373 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
6374  uint64x1_t __ret;
6375  __ret = (uint64x1_t) {__p0};
6376  return __ret;
6377 }
6378 #ifdef __LITTLE_ENDIAN__
6379 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
6380  uint16x4_t __ret;
6381  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
6382  return __ret;
6383 }
6384 #else
6385 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
6386  uint16x4_t __ret;
6387  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
6388  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6389  return __ret;
6390 }
6391 #endif
6392 
6393 #ifdef __LITTLE_ENDIAN__
6394 __ai int8x8_t vdup_n_s8(int8_t __p0) {
6395  int8x8_t __ret;
6396  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6397  return __ret;
6398 }
6399 #else
6400 __ai int8x8_t vdup_n_s8(int8_t __p0) {
6401  int8x8_t __ret;
6402  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
6403  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6404  return __ret;
6405 }
6406 #endif
6407 
6408 #ifdef __LITTLE_ENDIAN__
6409 __ai float32x2_t vdup_n_f32(float32_t __p0) {
6410  float32x2_t __ret;
6411  __ret = (float32x2_t) {__p0, __p0};
6412  return __ret;
6413 }
6414 #else
6415 __ai float32x2_t vdup_n_f32(float32_t __p0) {
6416  float32x2_t __ret;
6417  __ret = (float32x2_t) {__p0, __p0};
6418  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6419  return __ret;
6420 }
6421 #endif
6422 
6423 #ifdef __LITTLE_ENDIAN__
6424 #define vdup_n_f16(__p0) __extension__ ({ \
6425  float16_t __s0 = __p0; \
6426  float16x4_t __ret; \
6427  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
6428  __ret; \
6429 })
6430 #else
6431 #define vdup_n_f16(__p0) __extension__ ({ \
6432  float16_t __s0 = __p0; \
6433  float16x4_t __ret; \
6434  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
6435  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6436  __ret; \
6437 })
6438 #endif
6439 
6440 #ifdef __LITTLE_ENDIAN__
6441 __ai int32x2_t vdup_n_s32(int32_t __p0) {
6442  int32x2_t __ret;
6443  __ret = (int32x2_t) {__p0, __p0};
6444  return __ret;
6445 }
6446 #else
6447 __ai int32x2_t vdup_n_s32(int32_t __p0) {
6448  int32x2_t __ret;
6449  __ret = (int32x2_t) {__p0, __p0};
6450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6451  return __ret;
6452 }
6453 #endif
6454 
6455 __ai int64x1_t vdup_n_s64(int64_t __p0) {
6456  int64x1_t __ret;
6457  __ret = (int64x1_t) {__p0};
6458  return __ret;
6459 }
6460 #ifdef __LITTLE_ENDIAN__
6461 __ai int16x4_t vdup_n_s16(int16_t __p0) {
6462  int16x4_t __ret;
6463  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
6464  return __ret;
6465 }
6466 #else
6467 __ai int16x4_t vdup_n_s16(int16_t __p0) {
6468  int16x4_t __ret;
6469  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
6470  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6471  return __ret;
6472 }
6473 #endif
6474 
6475 #ifdef __LITTLE_ENDIAN__
6476 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
6477  uint8x16_t __ret;
6478  __ret = __p0 ^ __p1;
6479  return __ret;
6480 }
6481 #else
6482 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
6483  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6484  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6485  uint8x16_t __ret;
6486  __ret = __rev0 ^ __rev1;
6487  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6488  return __ret;
6489 }
6490 #endif
6491 
6492 #ifdef __LITTLE_ENDIAN__
6493 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
6494  uint32x4_t __ret;
6495  __ret = __p0 ^ __p1;
6496  return __ret;
6497 }
6498 #else
6499 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
6500  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6501  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6502  uint32x4_t __ret;
6503  __ret = __rev0 ^ __rev1;
6504  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6505  return __ret;
6506 }
6507 #endif
6508 
6509 #ifdef __LITTLE_ENDIAN__
6510 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
6511  uint64x2_t __ret;
6512  __ret = __p0 ^ __p1;
6513  return __ret;
6514 }
6515 #else
6516 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
6517  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6518  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6519  uint64x2_t __ret;
6520  __ret = __rev0 ^ __rev1;
6521  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6522  return __ret;
6523 }
6524 #endif
6525 
6526 #ifdef __LITTLE_ENDIAN__
6527 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
6528  uint16x8_t __ret;
6529  __ret = __p0 ^ __p1;
6530  return __ret;
6531 }
6532 #else
6533 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
6534  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6535  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6536  uint16x8_t __ret;
6537  __ret = __rev0 ^ __rev1;
6538  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6539  return __ret;
6540 }
6541 #endif
6542 
6543 #ifdef __LITTLE_ENDIAN__
6544 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
6545  int8x16_t __ret;
6546  __ret = __p0 ^ __p1;
6547  return __ret;
6548 }
6549 #else
6550 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
6551  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6552  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6553  int8x16_t __ret;
6554  __ret = __rev0 ^ __rev1;
6555  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6556  return __ret;
6557 }
6558 #endif
6559 
6560 #ifdef __LITTLE_ENDIAN__
6561 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
6562  int32x4_t __ret;
6563  __ret = __p0 ^ __p1;
6564  return __ret;
6565 }
6566 #else
6567 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
6568  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6569  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6570  int32x4_t __ret;
6571  __ret = __rev0 ^ __rev1;
6572  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6573  return __ret;
6574 }
6575 #endif
6576 
6577 #ifdef __LITTLE_ENDIAN__
6578 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
6579  int64x2_t __ret;
6580  __ret = __p0 ^ __p1;
6581  return __ret;
6582 }
6583 #else
6584 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
6585  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6586  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6587  int64x2_t __ret;
6588  __ret = __rev0 ^ __rev1;
6589  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6590  return __ret;
6591 }
6592 #endif
6593 
6594 #ifdef __LITTLE_ENDIAN__
6595 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
6596  int16x8_t __ret;
6597  __ret = __p0 ^ __p1;
6598  return __ret;
6599 }
6600 #else
6601 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
6602  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6603  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6604  int16x8_t __ret;
6605  __ret = __rev0 ^ __rev1;
6606  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6607  return __ret;
6608 }
6609 #endif
6610 
6611 #ifdef __LITTLE_ENDIAN__
6612 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
6613  uint8x8_t __ret;
6614  __ret = __p0 ^ __p1;
6615  return __ret;
6616 }
6617 #else
6618 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
6619  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6620  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6621  uint8x8_t __ret;
6622  __ret = __rev0 ^ __rev1;
6623  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6624  return __ret;
6625 }
6626 #endif
6627 
6628 #ifdef __LITTLE_ENDIAN__
6629 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
6630  uint32x2_t __ret;
6631  __ret = __p0 ^ __p1;
6632  return __ret;
6633 }
6634 #else
6635 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
6636  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6637  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6638  uint32x2_t __ret;
6639  __ret = __rev0 ^ __rev1;
6640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6641  return __ret;
6642 }
6643 #endif
6644 
6645 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
6646  uint64x1_t __ret;
6647  __ret = __p0 ^ __p1;
6648  return __ret;
6649 }
6650 #ifdef __LITTLE_ENDIAN__
6651 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
6652  uint16x4_t __ret;
6653  __ret = __p0 ^ __p1;
6654  return __ret;
6655 }
6656 #else
6657 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
6658  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6659  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6660  uint16x4_t __ret;
6661  __ret = __rev0 ^ __rev1;
6662  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6663  return __ret;
6664 }
6665 #endif
6666 
6667 #ifdef __LITTLE_ENDIAN__
6668 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
6669  int8x8_t __ret;
6670  __ret = __p0 ^ __p1;
6671  return __ret;
6672 }
6673 #else
6674 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
6675  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6676  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
6677  int8x8_t __ret;
6678  __ret = __rev0 ^ __rev1;
6679  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6680  return __ret;
6681 }
6682 #endif
6683 
6684 #ifdef __LITTLE_ENDIAN__
6685 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
6686  int32x2_t __ret;
6687  __ret = __p0 ^ __p1;
6688  return __ret;
6689 }
6690 #else
6691 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
6692  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6693  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6694  int32x2_t __ret;
6695  __ret = __rev0 ^ __rev1;
6696  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6697  return __ret;
6698 }
6699 #endif
6700 
6701 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
6702  int64x1_t __ret;
6703  __ret = __p0 ^ __p1;
6704  return __ret;
6705 }
6706 #ifdef __LITTLE_ENDIAN__
6707 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
6708  int16x4_t __ret;
6709  __ret = __p0 ^ __p1;
6710  return __ret;
6711 }
6712 #else
6713 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
6714  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6715  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6716  int16x4_t __ret;
6717  __ret = __rev0 ^ __rev1;
6718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6719  return __ret;
6720 }
6721 #endif
6722 
6723 #ifdef __LITTLE_ENDIAN__
6724 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
6725  poly8x8_t __s0 = __p0; \
6726  poly8x8_t __s1 = __p1; \
6727  poly8x8_t __ret; \
6728  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
6729  __ret; \
6730 })
6731 #else
6732 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
6733  poly8x8_t __s0 = __p0; \
6734  poly8x8_t __s1 = __p1; \
6735  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6736  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6737  poly8x8_t __ret; \
6738  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
6739  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6740  __ret; \
6741 })
6742 #endif
6743 
6744 #ifdef __LITTLE_ENDIAN__
6745 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
6746  poly16x4_t __s0 = __p0; \
6747  poly16x4_t __s1 = __p1; \
6748  poly16x4_t __ret; \
6749  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
6750  __ret; \
6751 })
6752 #else
6753 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
6754  poly16x4_t __s0 = __p0; \
6755  poly16x4_t __s1 = __p1; \
6756  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6757  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6758  poly16x4_t __ret; \
6759  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
6760  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6761  __ret; \
6762 })
6763 #endif
6764 
6765 #ifdef __LITTLE_ENDIAN__
6766 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
6767  poly8x16_t __s0 = __p0; \
6768  poly8x16_t __s1 = __p1; \
6769  poly8x16_t __ret; \
6770  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
6771  __ret; \
6772 })
6773 #else
6774 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
6775  poly8x16_t __s0 = __p0; \
6776  poly8x16_t __s1 = __p1; \
6777  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6778  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6779  poly8x16_t __ret; \
6780  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
6781  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6782  __ret; \
6783 })
6784 #endif
6785 
6786 #ifdef __LITTLE_ENDIAN__
6787 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
6788  poly16x8_t __s0 = __p0; \
6789  poly16x8_t __s1 = __p1; \
6790  poly16x8_t __ret; \
6791  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
6792  __ret; \
6793 })
6794 #else
6795 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
6796  poly16x8_t __s0 = __p0; \
6797  poly16x8_t __s1 = __p1; \
6798  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6799  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6800  poly16x8_t __ret; \
6801  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
6802  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6803  __ret; \
6804 })
6805 #endif
6806 
6807 #ifdef __LITTLE_ENDIAN__
6808 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
6809  uint8x16_t __s0 = __p0; \
6810  uint8x16_t __s1 = __p1; \
6811  uint8x16_t __ret; \
6812  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
6813  __ret; \
6814 })
6815 #else
6816 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
6817  uint8x16_t __s0 = __p0; \
6818  uint8x16_t __s1 = __p1; \
6819  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6820  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6821  uint8x16_t __ret; \
6822  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
6823  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6824  __ret; \
6825 })
6826 #endif
6827 
6828 #ifdef __LITTLE_ENDIAN__
6829 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
6830  uint32x4_t __s0 = __p0; \
6831  uint32x4_t __s1 = __p1; \
6832  uint32x4_t __ret; \
6833  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
6834  __ret; \
6835 })
6836 #else
6837 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
6838  uint32x4_t __s0 = __p0; \
6839  uint32x4_t __s1 = __p1; \
6840  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6841  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6842  uint32x4_t __ret; \
6843  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
6844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6845  __ret; \
6846 })
6847 #endif
6848 
6849 #ifdef __LITTLE_ENDIAN__
6850 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
6851  uint64x2_t __s0 = __p0; \
6852  uint64x2_t __s1 = __p1; \
6853  uint64x2_t __ret; \
6854  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
6855  __ret; \
6856 })
6857 #else
6858 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
6859  uint64x2_t __s0 = __p0; \
6860  uint64x2_t __s1 = __p1; \
6861  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6862  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6863  uint64x2_t __ret; \
6864  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
6865  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6866  __ret; \
6867 })
6868 #endif
6869 
6870 #ifdef __LITTLE_ENDIAN__
6871 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
6872  uint16x8_t __s0 = __p0; \
6873  uint16x8_t __s1 = __p1; \
6874  uint16x8_t __ret; \
6875  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
6876  __ret; \
6877 })
6878 #else
6879 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
6880  uint16x8_t __s0 = __p0; \
6881  uint16x8_t __s1 = __p1; \
6882  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6883  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6884  uint16x8_t __ret; \
6885  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
6886  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6887  __ret; \
6888 })
6889 #endif
6890 
6891 #ifdef __LITTLE_ENDIAN__
6892 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
6893  int8x16_t __s0 = __p0; \
6894  int8x16_t __s1 = __p1; \
6895  int8x16_t __ret; \
6896  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
6897  __ret; \
6898 })
6899 #else
6900 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
6901  int8x16_t __s0 = __p0; \
6902  int8x16_t __s1 = __p1; \
6903  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6904  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6905  int8x16_t __ret; \
6906  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
6907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6908  __ret; \
6909 })
6910 #endif
6911 
6912 #ifdef __LITTLE_ENDIAN__
6913 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
6914  float32x4_t __s0 = __p0; \
6915  float32x4_t __s1 = __p1; \
6916  float32x4_t __ret; \
6917  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
6918  __ret; \
6919 })
6920 #else
6921 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
6922  float32x4_t __s0 = __p0; \
6923  float32x4_t __s1 = __p1; \
6924  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6925  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6926  float32x4_t __ret; \
6927  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
6928  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6929  __ret; \
6930 })
6931 #endif
6932 
6933 #ifdef __LITTLE_ENDIAN__
6934 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
6935  int32x4_t __s0 = __p0; \
6936  int32x4_t __s1 = __p1; \
6937  int32x4_t __ret; \
6938  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
6939  __ret; \
6940 })
6941 #else
6942 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
6943  int32x4_t __s0 = __p0; \
6944  int32x4_t __s1 = __p1; \
6945  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6946  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6947  int32x4_t __ret; \
6948  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
6949  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6950  __ret; \
6951 })
6952 #endif
6953 
6954 #ifdef __LITTLE_ENDIAN__
6955 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
6956  int64x2_t __s0 = __p0; \
6957  int64x2_t __s1 = __p1; \
6958  int64x2_t __ret; \
6959  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
6960  __ret; \
6961 })
6962 #else
6963 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
6964  int64x2_t __s0 = __p0; \
6965  int64x2_t __s1 = __p1; \
6966  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6967  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6968  int64x2_t __ret; \
6969  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
6970  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6971  __ret; \
6972 })
6973 #endif
6974 
6975 #ifdef __LITTLE_ENDIAN__
6976 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
6977  int16x8_t __s0 = __p0; \
6978  int16x8_t __s1 = __p1; \
6979  int16x8_t __ret; \
6980  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
6981  __ret; \
6982 })
6983 #else
6984 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
6985  int16x8_t __s0 = __p0; \
6986  int16x8_t __s1 = __p1; \
6987  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6988  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6989  int16x8_t __ret; \
6990  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
6991  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6992  __ret; \
6993 })
6994 #endif
6995 
6996 #ifdef __LITTLE_ENDIAN__
6997 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
6998  uint8x8_t __s0 = __p0; \
6999  uint8x8_t __s1 = __p1; \
7000  uint8x8_t __ret; \
7001  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
7002  __ret; \
7003 })
7004 #else
7005 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
7006  uint8x8_t __s0 = __p0; \
7007  uint8x8_t __s1 = __p1; \
7008  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7009  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
7010  uint8x8_t __ret; \
7011  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
7012  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7013  __ret; \
7014 })
7015 #endif
7016 
7017 #ifdef __LITTLE_ENDIAN__
7018 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
7019  uint32x2_t __s0 = __p0; \
7020  uint32x2_t __s1 = __p1; \
7021  uint32x2_t __ret; \
7022  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
7023  __ret; \
7024 })
7025 #else
7026 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
7027  uint32x2_t __s0 = __p0; \
7028  uint32x2_t __s1 = __p1; \
7029  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7030  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
7031  uint32x2_t __ret; \
7032  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
7033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7034  __ret; \
7035 })
7036 #endif
7037 
7038 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
7039  uint64x1_t __s0 = __p0; \
7040  uint64x1_t __s1 = __p1; \
7041  uint64x1_t __ret; \
7042  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
7043  __ret; \
7044 })
7045 #ifdef __LITTLE_ENDIAN__
7046 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
7047  uint16x4_t __s0 = __p0; \
7048  uint16x4_t __s1 = __p1; \
7049  uint16x4_t __ret; \
7050  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
7051  __ret; \
7052 })
7053 #else
7054 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
7055  uint16x4_t __s0 = __p0; \
7056  uint16x4_t __s1 = __p1; \
7057  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7058  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
7059  uint16x4_t __ret; \
7060  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
7061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7062  __ret; \
7063 })
7064 #endif
7065 
7066 #ifdef __LITTLE_ENDIAN__
7067 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
7068  int8x8_t __s0 = __p0; \
7069  int8x8_t __s1 = __p1; \
7070  int8x8_t __ret; \
7071  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
7072  __ret; \
7073 })
7074 #else
7075 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
7076  int8x8_t __s0 = __p0; \
7077  int8x8_t __s1 = __p1; \
7078  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7079  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
7080  int8x8_t __ret; \
7081  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
7082  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7083  __ret; \
7084 })
7085 #endif
7086 
7087 #ifdef __LITTLE_ENDIAN__
7088 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
7089  float32x2_t __s0 = __p0; \
7090  float32x2_t __s1 = __p1; \
7091  float32x2_t __ret; \
7092  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
7093  __ret; \
7094 })
7095 #else
7096 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
7097  float32x2_t __s0 = __p0; \
7098  float32x2_t __s1 = __p1; \
7099  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7100  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
7101  float32x2_t __ret; \
7102  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
7103  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7104  __ret; \
7105 })
7106 #endif
7107 
7108 #ifdef __LITTLE_ENDIAN__
7109 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
7110  int32x2_t __s0 = __p0; \
7111  int32x2_t __s1 = __p1; \
7112  int32x2_t __ret; \
7113  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
7114  __ret; \
7115 })
7116 #else
7117 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
7118  int32x2_t __s0 = __p0; \
7119  int32x2_t __s1 = __p1; \
7120  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7121  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
7122  int32x2_t __ret; \
7123  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
7124  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7125  __ret; \
7126 })
7127 #endif
7128 
7129 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
7130  int64x1_t __s0 = __p0; \
7131  int64x1_t __s1 = __p1; \
7132  int64x1_t __ret; \
7133  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
7134  __ret; \
7135 })
7136 #ifdef __LITTLE_ENDIAN__
7137 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
7138  int16x4_t __s0 = __p0; \
7139  int16x4_t __s1 = __p1; \
7140  int16x4_t __ret; \
7141  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
7142  __ret; \
7143 })
7144 #else
7145 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
7146  int16x4_t __s0 = __p0; \
7147  int16x4_t __s1 = __p1; \
7148  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7149  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
7150  int16x4_t __ret; \
7151  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
7152  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7153  __ret; \
7154 })
7155 #endif
7156 
7157 #ifdef __LITTLE_ENDIAN__
7158 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
7159  poly8x8_t __ret;
7160  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7161  return __ret;
7162 }
7163 #else
7164 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
7165  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7166  poly8x8_t __ret;
7167  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7168  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7169  return __ret;
7170 }
7171 __ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
7172  poly8x8_t __ret;
7173  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7174  return __ret;
7175 }
7176 #endif
7177 
7178 #ifdef __LITTLE_ENDIAN__
7179 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
7180  poly16x4_t __ret;
7181  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7182  return __ret;
7183 }
7184 #else
7185 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
7186  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7187  poly16x4_t __ret;
7188  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7189  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7190  return __ret;
7191 }
7192 #endif
7193 
7194 #ifdef __LITTLE_ENDIAN__
7195 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
7196  uint8x8_t __ret;
7197  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7198  return __ret;
7199 }
7200 #else
7201 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
7202  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7203  uint8x8_t __ret;
7204  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7205  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7206  return __ret;
7207 }
7208 __ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
7209  uint8x8_t __ret;
7210  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7211  return __ret;
7212 }
7213 #endif
7214 
7215 #ifdef __LITTLE_ENDIAN__
7216 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
7217  uint32x2_t __ret;
7218  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7219  return __ret;
7220 }
7221 #else
7222 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
7223  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7224  uint32x2_t __ret;
7225  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7227  return __ret;
7228 }
7229 __ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
7230  uint32x2_t __ret;
7231  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7232  return __ret;
7233 }
7234 #endif
7235 
7236 #ifdef __LITTLE_ENDIAN__
7237 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
7238  uint64x1_t __ret;
7239  __ret = __builtin_shufflevector(__p0, __p0, 1);
7240  return __ret;
7241 }
7242 #else
7243 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
7244  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7245  uint64x1_t __ret;
7246  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
7247  return __ret;
7248 }
7249 #endif
7250 
7251 #ifdef __LITTLE_ENDIAN__
7252 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
7253  uint16x4_t __ret;
7254  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7255  return __ret;
7256 }
7257 #else
7258 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
7259  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7260  uint16x4_t __ret;
7261  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7262  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7263  return __ret;
7264 }
7265 __ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
7266  uint16x4_t __ret;
7267  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7268  return __ret;
7269 }
7270 #endif
7271 
7272 #ifdef __LITTLE_ENDIAN__
7273 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
7274  int8x8_t __ret;
7275  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7276  return __ret;
7277 }
7278 #else
7279 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
7280  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7281  int8x8_t __ret;
7282  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
7283  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7284  return __ret;
7285 }
7286 __ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
7287  int8x8_t __ret;
7288  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
7289  return __ret;
7290 }
7291 #endif
7292 
7293 #ifdef __LITTLE_ENDIAN__
7294 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
7295  float32x2_t __ret;
7296  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7297  return __ret;
7298 }
7299 #else
7300 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
7301  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7302  float32x2_t __ret;
7303  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7305  return __ret;
7306 }
7307 __ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
7308  float32x2_t __ret;
7309  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7310  return __ret;
7311 }
7312 #endif
7313 
7314 #ifdef __LITTLE_ENDIAN__
7315 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
7316  float16x4_t __ret;
7317  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7318  return __ret;
7319 }
7320 #else
7321 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
7322  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7323  float16x4_t __ret;
7324  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7325  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7326  return __ret;
7327 }
7328 __ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
7329  float16x4_t __ret;
7330  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7331  return __ret;
7332 }
7333 #endif
7334 
7335 #ifdef __LITTLE_ENDIAN__
7336 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
7337  int32x2_t __ret;
7338  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7339  return __ret;
7340 }
7341 #else
7342 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
7343  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7344  int32x2_t __ret;
7345  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
7346  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7347  return __ret;
7348 }
7349 __ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
7350  int32x2_t __ret;
7351  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
7352  return __ret;
7353 }
7354 #endif
7355 
7356 #ifdef __LITTLE_ENDIAN__
7357 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
7358  int64x1_t __ret;
7359  __ret = __builtin_shufflevector(__p0, __p0, 1);
7360  return __ret;
7361 }
7362 #else
7363 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
7364  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7365  int64x1_t __ret;
7366  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
7367  return __ret;
7368 }
7369 #endif
7370 
7371 #ifdef __LITTLE_ENDIAN__
7372 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
7373  int16x4_t __ret;
7374  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7375  return __ret;
7376 }
7377 #else
7378 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
7379  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7380  int16x4_t __ret;
7381  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
7382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7383  return __ret;
7384 }
7385 __ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
7386  int16x4_t __ret;
7387  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
7388  return __ret;
7389 }
7390 #endif
7391 
7392 #ifdef __LITTLE_ENDIAN__
7393 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
7394  poly8x8_t __s0 = __p0; \
7395  poly8_t __ret; \
7396  __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__s0, __p1); \
7397  __ret; \
7398 })
7399 #else
7400 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
7401  poly8x8_t __s0 = __p0; \
7402  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7403  poly8_t __ret; \
7404  __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__rev0, __p1); \
7405  __ret; \
7406 })
7407 #define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
7408  poly8x8_t __s0 = __p0; \
7409  poly8_t __ret; \
7410  __ret = (poly8_t) __builtin_neon_vget_lane_i8((poly8x8_t)__s0, __p1); \
7411  __ret; \
7412 })
7413 #endif
7414 
7415 #ifdef __LITTLE_ENDIAN__
7416 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
7417  poly16x4_t __s0 = __p0; \
7418  poly16_t __ret; \
7419  __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__s0, __p1); \
7420  __ret; \
7421 })
7422 #else
7423 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
7424  poly16x4_t __s0 = __p0; \
7425  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7426  poly16_t __ret; \
7427  __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__rev0, __p1); \
7428  __ret; \
7429 })
7430 #define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
7431  poly16x4_t __s0 = __p0; \
7432  poly16_t __ret; \
7433  __ret = (poly16_t) __builtin_neon_vget_lane_i16((poly16x4_t)__s0, __p1); \
7434  __ret; \
7435 })
7436 #endif
7437 
7438 #ifdef __LITTLE_ENDIAN__
7439 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7440  poly8x16_t __s0 = __p0; \
7441  poly8_t __ret; \
7442  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__s0, __p1); \
7443  __ret; \
7444 })
7445 #else
7446 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7447  poly8x16_t __s0 = __p0; \
7448  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7449  poly8_t __ret; \
7450  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__rev0, __p1); \
7451  __ret; \
7452 })
7453 #define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
7454  poly8x16_t __s0 = __p0; \
7455  poly8_t __ret; \
7456  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((poly8x16_t)__s0, __p1); \
7457  __ret; \
7458 })
7459 #endif
7460 
7461 #ifdef __LITTLE_ENDIAN__
7462 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7463  poly16x8_t __s0 = __p0; \
7464  poly16_t __ret; \
7465  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__s0, __p1); \
7466  __ret; \
7467 })
7468 #else
7469 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7470  poly16x8_t __s0 = __p0; \
7471  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7472  poly16_t __ret; \
7473  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__rev0, __p1); \
7474  __ret; \
7475 })
7476 #define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
7477  poly16x8_t __s0 = __p0; \
7478  poly16_t __ret; \
7479  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((poly16x8_t)__s0, __p1); \
7480  __ret; \
7481 })
7482 #endif
7483 
7484 #ifdef __LITTLE_ENDIAN__
7485 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7486  uint8x16_t __s0 = __p0; \
7487  uint8_t __ret; \
7488  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7489  __ret; \
7490 })
7491 #else
7492 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7493  uint8x16_t __s0 = __p0; \
7494  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7495  uint8_t __ret; \
7496  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
7497  __ret; \
7498 })
7499 #define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
7500  uint8x16_t __s0 = __p0; \
7501  uint8_t __ret; \
7502  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7503  __ret; \
7504 })
7505 #endif
7506 
7507 #ifdef __LITTLE_ENDIAN__
7508 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7509  uint32x4_t __s0 = __p0; \
7510  uint32_t __ret; \
7511  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7512  __ret; \
7513 })
7514 #else
7515 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7516  uint32x4_t __s0 = __p0; \
7517  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7518  uint32_t __ret; \
7519  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__rev0, __p1); \
7520  __ret; \
7521 })
7522 #define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
7523  uint32x4_t __s0 = __p0; \
7524  uint32_t __ret; \
7525  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7526  __ret; \
7527 })
7528 #endif
7529 
7530 #ifdef __LITTLE_ENDIAN__
7531 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7532  uint64x2_t __s0 = __p0; \
7533  uint64_t __ret; \
7534  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7535  __ret; \
7536 })
7537 #else
7538 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7539  uint64x2_t __s0 = __p0; \
7540  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7541  uint64_t __ret; \
7542  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__rev0, __p1); \
7543  __ret; \
7544 })
7545 #define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
7546  uint64x2_t __s0 = __p0; \
7547  uint64_t __ret; \
7548  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7549  __ret; \
7550 })
7551 #endif
7552 
7553 #ifdef __LITTLE_ENDIAN__
7554 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7555  uint16x8_t __s0 = __p0; \
7556  uint16_t __ret; \
7557  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7558  __ret; \
7559 })
7560 #else
7561 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7562  uint16x8_t __s0 = __p0; \
7563  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7564  uint16_t __ret; \
7565  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__rev0, __p1); \
7566  __ret; \
7567 })
7568 #define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
7569  uint16x8_t __s0 = __p0; \
7570  uint16_t __ret; \
7571  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7572  __ret; \
7573 })
7574 #endif
7575 
7576 #ifdef __LITTLE_ENDIAN__
7577 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7578  int8x16_t __s0 = __p0; \
7579  int8_t __ret; \
7580  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7581  __ret; \
7582 })
7583 #else
7584 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7585  int8x16_t __s0 = __p0; \
7586  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7587  int8_t __ret; \
7588  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
7589  __ret; \
7590 })
7591 #define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
7592  int8x16_t __s0 = __p0; \
7593  int8_t __ret; \
7594  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
7595  __ret; \
7596 })
7597 #endif
7598 
7599 #ifdef __LITTLE_ENDIAN__
7600 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7601  float32x4_t __s0 = __p0; \
7602  float32_t __ret; \
7603  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__s0, __p1); \
7604  __ret; \
7605 })
7606 #else
7607 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7608  float32x4_t __s0 = __p0; \
7609  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7610  float32_t __ret; \
7611  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__rev0, __p1); \
7612  __ret; \
7613 })
7614 #define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
7615  float32x4_t __s0 = __p0; \
7616  float32_t __ret; \
7617  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((float32x4_t)__s0, __p1); \
7618  __ret; \
7619 })
7620 #endif
7621 
7622 #ifdef __LITTLE_ENDIAN__
7623 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7624  int32x4_t __s0 = __p0; \
7625  int32_t __ret; \
7626  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7627  __ret; \
7628 })
7629 #else
7630 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7631  int32x4_t __s0 = __p0; \
7632  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7633  int32_t __ret; \
7634  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__rev0, __p1); \
7635  __ret; \
7636 })
7637 #define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
7638  int32x4_t __s0 = __p0; \
7639  int32_t __ret; \
7640  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int32x4_t)__s0, __p1); \
7641  __ret; \
7642 })
7643 #endif
7644 
7645 #ifdef __LITTLE_ENDIAN__
7646 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7647  int64x2_t __s0 = __p0; \
7648  int64_t __ret; \
7649  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7650  __ret; \
7651 })
7652 #else
7653 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7654  int64x2_t __s0 = __p0; \
7655  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7656  int64_t __ret; \
7657  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__rev0, __p1); \
7658  __ret; \
7659 })
7660 #define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
7661  int64x2_t __s0 = __p0; \
7662  int64_t __ret; \
7663  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int64x2_t)__s0, __p1); \
7664  __ret; \
7665 })
7666 #endif
7667 
7668 #ifdef __LITTLE_ENDIAN__
7669 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7670  int16x8_t __s0 = __p0; \
7671  int16_t __ret; \
7672  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7673  __ret; \
7674 })
7675 #else
7676 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7677  int16x8_t __s0 = __p0; \
7678  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7679  int16_t __ret; \
7680  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__rev0, __p1); \
7681  __ret; \
7682 })
7683 #define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
7684  int16x8_t __s0 = __p0; \
7685  int16_t __ret; \
7686  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int16x8_t)__s0, __p1); \
7687  __ret; \
7688 })
7689 #endif
7690 
7691 #ifdef __LITTLE_ENDIAN__
7692 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
7693  uint8x8_t __s0 = __p0; \
7694  uint8_t __ret; \
7695  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7696  __ret; \
7697 })
7698 #else
7699 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
7700  uint8x8_t __s0 = __p0; \
7701  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7702  uint8_t __ret; \
7703  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
7704  __ret; \
7705 })
7706 #define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
7707  uint8x8_t __s0 = __p0; \
7708  uint8_t __ret; \
7709  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7710  __ret; \
7711 })
7712 #endif
7713 
7714 #ifdef __LITTLE_ENDIAN__
7715 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
7716  uint32x2_t __s0 = __p0; \
7717  uint32_t __ret; \
7718  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7719  __ret; \
7720 })
7721 #else
7722 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
7723  uint32x2_t __s0 = __p0; \
7724  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7725  uint32_t __ret; \
7726  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__rev0, __p1); \
7727  __ret; \
7728 })
7729 #define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
7730  uint32x2_t __s0 = __p0; \
7731  uint32_t __ret; \
7732  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7733  __ret; \
7734 })
7735 #endif
7736 
7737 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
7738  uint64x1_t __s0 = __p0; \
7739  uint64_t __ret; \
7740  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int64x1_t)__s0, __p1); \
7741  __ret; \
7742 })
7743 #ifdef __LITTLE_ENDIAN__
7744 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
7745  uint16x4_t __s0 = __p0; \
7746  uint16_t __ret; \
7747  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7748  __ret; \
7749 })
7750 #else
7751 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
7752  uint16x4_t __s0 = __p0; \
7753  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7754  uint16_t __ret; \
7755  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__rev0, __p1); \
7756  __ret; \
7757 })
7758 #define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
7759  uint16x4_t __s0 = __p0; \
7760  uint16_t __ret; \
7761  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7762  __ret; \
7763 })
7764 #endif
7765 
7766 #ifdef __LITTLE_ENDIAN__
7767 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
7768  int8x8_t __s0 = __p0; \
7769  int8_t __ret; \
7770  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7771  __ret; \
7772 })
7773 #else
7774 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
7775  int8x8_t __s0 = __p0; \
7776  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
7777  int8_t __ret; \
7778  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
7779  __ret; \
7780 })
7781 #define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
7782  int8x8_t __s0 = __p0; \
7783  int8_t __ret; \
7784  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
7785  __ret; \
7786 })
7787 #endif
7788 
7789 #ifdef __LITTLE_ENDIAN__
7790 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
7791  float32x2_t __s0 = __p0; \
7792  float32_t __ret; \
7793  __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__s0, __p1); \
7794  __ret; \
7795 })
7796 #else
7797 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
7798  float32x2_t __s0 = __p0; \
7799  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7800  float32_t __ret; \
7801  __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__rev0, __p1); \
7802  __ret; \
7803 })
7804 #define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
7805  float32x2_t __s0 = __p0; \
7806  float32_t __ret; \
7807  __ret = (float32_t) __builtin_neon_vget_lane_f32((float32x2_t)__s0, __p1); \
7808  __ret; \
7809 })
7810 #endif
7811 
7812 #ifdef __LITTLE_ENDIAN__
7813 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
7814  int32x2_t __s0 = __p0; \
7815  int32_t __ret; \
7816  __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7817  __ret; \
7818 })
7819 #else
7820 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
7821  int32x2_t __s0 = __p0; \
7822  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
7823  int32_t __ret; \
7824  __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__rev0, __p1); \
7825  __ret; \
7826 })
7827 #define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
7828  int32x2_t __s0 = __p0; \
7829  int32_t __ret; \
7830  __ret = (int32_t) __builtin_neon_vget_lane_i32((int32x2_t)__s0, __p1); \
7831  __ret; \
7832 })
7833 #endif
7834 
7835 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
7836  int64x1_t __s0 = __p0; \
7837  int64_t __ret; \
7838  __ret = (int64_t) __builtin_neon_vget_lane_i64((int64x1_t)__s0, __p1); \
7839  __ret; \
7840 })
7841 #ifdef __LITTLE_ENDIAN__
7842 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
7843  int16x4_t __s0 = __p0; \
7844  int16_t __ret; \
7845  __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7846  __ret; \
7847 })
7848 #else
7849 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
7850  int16x4_t __s0 = __p0; \
7851  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
7852  int16_t __ret; \
7853  __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__rev0, __p1); \
7854  __ret; \
7855 })
7856 #define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
7857  int16x4_t __s0 = __p0; \
7858  int16_t __ret; \
7859  __ret = (int16_t) __builtin_neon_vget_lane_i16((int16x4_t)__s0, __p1); \
7860  __ret; \
7861 })
7862 #endif
7863 
7864 #ifdef __LITTLE_ENDIAN__
7865 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
7866  poly8x8_t __ret;
7867  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7868  return __ret;
7869 }
7870 #else
7871 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
7872  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7873  poly8x8_t __ret;
7874  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7875  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7876  return __ret;
7877 }
7878 #endif
7879 
7880 #ifdef __LITTLE_ENDIAN__
7881 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
7882  poly16x4_t __ret;
7883  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7884  return __ret;
7885 }
7886 #else
7887 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
7888  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7889  poly16x4_t __ret;
7890  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7891  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7892  return __ret;
7893 }
7894 #endif
7895 
7896 #ifdef __LITTLE_ENDIAN__
7897 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
7898  uint8x8_t __ret;
7899  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7900  return __ret;
7901 }
7902 #else
7903 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
7904  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7905  uint8x8_t __ret;
7906  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7907  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7908  return __ret;
7909 }
7910 #endif
7911 
7912 #ifdef __LITTLE_ENDIAN__
7913 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
7914  uint32x2_t __ret;
7915  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7916  return __ret;
7917 }
7918 #else
7919 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
7920  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7921  uint32x2_t __ret;
7922  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7923  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7924  return __ret;
7925 }
7926 #endif
7927 
7928 #ifdef __LITTLE_ENDIAN__
7929 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7930  uint64x1_t __ret;
7931  __ret = __builtin_shufflevector(__p0, __p0, 0);
7932  return __ret;
7933 }
7934 #else
7935 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7936  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7937  uint64x1_t __ret;
7938  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7939  return __ret;
7940 }
7941 #endif
7942 
7943 #ifdef __LITTLE_ENDIAN__
7944 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7945  uint16x4_t __ret;
7946  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7947  return __ret;
7948 }
7949 #else
7950 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7951  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7952  uint16x4_t __ret;
7953  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7954  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7955  return __ret;
7956 }
7957 #endif
7958 
7959 #ifdef __LITTLE_ENDIAN__
7960 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7961  int8x8_t __ret;
7962  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7963  return __ret;
7964 }
7965 #else
7966 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7967  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7968  int8x8_t __ret;
7969  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7971  return __ret;
7972 }
7973 #endif
7974 
7975 #ifdef __LITTLE_ENDIAN__
7976 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7977  float32x2_t __ret;
7978  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7979  return __ret;
7980 }
7981 #else
7982 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7983  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7984  float32x2_t __ret;
7985  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7986  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7987  return __ret;
7988 }
7989 #endif
7990 
7991 #ifdef __LITTLE_ENDIAN__
7992 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7993  float16x4_t __ret;
7994  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7995  return __ret;
7996 }
7997 #else
7998 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7999  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8000  float16x4_t __ret;
8001  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
8002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8003  return __ret;
8004 }
8005 #endif
8006 
8007 #ifdef __LITTLE_ENDIAN__
8008 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
8009  int32x2_t __ret;
8010  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
8011  return __ret;
8012 }
8013 #else
8014 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
8015  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8016  int32x2_t __ret;
8017  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
8018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8019  return __ret;
8020 }
8021 #endif
8022 
8023 #ifdef __LITTLE_ENDIAN__
8024 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
8025  int64x1_t __ret;
8026  __ret = __builtin_shufflevector(__p0, __p0, 0);
8027  return __ret;
8028 }
8029 #else
8030 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
8031  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8032  int64x1_t __ret;
8033  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
8034  return __ret;
8035 }
8036 #endif
8037 
8038 #ifdef __LITTLE_ENDIAN__
8039 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
8040  int16x4_t __ret;
8041  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
8042  return __ret;
8043 }
8044 #else
8045 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
8046  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8047  int16x4_t __ret;
8048  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
8049  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8050  return __ret;
8051 }
8052 #endif
8053 
8054 #ifdef __LITTLE_ENDIAN__
8055 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8056  uint8x16_t __ret;
8057  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
8058  return __ret;
8059 }
8060 #else
8061 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8062  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8063  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8064  uint8x16_t __ret;
8065  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
8066  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8067  return __ret;
8068 }
8069 #endif
8070 
8071 #ifdef __LITTLE_ENDIAN__
8072 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8073  uint32x4_t __ret;
8074  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
8075  return __ret;
8076 }
8077 #else
8078 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8079  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8080  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8081  uint32x4_t __ret;
8082  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
8083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8084  return __ret;
8085 }
8086 #endif
8087 
8088 #ifdef __LITTLE_ENDIAN__
8089 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8090  uint16x8_t __ret;
8091  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
8092  return __ret;
8093 }
8094 #else
8095 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8096  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8097  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8098  uint16x8_t __ret;
8099  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
8100  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8101  return __ret;
8102 }
8103 #endif
8104 
8105 #ifdef __LITTLE_ENDIAN__
8106 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
8107  int8x16_t __ret;
8108  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
8109  return __ret;
8110 }
8111 #else
8112 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
8113  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8114  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8115  int8x16_t __ret;
8116  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
8117  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8118  return __ret;
8119 }
8120 #endif
8121 
8122 #ifdef __LITTLE_ENDIAN__
8123 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
8124  int32x4_t __ret;
8125  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
8126  return __ret;
8127 }
8128 #else
8129 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
8130  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8131  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8132  int32x4_t __ret;
8133  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
8134  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8135  return __ret;
8136 }
8137 #endif
8138 
8139 #ifdef __LITTLE_ENDIAN__
8140 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
8141  int16x8_t __ret;
8142  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
8143  return __ret;
8144 }
8145 #else
8146 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
8147  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8148  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8149  int16x8_t __ret;
8150  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
8151  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8152  return __ret;
8153 }
8154 #endif
8155 
8156 #ifdef __LITTLE_ENDIAN__
8157 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
8158  uint8x8_t __ret;
8159  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
8160  return __ret;
8161 }
8162 #else
8163 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
8164  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8165  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8166  uint8x8_t __ret;
8167  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
8168  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8169  return __ret;
8170 }
8171 #endif
8172 
8173 #ifdef __LITTLE_ENDIAN__
8174 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
8175  uint32x2_t __ret;
8176  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
8177  return __ret;
8178 }
8179 #else
8180 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
8181  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8182  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8183  uint32x2_t __ret;
8184  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
8185  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8186  return __ret;
8187 }
8188 #endif
8189 
8190 #ifdef __LITTLE_ENDIAN__
8191 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
8192  uint16x4_t __ret;
8193  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
8194  return __ret;
8195 }
8196 #else
8197 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
8198  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8199  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8200  uint16x4_t __ret;
8201  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
8202  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8203  return __ret;
8204 }
8205 #endif
8206 
8207 #ifdef __LITTLE_ENDIAN__
8208 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
8209  int8x8_t __ret;
8210  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
8211  return __ret;
8212 }
8213 #else
8214 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
8215  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8216  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8217  int8x8_t __ret;
8218  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
8219  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8220  return __ret;
8221 }
8222 #endif
8223 
8224 #ifdef __LITTLE_ENDIAN__
8225 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
8226  int32x2_t __ret;
8227  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
8228  return __ret;
8229 }
8230 #else
8231 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
8232  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8233  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8234  int32x2_t __ret;
8235  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
8236  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8237  return __ret;
8238 }
8239 #endif
8240 
8241 #ifdef __LITTLE_ENDIAN__
8242 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
8243  int16x4_t __ret;
8244  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
8245  return __ret;
8246 }
8247 #else
8248 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
8249  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8250  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8251  int16x4_t __ret;
8252  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
8253  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8254  return __ret;
8255 }
8256 #endif
8257 
8258 #ifdef __LITTLE_ENDIAN__
8259 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8260  uint8x16_t __ret;
8261  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
8262  return __ret;
8263 }
8264 #else
8265 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
8266  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8267  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8268  uint8x16_t __ret;
8269  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
8270  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8271  return __ret;
8272 }
8273 #endif
8274 
8275 #ifdef __LITTLE_ENDIAN__
8276 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8277  uint32x4_t __ret;
8278  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
8279  return __ret;
8280 }
8281 #else
8282 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
8283  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8284  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8285  uint32x4_t __ret;
8286  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
8287  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8288  return __ret;
8289 }
8290 #endif
8291 
8292 #ifdef __LITTLE_ENDIAN__
8293 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8294  uint16x8_t __ret;
8295  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
8296  return __ret;
8297 }
8298 #else
8299 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
8300  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8301  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8302  uint16x8_t __ret;
8303  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
8304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8305  return __ret;
8306 }
8307 #endif
8308 
8309 #ifdef __LITTLE_ENDIAN__
8310 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
8311  int8x16_t __ret;
8312  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
8313  return __ret;
8314 }
8315 #else
8316 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
8317  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8318  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8319  int8x16_t __ret;
8320  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
8321  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
8322  return __ret;
8323 }
8324 #endif
8325 
8326 #ifdef __LITTLE_ENDIAN__
8327 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
8328  int32x4_t __ret;
8329  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
8330  return __ret;
8331 }
8332 #else
8333 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
8334  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8335  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8336  int32x4_t __ret;
8337  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
8338  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8339  return __ret;
8340 }
8341 #endif
8342 
8343 #ifdef __LITTLE_ENDIAN__
8344 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
8345  int16x8_t __ret;
8346  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
8347  return __ret;
8348 }
8349 #else
8350 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
8351  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8352  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8353  int16x8_t __ret;
8354  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
8355  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8356  return __ret;
8357 }
8358 #endif
8359 
8360 #ifdef __LITTLE_ENDIAN__
8361 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
8362  uint8x8_t __ret;
8363  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
8364  return __ret;
8365 }
8366 #else
8367 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
8368  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8369  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8370  uint8x8_t __ret;
8371  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
8372  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8373  return __ret;
8374 }
8375 #endif
8376 
8377 #ifdef __LITTLE_ENDIAN__
8378 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
8379  uint32x2_t __ret;
8380  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
8381  return __ret;
8382 }
8383 #else
8384 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
8385  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8386  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8387  uint32x2_t __ret;
8388  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
8389  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8390  return __ret;
8391 }
8392 #endif
8393 
8394 #ifdef __LITTLE_ENDIAN__
8395 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
8396  uint16x4_t __ret;
8397  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
8398  return __ret;
8399 }
8400 #else
8401 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
8402  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8403  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8404  uint16x4_t __ret;
8405  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
8406  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8407  return __ret;
8408 }
8409 #endif
8410 
8411 #ifdef __LITTLE_ENDIAN__
8412 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
8413  int8x8_t __ret;
8414  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
8415  return __ret;
8416 }
8417 #else
8418 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
8419  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
8420  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
8421  int8x8_t __ret;
8422  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
8423  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
8424  return __ret;
8425 }
8426 #endif
8427 
8428 #ifdef __LITTLE_ENDIAN__
8429 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
8430  int32x2_t __ret;
8431  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
8432  return __ret;
8433 }
8434 #else
8435 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
8436  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
8437  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
8438  int32x2_t __ret;
8439  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
8440  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
8441  return __ret;
8442 }
8443 #endif
8444 
8445 #ifdef __LITTLE_ENDIAN__
8446 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
8447  int16x4_t __ret;
8448  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
8449  return __ret;
8450 }
8451 #else
8452 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
8453  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
8454  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
8455  int16x4_t __ret;
8456  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
8457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
8458  return __ret;
8459 }
8460 #endif
8461 
8462 #ifdef __LITTLE_ENDIAN__
8463 #define vld1_p8(__p0) __extension__ ({ \
8464  poly8x8_t __ret; \
8465  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
8466  __ret; \
8467 })
8468 #else
8469 #define vld1_p8(__p0) __extension__ ({ \
8470  poly8x8_t __ret; \
8471  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
8472  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8473  __ret; \
8474 })
8475 #endif
8476 
8477 #ifdef __LITTLE_ENDIAN__
8478 #define vld1_p16(__p0) __extension__ ({ \
8479  poly16x4_t __ret; \
8480  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
8481  __ret; \
8482 })
8483 #else
8484 #define vld1_p16(__p0) __extension__ ({ \
8485  poly16x4_t __ret; \
8486  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
8487  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8488  __ret; \
8489 })
8490 #endif
8491 
8492 #ifdef __LITTLE_ENDIAN__
8493 #define vld1q_p8(__p0) __extension__ ({ \
8494  poly8x16_t __ret; \
8495  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
8496  __ret; \
8497 })
8498 #else
8499 #define vld1q_p8(__p0) __extension__ ({ \
8500  poly8x16_t __ret; \
8501  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
8502  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8503  __ret; \
8504 })
8505 #endif
8506 
8507 #ifdef __LITTLE_ENDIAN__
8508 #define vld1q_p16(__p0) __extension__ ({ \
8509  poly16x8_t __ret; \
8510  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
8511  __ret; \
8512 })
8513 #else
8514 #define vld1q_p16(__p0) __extension__ ({ \
8515  poly16x8_t __ret; \
8516  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
8517  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8518  __ret; \
8519 })
8520 #endif
8521 
8522 #ifdef __LITTLE_ENDIAN__
8523 #define vld1q_u8(__p0) __extension__ ({ \
8524  uint8x16_t __ret; \
8525  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
8526  __ret; \
8527 })
8528 #else
8529 #define vld1q_u8(__p0) __extension__ ({ \
8530  uint8x16_t __ret; \
8531  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
8532  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8533  __ret; \
8534 })
8535 #endif
8536 
8537 #ifdef __LITTLE_ENDIAN__
8538 #define vld1q_u32(__p0) __extension__ ({ \
8539  uint32x4_t __ret; \
8540  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
8541  __ret; \
8542 })
8543 #else
8544 #define vld1q_u32(__p0) __extension__ ({ \
8545  uint32x4_t __ret; \
8546  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
8547  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8548  __ret; \
8549 })
8550 #endif
8551 
8552 #ifdef __LITTLE_ENDIAN__
8553 #define vld1q_u64(__p0) __extension__ ({ \
8554  uint64x2_t __ret; \
8555  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
8556  __ret; \
8557 })
8558 #else
8559 #define vld1q_u64(__p0) __extension__ ({ \
8560  uint64x2_t __ret; \
8561  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
8562  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8563  __ret; \
8564 })
8565 #endif
8566 
8567 #ifdef __LITTLE_ENDIAN__
8568 #define vld1q_u16(__p0) __extension__ ({ \
8569  uint16x8_t __ret; \
8570  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
8571  __ret; \
8572 })
8573 #else
8574 #define vld1q_u16(__p0) __extension__ ({ \
8575  uint16x8_t __ret; \
8576  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
8577  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8578  __ret; \
8579 })
8580 #endif
8581 
8582 #ifdef __LITTLE_ENDIAN__
8583 #define vld1q_s8(__p0) __extension__ ({ \
8584  int8x16_t __ret; \
8585  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
8586  __ret; \
8587 })
8588 #else
8589 #define vld1q_s8(__p0) __extension__ ({ \
8590  int8x16_t __ret; \
8591  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
8592  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8593  __ret; \
8594 })
8595 #endif
8596 
8597 #ifdef __LITTLE_ENDIAN__
8598 #define vld1q_f32(__p0) __extension__ ({ \
8599  float32x4_t __ret; \
8600  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
8601  __ret; \
8602 })
8603 #else
8604 #define vld1q_f32(__p0) __extension__ ({ \
8605  float32x4_t __ret; \
8606  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
8607  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8608  __ret; \
8609 })
8610 #endif
8611 
8612 #ifdef __LITTLE_ENDIAN__
8613 #define vld1q_s32(__p0) __extension__ ({ \
8614  int32x4_t __ret; \
8615  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
8616  __ret; \
8617 })
8618 #else
8619 #define vld1q_s32(__p0) __extension__ ({ \
8620  int32x4_t __ret; \
8621  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
8622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8623  __ret; \
8624 })
8625 #endif
8626 
8627 #ifdef __LITTLE_ENDIAN__
8628 #define vld1q_s64(__p0) __extension__ ({ \
8629  int64x2_t __ret; \
8630  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
8631  __ret; \
8632 })
8633 #else
8634 #define vld1q_s64(__p0) __extension__ ({ \
8635  int64x2_t __ret; \
8636  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
8637  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8638  __ret; \
8639 })
8640 #endif
8641 
8642 #ifdef __LITTLE_ENDIAN__
8643 #define vld1q_s16(__p0) __extension__ ({ \
8644  int16x8_t __ret; \
8645  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
8646  __ret; \
8647 })
8648 #else
8649 #define vld1q_s16(__p0) __extension__ ({ \
8650  int16x8_t __ret; \
8651  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
8652  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8653  __ret; \
8654 })
8655 #endif
8656 
8657 #ifdef __LITTLE_ENDIAN__
8658 #define vld1_u8(__p0) __extension__ ({ \
8659  uint8x8_t __ret; \
8660  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
8661  __ret; \
8662 })
8663 #else
8664 #define vld1_u8(__p0) __extension__ ({ \
8665  uint8x8_t __ret; \
8666  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
8667  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8668  __ret; \
8669 })
8670 #endif
8671 
8672 #ifdef __LITTLE_ENDIAN__
8673 #define vld1_u32(__p0) __extension__ ({ \
8674  uint32x2_t __ret; \
8675  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
8676  __ret; \
8677 })
8678 #else
8679 #define vld1_u32(__p0) __extension__ ({ \
8680  uint32x2_t __ret; \
8681  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
8682  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8683  __ret; \
8684 })
8685 #endif
8686 
8687 #define vld1_u64(__p0) __extension__ ({ \
8688  uint64x1_t __ret; \
8689  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
8690  __ret; \
8691 })
8692 #ifdef __LITTLE_ENDIAN__
8693 #define vld1_u16(__p0) __extension__ ({ \
8694  uint16x4_t __ret; \
8695  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
8696  __ret; \
8697 })
8698 #else
8699 #define vld1_u16(__p0) __extension__ ({ \
8700  uint16x4_t __ret; \
8701  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
8702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8703  __ret; \
8704 })
8705 #endif
8706 
8707 #ifdef __LITTLE_ENDIAN__
8708 #define vld1_s8(__p0) __extension__ ({ \
8709  int8x8_t __ret; \
8710  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
8711  __ret; \
8712 })
8713 #else
8714 #define vld1_s8(__p0) __extension__ ({ \
8715  int8x8_t __ret; \
8716  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
8717  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8718  __ret; \
8719 })
8720 #endif
8721 
8722 #ifdef __LITTLE_ENDIAN__
8723 #define vld1_f32(__p0) __extension__ ({ \
8724  float32x2_t __ret; \
8725  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
8726  __ret; \
8727 })
8728 #else
8729 #define vld1_f32(__p0) __extension__ ({ \
8730  float32x2_t __ret; \
8731  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
8732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8733  __ret; \
8734 })
8735 #endif
8736 
8737 #ifdef __LITTLE_ENDIAN__
8738 #define vld1_s32(__p0) __extension__ ({ \
8739  int32x2_t __ret; \
8740  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
8741  __ret; \
8742 })
8743 #else
8744 #define vld1_s32(__p0) __extension__ ({ \
8745  int32x2_t __ret; \
8746  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
8747  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8748  __ret; \
8749 })
8750 #endif
8751 
8752 #define vld1_s64(__p0) __extension__ ({ \
8753  int64x1_t __ret; \
8754  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
8755  __ret; \
8756 })
8757 #ifdef __LITTLE_ENDIAN__
8758 #define vld1_s16(__p0) __extension__ ({ \
8759  int16x4_t __ret; \
8760  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
8761  __ret; \
8762 })
8763 #else
8764 #define vld1_s16(__p0) __extension__ ({ \
8765  int16x4_t __ret; \
8766  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
8767  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8768  __ret; \
8769 })
8770 #endif
8771 
8772 #ifdef __LITTLE_ENDIAN__
8773 #define vld1_dup_p8(__p0) __extension__ ({ \
8774  poly8x8_t __ret; \
8775  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
8776  __ret; \
8777 })
8778 #else
8779 #define vld1_dup_p8(__p0) __extension__ ({ \
8780  poly8x8_t __ret; \
8781  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
8782  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8783  __ret; \
8784 })
8785 #endif
8786 
8787 #ifdef __LITTLE_ENDIAN__
8788 #define vld1_dup_p16(__p0) __extension__ ({ \
8789  poly16x4_t __ret; \
8790  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
8791  __ret; \
8792 })
8793 #else
8794 #define vld1_dup_p16(__p0) __extension__ ({ \
8795  poly16x4_t __ret; \
8796  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
8797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8798  __ret; \
8799 })
8800 #endif
8801 
8802 #ifdef __LITTLE_ENDIAN__
8803 #define vld1q_dup_p8(__p0) __extension__ ({ \
8804  poly8x16_t __ret; \
8805  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
8806  __ret; \
8807 })
8808 #else
8809 #define vld1q_dup_p8(__p0) __extension__ ({ \
8810  poly8x16_t __ret; \
8811  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
8812  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8813  __ret; \
8814 })
8815 #endif
8816 
8817 #ifdef __LITTLE_ENDIAN__
8818 #define vld1q_dup_p16(__p0) __extension__ ({ \
8819  poly16x8_t __ret; \
8820  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
8821  __ret; \
8822 })
8823 #else
8824 #define vld1q_dup_p16(__p0) __extension__ ({ \
8825  poly16x8_t __ret; \
8826  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
8827  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8828  __ret; \
8829 })
8830 #endif
8831 
8832 #ifdef __LITTLE_ENDIAN__
8833 #define vld1q_dup_u8(__p0) __extension__ ({ \
8834  uint8x16_t __ret; \
8835  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
8836  __ret; \
8837 })
8838 #else
8839 #define vld1q_dup_u8(__p0) __extension__ ({ \
8840  uint8x16_t __ret; \
8841  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
8842  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8843  __ret; \
8844 })
8845 #endif
8846 
8847 #ifdef __LITTLE_ENDIAN__
8848 #define vld1q_dup_u32(__p0) __extension__ ({ \
8849  uint32x4_t __ret; \
8850  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
8851  __ret; \
8852 })
8853 #else
8854 #define vld1q_dup_u32(__p0) __extension__ ({ \
8855  uint32x4_t __ret; \
8856  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
8857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8858  __ret; \
8859 })
8860 #endif
8861 
8862 #ifdef __LITTLE_ENDIAN__
8863 #define vld1q_dup_u64(__p0) __extension__ ({ \
8864  uint64x2_t __ret; \
8865  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
8866  __ret; \
8867 })
8868 #else
8869 #define vld1q_dup_u64(__p0) __extension__ ({ \
8870  uint64x2_t __ret; \
8871  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
8872  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8873  __ret; \
8874 })
8875 #endif
8876 
8877 #ifdef __LITTLE_ENDIAN__
8878 #define vld1q_dup_u16(__p0) __extension__ ({ \
8879  uint16x8_t __ret; \
8880  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8881  __ret; \
8882 })
8883 #else
8884 #define vld1q_dup_u16(__p0) __extension__ ({ \
8885  uint16x8_t __ret; \
8886  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8887  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8888  __ret; \
8889 })
8890 #endif
8891 
8892 #ifdef __LITTLE_ENDIAN__
8893 #define vld1q_dup_s8(__p0) __extension__ ({ \
8894  int8x16_t __ret; \
8895  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8896  __ret; \
8897 })
8898 #else
8899 #define vld1q_dup_s8(__p0) __extension__ ({ \
8900  int8x16_t __ret; \
8901  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8902  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8903  __ret; \
8904 })
8905 #endif
8906 
8907 #ifdef __LITTLE_ENDIAN__
8908 #define vld1q_dup_f32(__p0) __extension__ ({ \
8909  float32x4_t __ret; \
8910  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8911  __ret; \
8912 })
8913 #else
8914 #define vld1q_dup_f32(__p0) __extension__ ({ \
8915  float32x4_t __ret; \
8916  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8917  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8918  __ret; \
8919 })
8920 #endif
8921 
8922 #ifdef __LITTLE_ENDIAN__
8923 #define vld1q_dup_s32(__p0) __extension__ ({ \
8924  int32x4_t __ret; \
8925  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8926  __ret; \
8927 })
8928 #else
8929 #define vld1q_dup_s32(__p0) __extension__ ({ \
8930  int32x4_t __ret; \
8931  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8932  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8933  __ret; \
8934 })
8935 #endif
8936 
8937 #ifdef __LITTLE_ENDIAN__
8938 #define vld1q_dup_s64(__p0) __extension__ ({ \
8939  int64x2_t __ret; \
8940  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8941  __ret; \
8942 })
8943 #else
8944 #define vld1q_dup_s64(__p0) __extension__ ({ \
8945  int64x2_t __ret; \
8946  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8947  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8948  __ret; \
8949 })
8950 #endif
8951 
8952 #ifdef __LITTLE_ENDIAN__
8953 #define vld1q_dup_s16(__p0) __extension__ ({ \
8954  int16x8_t __ret; \
8955  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8956  __ret; \
8957 })
8958 #else
8959 #define vld1q_dup_s16(__p0) __extension__ ({ \
8960  int16x8_t __ret; \
8961  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8963  __ret; \
8964 })
8965 #endif
8966 
8967 #ifdef __LITTLE_ENDIAN__
8968 #define vld1_dup_u8(__p0) __extension__ ({ \
8969  uint8x8_t __ret; \
8970  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8971  __ret; \
8972 })
8973 #else
8974 #define vld1_dup_u8(__p0) __extension__ ({ \
8975  uint8x8_t __ret; \
8976  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8977  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8978  __ret; \
8979 })
8980 #endif
8981 
8982 #ifdef __LITTLE_ENDIAN__
8983 #define vld1_dup_u32(__p0) __extension__ ({ \
8984  uint32x2_t __ret; \
8985  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8986  __ret; \
8987 })
8988 #else
8989 #define vld1_dup_u32(__p0) __extension__ ({ \
8990  uint32x2_t __ret; \
8991  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8992  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8993  __ret; \
8994 })
8995 #endif
8996 
8997 #define vld1_dup_u64(__p0) __extension__ ({ \
8998  uint64x1_t __ret; \
8999  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
9000  __ret; \
9001 })
9002 #ifdef __LITTLE_ENDIAN__
9003 #define vld1_dup_u16(__p0) __extension__ ({ \
9004  uint16x4_t __ret; \
9005  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
9006  __ret; \
9007 })
9008 #else
9009 #define vld1_dup_u16(__p0) __extension__ ({ \
9010  uint16x4_t __ret; \
9011  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
9012  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9013  __ret; \
9014 })
9015 #endif
9016 
9017 #ifdef __LITTLE_ENDIAN__
9018 #define vld1_dup_s8(__p0) __extension__ ({ \
9019  int8x8_t __ret; \
9020  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
9021  __ret; \
9022 })
9023 #else
9024 #define vld1_dup_s8(__p0) __extension__ ({ \
9025  int8x8_t __ret; \
9026  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
9027  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9028  __ret; \
9029 })
9030 #endif
9031 
9032 #ifdef __LITTLE_ENDIAN__
9033 #define vld1_dup_f32(__p0) __extension__ ({ \
9034  float32x2_t __ret; \
9035  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
9036  __ret; \
9037 })
9038 #else
9039 #define vld1_dup_f32(__p0) __extension__ ({ \
9040  float32x2_t __ret; \
9041  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
9042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9043  __ret; \
9044 })
9045 #endif
9046 
9047 #ifdef __LITTLE_ENDIAN__
9048 #define vld1_dup_s32(__p0) __extension__ ({ \
9049  int32x2_t __ret; \
9050  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
9051  __ret; \
9052 })
9053 #else
9054 #define vld1_dup_s32(__p0) __extension__ ({ \
9055  int32x2_t __ret; \
9056  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
9057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9058  __ret; \
9059 })
9060 #endif
9061 
9062 #define vld1_dup_s64(__p0) __extension__ ({ \
9063  int64x1_t __ret; \
9064  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
9065  __ret; \
9066 })
9067 #ifdef __LITTLE_ENDIAN__
9068 #define vld1_dup_s16(__p0) __extension__ ({ \
9069  int16x4_t __ret; \
9070  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
9071  __ret; \
9072 })
9073 #else
9074 #define vld1_dup_s16(__p0) __extension__ ({ \
9075  int16x4_t __ret; \
9076  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
9077  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9078  __ret; \
9079 })
9080 #endif
9081 
9082 #ifdef __LITTLE_ENDIAN__
9083 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9084  poly8x8_t __s1 = __p1; \
9085  poly8x8_t __ret; \
9086  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
9087  __ret; \
9088 })
9089 #else
9090 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9091  poly8x8_t __s1 = __p1; \
9092  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9093  poly8x8_t __ret; \
9094  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
9095  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9096  __ret; \
9097 })
9098 #endif
9099 
9100 #ifdef __LITTLE_ENDIAN__
9101 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9102  poly16x4_t __s1 = __p1; \
9103  poly16x4_t __ret; \
9104  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
9105  __ret; \
9106 })
9107 #else
9108 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9109  poly16x4_t __s1 = __p1; \
9110  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9111  poly16x4_t __ret; \
9112  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
9113  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9114  __ret; \
9115 })
9116 #endif
9117 
9118 #ifdef __LITTLE_ENDIAN__
9119 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9120  poly8x16_t __s1 = __p1; \
9121  poly8x16_t __ret; \
9122  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
9123  __ret; \
9124 })
9125 #else
9126 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9127  poly8x16_t __s1 = __p1; \
9128  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9129  poly8x16_t __ret; \
9130  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
9131  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9132  __ret; \
9133 })
9134 #endif
9135 
9136 #ifdef __LITTLE_ENDIAN__
9137 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9138  poly16x8_t __s1 = __p1; \
9139  poly16x8_t __ret; \
9140  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
9141  __ret; \
9142 })
9143 #else
9144 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9145  poly16x8_t __s1 = __p1; \
9146  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9147  poly16x8_t __ret; \
9148  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
9149  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9150  __ret; \
9151 })
9152 #endif
9153 
9154 #ifdef __LITTLE_ENDIAN__
9155 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9156  uint8x16_t __s1 = __p1; \
9157  uint8x16_t __ret; \
9158  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
9159  __ret; \
9160 })
9161 #else
9162 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9163  uint8x16_t __s1 = __p1; \
9164  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9165  uint8x16_t __ret; \
9166  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
9167  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9168  __ret; \
9169 })
9170 #endif
9171 
9172 #ifdef __LITTLE_ENDIAN__
9173 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9174  uint32x4_t __s1 = __p1; \
9175  uint32x4_t __ret; \
9176  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
9177  __ret; \
9178 })
9179 #else
9180 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9181  uint32x4_t __s1 = __p1; \
9182  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9183  uint32x4_t __ret; \
9184  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
9185  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9186  __ret; \
9187 })
9188 #endif
9189 
9190 #ifdef __LITTLE_ENDIAN__
9191 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9192  uint64x2_t __s1 = __p1; \
9193  uint64x2_t __ret; \
9194  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
9195  __ret; \
9196 })
9197 #else
9198 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9199  uint64x2_t __s1 = __p1; \
9200  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9201  uint64x2_t __ret; \
9202  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
9203  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9204  __ret; \
9205 })
9206 #endif
9207 
9208 #ifdef __LITTLE_ENDIAN__
9209 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9210  uint16x8_t __s1 = __p1; \
9211  uint16x8_t __ret; \
9212  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
9213  __ret; \
9214 })
9215 #else
9216 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9217  uint16x8_t __s1 = __p1; \
9218  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9219  uint16x8_t __ret; \
9220  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
9221  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9222  __ret; \
9223 })
9224 #endif
9225 
9226 #ifdef __LITTLE_ENDIAN__
9227 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9228  int8x16_t __s1 = __p1; \
9229  int8x16_t __ret; \
9230  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
9231  __ret; \
9232 })
9233 #else
9234 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9235  int8x16_t __s1 = __p1; \
9236  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9237  int8x16_t __ret; \
9238  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
9239  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9240  __ret; \
9241 })
9242 #endif
9243 
9244 #ifdef __LITTLE_ENDIAN__
9245 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9246  float32x4_t __s1 = __p1; \
9247  float32x4_t __ret; \
9248  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
9249  __ret; \
9250 })
9251 #else
9252 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9253  float32x4_t __s1 = __p1; \
9254  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9255  float32x4_t __ret; \
9256  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
9257  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9258  __ret; \
9259 })
9260 #endif
9261 
9262 #ifdef __LITTLE_ENDIAN__
9263 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9264  int32x4_t __s1 = __p1; \
9265  int32x4_t __ret; \
9266  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
9267  __ret; \
9268 })
9269 #else
9270 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9271  int32x4_t __s1 = __p1; \
9272  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9273  int32x4_t __ret; \
9274  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
9275  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9276  __ret; \
9277 })
9278 #endif
9279 
9280 #ifdef __LITTLE_ENDIAN__
9281 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9282  int64x2_t __s1 = __p1; \
9283  int64x2_t __ret; \
9284  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
9285  __ret; \
9286 })
9287 #else
9288 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9289  int64x2_t __s1 = __p1; \
9290  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9291  int64x2_t __ret; \
9292  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
9293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9294  __ret; \
9295 })
9296 #endif
9297 
9298 #ifdef __LITTLE_ENDIAN__
9299 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9300  int16x8_t __s1 = __p1; \
9301  int16x8_t __ret; \
9302  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
9303  __ret; \
9304 })
9305 #else
9306 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9307  int16x8_t __s1 = __p1; \
9308  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9309  int16x8_t __ret; \
9310  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
9311  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9312  __ret; \
9313 })
9314 #endif
9315 
9316 #ifdef __LITTLE_ENDIAN__
9317 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9318  uint8x8_t __s1 = __p1; \
9319  uint8x8_t __ret; \
9320  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
9321  __ret; \
9322 })
9323 #else
9324 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9325  uint8x8_t __s1 = __p1; \
9326  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9327  uint8x8_t __ret; \
9328  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
9329  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9330  __ret; \
9331 })
9332 #endif
9333 
9334 #ifdef __LITTLE_ENDIAN__
9335 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9336  uint32x2_t __s1 = __p1; \
9337  uint32x2_t __ret; \
9338  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
9339  __ret; \
9340 })
9341 #else
9342 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9343  uint32x2_t __s1 = __p1; \
9344  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9345  uint32x2_t __ret; \
9346  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
9347  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9348  __ret; \
9349 })
9350 #endif
9351 
9352 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
9353  uint64x1_t __s1 = __p1; \
9354  uint64x1_t __ret; \
9355  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
9356  __ret; \
9357 })
9358 #ifdef __LITTLE_ENDIAN__
9359 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9360  uint16x4_t __s1 = __p1; \
9361  uint16x4_t __ret; \
9362  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
9363  __ret; \
9364 })
9365 #else
9366 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9367  uint16x4_t __s1 = __p1; \
9368  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9369  uint16x4_t __ret; \
9370  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
9371  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9372  __ret; \
9373 })
9374 #endif
9375 
9376 #ifdef __LITTLE_ENDIAN__
9377 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9378  int8x8_t __s1 = __p1; \
9379  int8x8_t __ret; \
9380  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
9381  __ret; \
9382 })
9383 #else
9384 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9385  int8x8_t __s1 = __p1; \
9386  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
9387  int8x8_t __ret; \
9388  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
9389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
9390  __ret; \
9391 })
9392 #endif
9393 
9394 #ifdef __LITTLE_ENDIAN__
9395 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9396  float32x2_t __s1 = __p1; \
9397  float32x2_t __ret; \
9398  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
9399  __ret; \
9400 })
9401 #else
9402 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9403  float32x2_t __s1 = __p1; \
9404  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9405  float32x2_t __ret; \
9406  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
9407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9408  __ret; \
9409 })
9410 #endif
9411 
9412 #ifdef __LITTLE_ENDIAN__
9413 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9414  int32x2_t __s1 = __p1; \
9415  int32x2_t __ret; \
9416  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
9417  __ret; \
9418 })
9419 #else
9420 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9421  int32x2_t __s1 = __p1; \
9422  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
9423  int32x2_t __ret; \
9424  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
9425  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
9426  __ret; \
9427 })
9428 #endif
9429 
9430 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
9431  int64x1_t __s1 = __p1; \
9432  int64x1_t __ret; \
9433  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
9434  __ret; \
9435 })
9436 #ifdef __LITTLE_ENDIAN__
9437 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9438  int16x4_t __s1 = __p1; \
9439  int16x4_t __ret; \
9440  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
9441  __ret; \
9442 })
9443 #else
9444 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9445  int16x4_t __s1 = __p1; \
9446  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
9447  int16x4_t __ret; \
9448  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
9449  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
9450  __ret; \
9451 })
9452 #endif
9453 
9454 #ifdef __LITTLE_ENDIAN__
9455 #define vld1_p8_x2(__p0) __extension__ ({ \
9456  poly8x8x2_t __ret; \
9457  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
9458  __ret; \
9459 })
9460 #else
9461 #define vld1_p8_x2(__p0) __extension__ ({ \
9462  poly8x8x2_t __ret; \
9463  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
9464  \
9465  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9466  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9467  __ret; \
9468 })
9469 #endif
9470 
9471 #ifdef __LITTLE_ENDIAN__
9472 #define vld1_p16_x2(__p0) __extension__ ({ \
9473  poly16x4x2_t __ret; \
9474  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
9475  __ret; \
9476 })
9477 #else
9478 #define vld1_p16_x2(__p0) __extension__ ({ \
9479  poly16x4x2_t __ret; \
9480  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
9481  \
9482  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9483  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9484  __ret; \
9485 })
9486 #endif
9487 
9488 #ifdef __LITTLE_ENDIAN__
9489 #define vld1q_p8_x2(__p0) __extension__ ({ \
9490  poly8x16x2_t __ret; \
9491  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
9492  __ret; \
9493 })
9494 #else
9495 #define vld1q_p8_x2(__p0) __extension__ ({ \
9496  poly8x16x2_t __ret; \
9497  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
9498  \
9499  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9500  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9501  __ret; \
9502 })
9503 #endif
9504 
9505 #ifdef __LITTLE_ENDIAN__
9506 #define vld1q_p16_x2(__p0) __extension__ ({ \
9507  poly16x8x2_t __ret; \
9508  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
9509  __ret; \
9510 })
9511 #else
9512 #define vld1q_p16_x2(__p0) __extension__ ({ \
9513  poly16x8x2_t __ret; \
9514  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
9515  \
9516  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9517  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9518  __ret; \
9519 })
9520 #endif
9521 
9522 #ifdef __LITTLE_ENDIAN__
9523 #define vld1q_u8_x2(__p0) __extension__ ({ \
9524  uint8x16x2_t __ret; \
9525  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
9526  __ret; \
9527 })
9528 #else
9529 #define vld1q_u8_x2(__p0) __extension__ ({ \
9530  uint8x16x2_t __ret; \
9531  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
9532  \
9533  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9534  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9535  __ret; \
9536 })
9537 #endif
9538 
9539 #ifdef __LITTLE_ENDIAN__
9540 #define vld1q_u32_x2(__p0) __extension__ ({ \
9541  uint32x4x2_t __ret; \
9542  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
9543  __ret; \
9544 })
9545 #else
9546 #define vld1q_u32_x2(__p0) __extension__ ({ \
9547  uint32x4x2_t __ret; \
9548  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
9549  \
9550  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9551  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9552  __ret; \
9553 })
9554 #endif
9555 
9556 #ifdef __LITTLE_ENDIAN__
9557 #define vld1q_u64_x2(__p0) __extension__ ({ \
9558  uint64x2x2_t __ret; \
9559  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
9560  __ret; \
9561 })
9562 #else
9563 #define vld1q_u64_x2(__p0) __extension__ ({ \
9564  uint64x2x2_t __ret; \
9565  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
9566  \
9567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9569  __ret; \
9570 })
9571 #endif
9572 
9573 #ifdef __LITTLE_ENDIAN__
9574 #define vld1q_u16_x2(__p0) __extension__ ({ \
9575  uint16x8x2_t __ret; \
9576  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
9577  __ret; \
9578 })
9579 #else
9580 #define vld1q_u16_x2(__p0) __extension__ ({ \
9581  uint16x8x2_t __ret; \
9582  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
9583  \
9584  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9585  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9586  __ret; \
9587 })
9588 #endif
9589 
9590 #ifdef __LITTLE_ENDIAN__
9591 #define vld1q_s8_x2(__p0) __extension__ ({ \
9592  int8x16x2_t __ret; \
9593  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
9594  __ret; \
9595 })
9596 #else
9597 #define vld1q_s8_x2(__p0) __extension__ ({ \
9598  int8x16x2_t __ret; \
9599  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
9600  \
9601  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9602  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9603  __ret; \
9604 })
9605 #endif
9606 
9607 #ifdef __LITTLE_ENDIAN__
9608 #define vld1q_f32_x2(__p0) __extension__ ({ \
9609  float32x4x2_t __ret; \
9610  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
9611  __ret; \
9612 })
9613 #else
9614 #define vld1q_f32_x2(__p0) __extension__ ({ \
9615  float32x4x2_t __ret; \
9616  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
9617  \
9618  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9619  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9620  __ret; \
9621 })
9622 #endif
9623 
9624 #ifdef __LITTLE_ENDIAN__
9625 #define vld1q_s32_x2(__p0) __extension__ ({ \
9626  int32x4x2_t __ret; \
9627  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
9628  __ret; \
9629 })
9630 #else
9631 #define vld1q_s32_x2(__p0) __extension__ ({ \
9632  int32x4x2_t __ret; \
9633  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
9634  \
9635  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9636  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9637  __ret; \
9638 })
9639 #endif
9640 
9641 #ifdef __LITTLE_ENDIAN__
9642 #define vld1q_s64_x2(__p0) __extension__ ({ \
9643  int64x2x2_t __ret; \
9644  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
9645  __ret; \
9646 })
9647 #else
9648 #define vld1q_s64_x2(__p0) __extension__ ({ \
9649  int64x2x2_t __ret; \
9650  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
9651  \
9652  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9653  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9654  __ret; \
9655 })
9656 #endif
9657 
9658 #ifdef __LITTLE_ENDIAN__
9659 #define vld1q_s16_x2(__p0) __extension__ ({ \
9660  int16x8x2_t __ret; \
9661  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
9662  __ret; \
9663 })
9664 #else
9665 #define vld1q_s16_x2(__p0) __extension__ ({ \
9666  int16x8x2_t __ret; \
9667  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
9668  \
9669  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9670  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9671  __ret; \
9672 })
9673 #endif
9674 
9675 #ifdef __LITTLE_ENDIAN__
9676 #define vld1_u8_x2(__p0) __extension__ ({ \
9677  uint8x8x2_t __ret; \
9678  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
9679  __ret; \
9680 })
9681 #else
9682 #define vld1_u8_x2(__p0) __extension__ ({ \
9683  uint8x8x2_t __ret; \
9684  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
9685  \
9686  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9687  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9688  __ret; \
9689 })
9690 #endif
9691 
9692 #ifdef __LITTLE_ENDIAN__
9693 #define vld1_u32_x2(__p0) __extension__ ({ \
9694  uint32x2x2_t __ret; \
9695  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
9696  __ret; \
9697 })
9698 #else
9699 #define vld1_u32_x2(__p0) __extension__ ({ \
9700  uint32x2x2_t __ret; \
9701  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
9702  \
9703  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9704  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9705  __ret; \
9706 })
9707 #endif
9708 
9709 #define vld1_u64_x2(__p0) __extension__ ({ \
9710  uint64x1x2_t __ret; \
9711  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
9712  __ret; \
9713 })
9714 #ifdef __LITTLE_ENDIAN__
9715 #define vld1_u16_x2(__p0) __extension__ ({ \
9716  uint16x4x2_t __ret; \
9717  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
9718  __ret; \
9719 })
9720 #else
9721 #define vld1_u16_x2(__p0) __extension__ ({ \
9722  uint16x4x2_t __ret; \
9723  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
9724  \
9725  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9726  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9727  __ret; \
9728 })
9729 #endif
9730 
9731 #ifdef __LITTLE_ENDIAN__
9732 #define vld1_s8_x2(__p0) __extension__ ({ \
9733  int8x8x2_t __ret; \
9734  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
9735  __ret; \
9736 })
9737 #else
9738 #define vld1_s8_x2(__p0) __extension__ ({ \
9739  int8x8x2_t __ret; \
9740  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
9741  \
9742  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9743  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9744  __ret; \
9745 })
9746 #endif
9747 
9748 #ifdef __LITTLE_ENDIAN__
9749 #define vld1_f32_x2(__p0) __extension__ ({ \
9750  float32x2x2_t __ret; \
9751  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
9752  __ret; \
9753 })
9754 #else
9755 #define vld1_f32_x2(__p0) __extension__ ({ \
9756  float32x2x2_t __ret; \
9757  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
9758  \
9759  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9760  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9761  __ret; \
9762 })
9763 #endif
9764 
9765 #ifdef __LITTLE_ENDIAN__
9766 #define vld1_s32_x2(__p0) __extension__ ({ \
9767  int32x2x2_t __ret; \
9768  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
9769  __ret; \
9770 })
9771 #else
9772 #define vld1_s32_x2(__p0) __extension__ ({ \
9773  int32x2x2_t __ret; \
9774  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
9775  \
9776  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9777  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9778  __ret; \
9779 })
9780 #endif
9781 
9782 #define vld1_s64_x2(__p0) __extension__ ({ \
9783  int64x1x2_t __ret; \
9784  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
9785  __ret; \
9786 })
9787 #ifdef __LITTLE_ENDIAN__
9788 #define vld1_s16_x2(__p0) __extension__ ({ \
9789  int16x4x2_t __ret; \
9790  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
9791  __ret; \
9792 })
9793 #else
9794 #define vld1_s16_x2(__p0) __extension__ ({ \
9795  int16x4x2_t __ret; \
9796  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
9797  \
9798  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9799  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9800  __ret; \
9801 })
9802 #endif
9803 
9804 #ifdef __LITTLE_ENDIAN__
9805 #define vld1_p8_x3(__p0) __extension__ ({ \
9806  poly8x8x3_t __ret; \
9807  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
9808  __ret; \
9809 })
9810 #else
9811 #define vld1_p8_x3(__p0) __extension__ ({ \
9812  poly8x8x3_t __ret; \
9813  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
9814  \
9815  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9816  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9817  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9818  __ret; \
9819 })
9820 #endif
9821 
9822 #ifdef __LITTLE_ENDIAN__
9823 #define vld1_p16_x3(__p0) __extension__ ({ \
9824  poly16x4x3_t __ret; \
9825  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
9826  __ret; \
9827 })
9828 #else
9829 #define vld1_p16_x3(__p0) __extension__ ({ \
9830  poly16x4x3_t __ret; \
9831  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
9832  \
9833  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9834  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9835  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9836  __ret; \
9837 })
9838 #endif
9839 
9840 #ifdef __LITTLE_ENDIAN__
9841 #define vld1q_p8_x3(__p0) __extension__ ({ \
9842  poly8x16x3_t __ret; \
9843  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
9844  __ret; \
9845 })
9846 #else
9847 #define vld1q_p8_x3(__p0) __extension__ ({ \
9848  poly8x16x3_t __ret; \
9849  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
9850  \
9851  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9852  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9853  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9854  __ret; \
9855 })
9856 #endif
9857 
9858 #ifdef __LITTLE_ENDIAN__
9859 #define vld1q_p16_x3(__p0) __extension__ ({ \
9860  poly16x8x3_t __ret; \
9861  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
9862  __ret; \
9863 })
9864 #else
9865 #define vld1q_p16_x3(__p0) __extension__ ({ \
9866  poly16x8x3_t __ret; \
9867  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
9868  \
9869  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9870  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9871  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9872  __ret; \
9873 })
9874 #endif
9875 
9876 #ifdef __LITTLE_ENDIAN__
9877 #define vld1q_u8_x3(__p0) __extension__ ({ \
9878  uint8x16x3_t __ret; \
9879  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
9880  __ret; \
9881 })
9882 #else
9883 #define vld1q_u8_x3(__p0) __extension__ ({ \
9884  uint8x16x3_t __ret; \
9885  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
9886  \
9887  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9888  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9889  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9890  __ret; \
9891 })
9892 #endif
9893 
9894 #ifdef __LITTLE_ENDIAN__
9895 #define vld1q_u32_x3(__p0) __extension__ ({ \
9896  uint32x4x3_t __ret; \
9897  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
9898  __ret; \
9899 })
9900 #else
9901 #define vld1q_u32_x3(__p0) __extension__ ({ \
9902  uint32x4x3_t __ret; \
9903  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
9904  \
9905  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9906  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9907  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9908  __ret; \
9909 })
9910 #endif
9911 
9912 #ifdef __LITTLE_ENDIAN__
9913 #define vld1q_u64_x3(__p0) __extension__ ({ \
9914  uint64x2x3_t __ret; \
9915  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
9916  __ret; \
9917 })
9918 #else
9919 #define vld1q_u64_x3(__p0) __extension__ ({ \
9920  uint64x2x3_t __ret; \
9921  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
9922  \
9923  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9924  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9925  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9926  __ret; \
9927 })
9928 #endif
9929 
9930 #ifdef __LITTLE_ENDIAN__
9931 #define vld1q_u16_x3(__p0) __extension__ ({ \
9932  uint16x8x3_t __ret; \
9933  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
9934  __ret; \
9935 })
9936 #else
9937 #define vld1q_u16_x3(__p0) __extension__ ({ \
9938  uint16x8x3_t __ret; \
9939  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
9940  \
9941  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9942  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9943  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9944  __ret; \
9945 })
9946 #endif
9947 
9948 #ifdef __LITTLE_ENDIAN__
9949 #define vld1q_s8_x3(__p0) __extension__ ({ \
9950  int8x16x3_t __ret; \
9951  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
9952  __ret; \
9953 })
9954 #else
9955 #define vld1q_s8_x3(__p0) __extension__ ({ \
9956  int8x16x3_t __ret; \
9957  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
9958  \
9959  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9960  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9961  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9962  __ret; \
9963 })
9964 #endif
9965 
9966 #ifdef __LITTLE_ENDIAN__
9967 #define vld1q_f32_x3(__p0) __extension__ ({ \
9968  float32x4x3_t __ret; \
9969  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
9970  __ret; \
9971 })
9972 #else
9973 #define vld1q_f32_x3(__p0) __extension__ ({ \
9974  float32x4x3_t __ret; \
9975  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
9976  \
9977  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9978  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9979  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9980  __ret; \
9981 })
9982 #endif
9983 
9984 #ifdef __LITTLE_ENDIAN__
9985 #define vld1q_s32_x3(__p0) __extension__ ({ \
9986  int32x4x3_t __ret; \
9987  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
9988  __ret; \
9989 })
9990 #else
9991 #define vld1q_s32_x3(__p0) __extension__ ({ \
9992  int32x4x3_t __ret; \
9993  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
9994  \
9995  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9996  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9997  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9998  __ret; \
9999 })
10000 #endif
10001 
10002 #ifdef __LITTLE_ENDIAN__
10003 #define vld1q_s64_x3(__p0) __extension__ ({ \
10004  int64x2x3_t __ret; \
10005  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
10006  __ret; \
10007 })
10008 #else
10009 #define vld1q_s64_x3(__p0) __extension__ ({ \
10010  int64x2x3_t __ret; \
10011  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
10012  \
10013  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10014  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10015  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10016  __ret; \
10017 })
10018 #endif
10019 
10020 #ifdef __LITTLE_ENDIAN__
10021 #define vld1q_s16_x3(__p0) __extension__ ({ \
10022  int16x8x3_t __ret; \
10023  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
10024  __ret; \
10025 })
10026 #else
10027 #define vld1q_s16_x3(__p0) __extension__ ({ \
10028  int16x8x3_t __ret; \
10029  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
10030  \
10031  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10032  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10033  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10034  __ret; \
10035 })
10036 #endif
10037 
10038 #ifdef __LITTLE_ENDIAN__
10039 #define vld1_u8_x3(__p0) __extension__ ({ \
10040  uint8x8x3_t __ret; \
10041  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
10042  __ret; \
10043 })
10044 #else
10045 #define vld1_u8_x3(__p0) __extension__ ({ \
10046  uint8x8x3_t __ret; \
10047  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
10048  \
10049  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10050  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10051  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10052  __ret; \
10053 })
10054 #endif
10055 
10056 #ifdef __LITTLE_ENDIAN__
10057 #define vld1_u32_x3(__p0) __extension__ ({ \
10058  uint32x2x3_t __ret; \
10059  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
10060  __ret; \
10061 })
10062 #else
10063 #define vld1_u32_x3(__p0) __extension__ ({ \
10064  uint32x2x3_t __ret; \
10065  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
10066  \
10067  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10068  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10069  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10070  __ret; \
10071 })
10072 #endif
10073 
10074 #define vld1_u64_x3(__p0) __extension__ ({ \
10075  uint64x1x3_t __ret; \
10076  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
10077  __ret; \
10078 })
10079 #ifdef __LITTLE_ENDIAN__
10080 #define vld1_u16_x3(__p0) __extension__ ({ \
10081  uint16x4x3_t __ret; \
10082  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
10083  __ret; \
10084 })
10085 #else
10086 #define vld1_u16_x3(__p0) __extension__ ({ \
10087  uint16x4x3_t __ret; \
10088  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
10089  \
10090  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10091  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10092  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10093  __ret; \
10094 })
10095 #endif
10096 
10097 #ifdef __LITTLE_ENDIAN__
10098 #define vld1_s8_x3(__p0) __extension__ ({ \
10099  int8x8x3_t __ret; \
10100  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
10101  __ret; \
10102 })
10103 #else
10104 #define vld1_s8_x3(__p0) __extension__ ({ \
10105  int8x8x3_t __ret; \
10106  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
10107  \
10108  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10109  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10110  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10111  __ret; \
10112 })
10113 #endif
10114 
10115 #ifdef __LITTLE_ENDIAN__
10116 #define vld1_f32_x3(__p0) __extension__ ({ \
10117  float32x2x3_t __ret; \
10118  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
10119  __ret; \
10120 })
10121 #else
10122 #define vld1_f32_x3(__p0) __extension__ ({ \
10123  float32x2x3_t __ret; \
10124  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
10125  \
10126  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10127  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10128  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10129  __ret; \
10130 })
10131 #endif
10132 
10133 #ifdef __LITTLE_ENDIAN__
10134 #define vld1_s32_x3(__p0) __extension__ ({ \
10135  int32x2x3_t __ret; \
10136  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
10137  __ret; \
10138 })
10139 #else
10140 #define vld1_s32_x3(__p0) __extension__ ({ \
10141  int32x2x3_t __ret; \
10142  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
10143  \
10144  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10145  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10146  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10147  __ret; \
10148 })
10149 #endif
10150 
10151 #define vld1_s64_x3(__p0) __extension__ ({ \
10152  int64x1x3_t __ret; \
10153  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
10154  __ret; \
10155 })
10156 #ifdef __LITTLE_ENDIAN__
10157 #define vld1_s16_x3(__p0) __extension__ ({ \
10158  int16x4x3_t __ret; \
10159  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
10160  __ret; \
10161 })
10162 #else
10163 #define vld1_s16_x3(__p0) __extension__ ({ \
10164  int16x4x3_t __ret; \
10165  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
10166  \
10167  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10168  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10169  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10170  __ret; \
10171 })
10172 #endif
10173 
10174 #ifdef __LITTLE_ENDIAN__
10175 #define vld1_p8_x4(__p0) __extension__ ({ \
10176  poly8x8x4_t __ret; \
10177  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
10178  __ret; \
10179 })
10180 #else
10181 #define vld1_p8_x4(__p0) __extension__ ({ \
10182  poly8x8x4_t __ret; \
10183  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
10184  \
10185  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10186  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10187  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10188  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10189  __ret; \
10190 })
10191 #endif
10192 
10193 #ifdef __LITTLE_ENDIAN__
10194 #define vld1_p16_x4(__p0) __extension__ ({ \
10195  poly16x4x4_t __ret; \
10196  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
10197  __ret; \
10198 })
10199 #else
10200 #define vld1_p16_x4(__p0) __extension__ ({ \
10201  poly16x4x4_t __ret; \
10202  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
10203  \
10204  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10205  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10206  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10207  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10208  __ret; \
10209 })
10210 #endif
10211 
10212 #ifdef __LITTLE_ENDIAN__
10213 #define vld1q_p8_x4(__p0) __extension__ ({ \
10214  poly8x16x4_t __ret; \
10215  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
10216  __ret; \
10217 })
10218 #else
10219 #define vld1q_p8_x4(__p0) __extension__ ({ \
10220  poly8x16x4_t __ret; \
10221  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
10222  \
10223  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10224  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10225  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10226  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10227  __ret; \
10228 })
10229 #endif
10230 
10231 #ifdef __LITTLE_ENDIAN__
10232 #define vld1q_p16_x4(__p0) __extension__ ({ \
10233  poly16x8x4_t __ret; \
10234  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
10235  __ret; \
10236 })
10237 #else
10238 #define vld1q_p16_x4(__p0) __extension__ ({ \
10239  poly16x8x4_t __ret; \
10240  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
10241  \
10242  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10243  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10244  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10245  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10246  __ret; \
10247 })
10248 #endif
10249 
10250 #ifdef __LITTLE_ENDIAN__
10251 #define vld1q_u8_x4(__p0) __extension__ ({ \
10252  uint8x16x4_t __ret; \
10253  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
10254  __ret; \
10255 })
10256 #else
10257 #define vld1q_u8_x4(__p0) __extension__ ({ \
10258  uint8x16x4_t __ret; \
10259  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
10260  \
10261  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10262  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10263  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10264  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10265  __ret; \
10266 })
10267 #endif
10268 
10269 #ifdef __LITTLE_ENDIAN__
10270 #define vld1q_u32_x4(__p0) __extension__ ({ \
10271  uint32x4x4_t __ret; \
10272  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
10273  __ret; \
10274 })
10275 #else
10276 #define vld1q_u32_x4(__p0) __extension__ ({ \
10277  uint32x4x4_t __ret; \
10278  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
10279  \
10280  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10281  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10282  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10283  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10284  __ret; \
10285 })
10286 #endif
10287 
10288 #ifdef __LITTLE_ENDIAN__
10289 #define vld1q_u64_x4(__p0) __extension__ ({ \
10290  uint64x2x4_t __ret; \
10291  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
10292  __ret; \
10293 })
10294 #else
10295 #define vld1q_u64_x4(__p0) __extension__ ({ \
10296  uint64x2x4_t __ret; \
10297  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
10298  \
10299  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10300  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10301  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10302  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10303  __ret; \
10304 })
10305 #endif
10306 
10307 #ifdef __LITTLE_ENDIAN__
10308 #define vld1q_u16_x4(__p0) __extension__ ({ \
10309  uint16x8x4_t __ret; \
10310  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
10311  __ret; \
10312 })
10313 #else
10314 #define vld1q_u16_x4(__p0) __extension__ ({ \
10315  uint16x8x4_t __ret; \
10316  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
10317  \
10318  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10319  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10320  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10321  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10322  __ret; \
10323 })
10324 #endif
10325 
10326 #ifdef __LITTLE_ENDIAN__
10327 #define vld1q_s8_x4(__p0) __extension__ ({ \
10328  int8x16x4_t __ret; \
10329  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
10330  __ret; \
10331 })
10332 #else
10333 #define vld1q_s8_x4(__p0) __extension__ ({ \
10334  int8x16x4_t __ret; \
10335  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
10336  \
10337  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10338  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10339  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10340  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10341  __ret; \
10342 })
10343 #endif
10344 
10345 #ifdef __LITTLE_ENDIAN__
10346 #define vld1q_f32_x4(__p0) __extension__ ({ \
10347  float32x4x4_t __ret; \
10348  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
10349  __ret; \
10350 })
10351 #else
10352 #define vld1q_f32_x4(__p0) __extension__ ({ \
10353  float32x4x4_t __ret; \
10354  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
10355  \
10356  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10357  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10358  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10359  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10360  __ret; \
10361 })
10362 #endif
10363 
10364 #ifdef __LITTLE_ENDIAN__
10365 #define vld1q_s32_x4(__p0) __extension__ ({ \
10366  int32x4x4_t __ret; \
10367  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
10368  __ret; \
10369 })
10370 #else
10371 #define vld1q_s32_x4(__p0) __extension__ ({ \
10372  int32x4x4_t __ret; \
10373  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
10374  \
10375  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10376  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10377  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10378  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10379  __ret; \
10380 })
10381 #endif
10382 
10383 #ifdef __LITTLE_ENDIAN__
10384 #define vld1q_s64_x4(__p0) __extension__ ({ \
10385  int64x2x4_t __ret; \
10386  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
10387  __ret; \
10388 })
10389 #else
10390 #define vld1q_s64_x4(__p0) __extension__ ({ \
10391  int64x2x4_t __ret; \
10392  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
10393  \
10394  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10395  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10396  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10397  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10398  __ret; \
10399 })
10400 #endif
10401 
10402 #ifdef __LITTLE_ENDIAN__
10403 #define vld1q_s16_x4(__p0) __extension__ ({ \
10404  int16x8x4_t __ret; \
10405  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
10406  __ret; \
10407 })
10408 #else
10409 #define vld1q_s16_x4(__p0) __extension__ ({ \
10410  int16x8x4_t __ret; \
10411  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
10412  \
10413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10415  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10416  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10417  __ret; \
10418 })
10419 #endif
10420 
10421 #ifdef __LITTLE_ENDIAN__
10422 #define vld1_u8_x4(__p0) __extension__ ({ \
10423  uint8x8x4_t __ret; \
10424  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
10425  __ret; \
10426 })
10427 #else
10428 #define vld1_u8_x4(__p0) __extension__ ({ \
10429  uint8x8x4_t __ret; \
10430  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
10431  \
10432  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10433  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10434  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10435  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10436  __ret; \
10437 })
10438 #endif
10439 
10440 #ifdef __LITTLE_ENDIAN__
10441 #define vld1_u32_x4(__p0) __extension__ ({ \
10442  uint32x2x4_t __ret; \
10443  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
10444  __ret; \
10445 })
10446 #else
10447 #define vld1_u32_x4(__p0) __extension__ ({ \
10448  uint32x2x4_t __ret; \
10449  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
10450  \
10451  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10452  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10453  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10454  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10455  __ret; \
10456 })
10457 #endif
10458 
10459 #define vld1_u64_x4(__p0) __extension__ ({ \
10460  uint64x1x4_t __ret; \
10461  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
10462  __ret; \
10463 })
10464 #ifdef __LITTLE_ENDIAN__
10465 #define vld1_u16_x4(__p0) __extension__ ({ \
10466  uint16x4x4_t __ret; \
10467  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
10468  __ret; \
10469 })
10470 #else
10471 #define vld1_u16_x4(__p0) __extension__ ({ \
10472  uint16x4x4_t __ret; \
10473  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
10474  \
10475  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10476  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10477  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10478  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10479  __ret; \
10480 })
10481 #endif
10482 
10483 #ifdef __LITTLE_ENDIAN__
10484 #define vld1_s8_x4(__p0) __extension__ ({ \
10485  int8x8x4_t __ret; \
10486  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
10487  __ret; \
10488 })
10489 #else
10490 #define vld1_s8_x4(__p0) __extension__ ({ \
10491  int8x8x4_t __ret; \
10492  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
10493  \
10494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10496  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10497  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10498  __ret; \
10499 })
10500 #endif
10501 
10502 #ifdef __LITTLE_ENDIAN__
10503 #define vld1_f32_x4(__p0) __extension__ ({ \
10504  float32x2x4_t __ret; \
10505  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
10506  __ret; \
10507 })
10508 #else
10509 #define vld1_f32_x4(__p0) __extension__ ({ \
10510  float32x2x4_t __ret; \
10511  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
10512  \
10513  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10514  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10515  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10516  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10517  __ret; \
10518 })
10519 #endif
10520 
10521 #ifdef __LITTLE_ENDIAN__
10522 #define vld1_s32_x4(__p0) __extension__ ({ \
10523  int32x2x4_t __ret; \
10524  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
10525  __ret; \
10526 })
10527 #else
10528 #define vld1_s32_x4(__p0) __extension__ ({ \
10529  int32x2x4_t __ret; \
10530  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
10531  \
10532  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10533  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10534  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10535  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10536  __ret; \
10537 })
10538 #endif
10539 
10540 #define vld1_s64_x4(__p0) __extension__ ({ \
10541  int64x1x4_t __ret; \
10542  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
10543  __ret; \
10544 })
10545 #ifdef __LITTLE_ENDIAN__
10546 #define vld1_s16_x4(__p0) __extension__ ({ \
10547  int16x4x4_t __ret; \
10548  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
10549  __ret; \
10550 })
10551 #else
10552 #define vld1_s16_x4(__p0) __extension__ ({ \
10553  int16x4x4_t __ret; \
10554  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
10555  \
10556  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10557  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10558  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10559  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10560  __ret; \
10561 })
10562 #endif
10563 
10564 #ifdef __LITTLE_ENDIAN__
10565 #define vld2_p8(__p0) __extension__ ({ \
10566  poly8x8x2_t __ret; \
10567  __builtin_neon_vld2_v(&__ret, __p0, 4); \
10568  __ret; \
10569 })
10570 #else
10571 #define vld2_p8(__p0) __extension__ ({ \
10572  poly8x8x2_t __ret; \
10573  __builtin_neon_vld2_v(&__ret, __p0, 4); \
10574  \
10575  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10576  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10577  __ret; \
10578 })
10579 #endif
10580 
10581 #ifdef __LITTLE_ENDIAN__
10582 #define vld2_p16(__p0) __extension__ ({ \
10583  poly16x4x2_t __ret; \
10584  __builtin_neon_vld2_v(&__ret, __p0, 5); \
10585  __ret; \
10586 })
10587 #else
10588 #define vld2_p16(__p0) __extension__ ({ \
10589  poly16x4x2_t __ret; \
10590  __builtin_neon_vld2_v(&__ret, __p0, 5); \
10591  \
10592  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10593  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10594  __ret; \
10595 })
10596 #endif
10597 
10598 #ifdef __LITTLE_ENDIAN__
10599 #define vld2q_p8(__p0) __extension__ ({ \
10600  poly8x16x2_t __ret; \
10601  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
10602  __ret; \
10603 })
10604 #else
10605 #define vld2q_p8(__p0) __extension__ ({ \
10606  poly8x16x2_t __ret; \
10607  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
10608  \
10609  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10610  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10611  __ret; \
10612 })
10613 #endif
10614 
10615 #ifdef __LITTLE_ENDIAN__
10616 #define vld2q_p16(__p0) __extension__ ({ \
10617  poly16x8x2_t __ret; \
10618  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
10619  __ret; \
10620 })
10621 #else
10622 #define vld2q_p16(__p0) __extension__ ({ \
10623  poly16x8x2_t __ret; \
10624  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
10625  \
10626  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10627  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10628  __ret; \
10629 })
10630 #endif
10631 
10632 #ifdef __LITTLE_ENDIAN__
10633 #define vld2q_u8(__p0) __extension__ ({ \
10634  uint8x16x2_t __ret; \
10635  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
10636  __ret; \
10637 })
10638 #else
10639 #define vld2q_u8(__p0) __extension__ ({ \
10640  uint8x16x2_t __ret; \
10641  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
10642  \
10643  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10644  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10645  __ret; \
10646 })
10647 #endif
10648 
10649 #ifdef __LITTLE_ENDIAN__
10650 #define vld2q_u32(__p0) __extension__ ({ \
10651  uint32x4x2_t __ret; \
10652  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
10653  __ret; \
10654 })
10655 #else
10656 #define vld2q_u32(__p0) __extension__ ({ \
10657  uint32x4x2_t __ret; \
10658  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
10659  \
10660  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10661  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10662  __ret; \
10663 })
10664 #endif
10665 
10666 #ifdef __LITTLE_ENDIAN__
10667 #define vld2q_u16(__p0) __extension__ ({ \
10668  uint16x8x2_t __ret; \
10669  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
10670  __ret; \
10671 })
10672 #else
10673 #define vld2q_u16(__p0) __extension__ ({ \
10674  uint16x8x2_t __ret; \
10675  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
10676  \
10677  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10678  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10679  __ret; \
10680 })
10681 #endif
10682 
10683 #ifdef __LITTLE_ENDIAN__
10684 #define vld2q_s8(__p0) __extension__ ({ \
10685  int8x16x2_t __ret; \
10686  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
10687  __ret; \
10688 })
10689 #else
10690 #define vld2q_s8(__p0) __extension__ ({ \
10691  int8x16x2_t __ret; \
10692  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
10693  \
10694  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10695  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10696  __ret; \
10697 })
10698 #endif
10699 
10700 #ifdef __LITTLE_ENDIAN__
10701 #define vld2q_f32(__p0) __extension__ ({ \
10702  float32x4x2_t __ret; \
10703  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
10704  __ret; \
10705 })
10706 #else
10707 #define vld2q_f32(__p0) __extension__ ({ \
10708  float32x4x2_t __ret; \
10709  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
10710  \
10711  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10712  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10713  __ret; \
10714 })
10715 #endif
10716 
10717 #ifdef __LITTLE_ENDIAN__
10718 #define vld2q_s32(__p0) __extension__ ({ \
10719  int32x4x2_t __ret; \
10720  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
10721  __ret; \
10722 })
10723 #else
10724 #define vld2q_s32(__p0) __extension__ ({ \
10725  int32x4x2_t __ret; \
10726  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
10727  \
10728  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10729  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10730  __ret; \
10731 })
10732 #endif
10733 
10734 #ifdef __LITTLE_ENDIAN__
10735 #define vld2q_s16(__p0) __extension__ ({ \
10736  int16x8x2_t __ret; \
10737  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
10738  __ret; \
10739 })
10740 #else
10741 #define vld2q_s16(__p0) __extension__ ({ \
10742  int16x8x2_t __ret; \
10743  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
10744  \
10745  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10746  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10747  __ret; \
10748 })
10749 #endif
10750 
10751 #ifdef __LITTLE_ENDIAN__
10752 #define vld2_u8(__p0) __extension__ ({ \
10753  uint8x8x2_t __ret; \
10754  __builtin_neon_vld2_v(&__ret, __p0, 16); \
10755  __ret; \
10756 })
10757 #else
10758 #define vld2_u8(__p0) __extension__ ({ \
10759  uint8x8x2_t __ret; \
10760  __builtin_neon_vld2_v(&__ret, __p0, 16); \
10761  \
10762  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10763  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10764  __ret; \
10765 })
10766 #endif
10767 
10768 #ifdef __LITTLE_ENDIAN__
10769 #define vld2_u32(__p0) __extension__ ({ \
10770  uint32x2x2_t __ret; \
10771  __builtin_neon_vld2_v(&__ret, __p0, 18); \
10772  __ret; \
10773 })
10774 #else
10775 #define vld2_u32(__p0) __extension__ ({ \
10776  uint32x2x2_t __ret; \
10777  __builtin_neon_vld2_v(&__ret, __p0, 18); \
10778  \
10779  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10780  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10781  __ret; \
10782 })
10783 #endif
10784 
10785 #define vld2_u64(__p0) __extension__ ({ \
10786  uint64x1x2_t __ret; \
10787  __builtin_neon_vld2_v(&__ret, __p0, 19); \
10788  __ret; \
10789 })
10790 #ifdef __LITTLE_ENDIAN__
10791 #define vld2_u16(__p0) __extension__ ({ \
10792  uint16x4x2_t __ret; \
10793  __builtin_neon_vld2_v(&__ret, __p0, 17); \
10794  __ret; \
10795 })
10796 #else
10797 #define vld2_u16(__p0) __extension__ ({ \
10798  uint16x4x2_t __ret; \
10799  __builtin_neon_vld2_v(&__ret, __p0, 17); \
10800  \
10801  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10802  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10803  __ret; \
10804 })
10805 #endif
10806 
10807 #ifdef __LITTLE_ENDIAN__
10808 #define vld2_s8(__p0) __extension__ ({ \
10809  int8x8x2_t __ret; \
10810  __builtin_neon_vld2_v(&__ret, __p0, 0); \
10811  __ret; \
10812 })
10813 #else
10814 #define vld2_s8(__p0) __extension__ ({ \
10815  int8x8x2_t __ret; \
10816  __builtin_neon_vld2_v(&__ret, __p0, 0); \
10817  \
10818  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10819  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10820  __ret; \
10821 })
10822 #endif
10823 
10824 #ifdef __LITTLE_ENDIAN__
10825 #define vld2_f32(__p0) __extension__ ({ \
10826  float32x2x2_t __ret; \
10827  __builtin_neon_vld2_v(&__ret, __p0, 9); \
10828  __ret; \
10829 })
10830 #else
10831 #define vld2_f32(__p0) __extension__ ({ \
10832  float32x2x2_t __ret; \
10833  __builtin_neon_vld2_v(&__ret, __p0, 9); \
10834  \
10835  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10836  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10837  __ret; \
10838 })
10839 #endif
10840 
10841 #ifdef __LITTLE_ENDIAN__
10842 #define vld2_s32(__p0) __extension__ ({ \
10843  int32x2x2_t __ret; \
10844  __builtin_neon_vld2_v(&__ret, __p0, 2); \
10845  __ret; \
10846 })
10847 #else
10848 #define vld2_s32(__p0) __extension__ ({ \
10849  int32x2x2_t __ret; \
10850  __builtin_neon_vld2_v(&__ret, __p0, 2); \
10851  \
10852  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10853  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10854  __ret; \
10855 })
10856 #endif
10857 
10858 #define vld2_s64(__p0) __extension__ ({ \
10859  int64x1x2_t __ret; \
10860  __builtin_neon_vld2_v(&__ret, __p0, 3); \
10861  __ret; \
10862 })
10863 #ifdef __LITTLE_ENDIAN__
10864 #define vld2_s16(__p0) __extension__ ({ \
10865  int16x4x2_t __ret; \
10866  __builtin_neon_vld2_v(&__ret, __p0, 1); \
10867  __ret; \
10868 })
10869 #else
10870 #define vld2_s16(__p0) __extension__ ({ \
10871  int16x4x2_t __ret; \
10872  __builtin_neon_vld2_v(&__ret, __p0, 1); \
10873  \
10874  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10875  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10876  __ret; \
10877 })
10878 #endif
10879 
10880 #ifdef __LITTLE_ENDIAN__
10881 #define vld2_dup_p8(__p0) __extension__ ({ \
10882  poly8x8x2_t __ret; \
10883  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
10884  __ret; \
10885 })
10886 #else
10887 #define vld2_dup_p8(__p0) __extension__ ({ \
10888  poly8x8x2_t __ret; \
10889  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
10890  \
10891  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10892  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10893  __ret; \
10894 })
10895 #endif
10896 
10897 #ifdef __LITTLE_ENDIAN__
10898 #define vld2_dup_p16(__p0) __extension__ ({ \
10899  poly16x4x2_t __ret; \
10900  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
10901  __ret; \
10902 })
10903 #else
10904 #define vld2_dup_p16(__p0) __extension__ ({ \
10905  poly16x4x2_t __ret; \
10906  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
10907  \
10908  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10909  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10910  __ret; \
10911 })
10912 #endif
10913 
10914 #ifdef __LITTLE_ENDIAN__
10915 #define vld2q_dup_p8(__p0) __extension__ ({ \
10916  poly8x16x2_t __ret; \
10917  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
10918  __ret; \
10919 })
10920 #else
10921 #define vld2q_dup_p8(__p0) __extension__ ({ \
10922  poly8x16x2_t __ret; \
10923  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
10924  \
10925  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10926  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10927  __ret; \
10928 })
10929 #endif
10930 
10931 #ifdef __LITTLE_ENDIAN__
10932 #define vld2q_dup_p16(__p0) __extension__ ({ \
10933  poly16x8x2_t __ret; \
10934  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
10935  __ret; \
10936 })
10937 #else
10938 #define vld2q_dup_p16(__p0) __extension__ ({ \
10939  poly16x8x2_t __ret; \
10940  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
10941  \
10942  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10943  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10944  __ret; \
10945 })
10946 #endif
10947 
10948 #ifdef __LITTLE_ENDIAN__
10949 #define vld2q_dup_u8(__p0) __extension__ ({ \
10950  uint8x16x2_t __ret; \
10951  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
10952  __ret; \
10953 })
10954 #else
10955 #define vld2q_dup_u8(__p0) __extension__ ({ \
10956  uint8x16x2_t __ret; \
10957  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
10958  \
10959  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10960  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10961  __ret; \
10962 })
10963 #endif
10964 
10965 #ifdef __LITTLE_ENDIAN__
10966 #define vld2q_dup_u32(__p0) __extension__ ({ \
10967  uint32x4x2_t __ret; \
10968  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
10969  __ret; \
10970 })
10971 #else
10972 #define vld2q_dup_u32(__p0) __extension__ ({ \
10973  uint32x4x2_t __ret; \
10974  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
10975  \
10976  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10977  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10978  __ret; \
10979 })
10980 #endif
10981 
10982 #ifdef __LITTLE_ENDIAN__
10983 #define vld2q_dup_u64(__p0) __extension__ ({ \
10984  uint64x2x2_t __ret; \
10985  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
10986  __ret; \
10987 })
10988 #else
10989 #define vld2q_dup_u64(__p0) __extension__ ({ \
10990  uint64x2x2_t __ret; \
10991  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
10992  \
10993  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10994  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10995  __ret; \
10996 })
10997 #endif
10998 
10999 #ifdef __LITTLE_ENDIAN__
11000 #define vld2q_dup_u16(__p0) __extension__ ({ \
11001  uint16x8x2_t __ret; \
11002  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
11003  __ret; \
11004 })
11005 #else
11006 #define vld2q_dup_u16(__p0) __extension__ ({ \
11007  uint16x8x2_t __ret; \
11008  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
11009  \
11010  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11011  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11012  __ret; \
11013 })
11014 #endif
11015 
11016 #ifdef __LITTLE_ENDIAN__
11017 #define vld2q_dup_s8(__p0) __extension__ ({ \
11018  int8x16x2_t __ret; \
11019  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
11020  __ret; \
11021 })
11022 #else
11023 #define vld2q_dup_s8(__p0) __extension__ ({ \
11024  int8x16x2_t __ret; \
11025  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
11026  \
11027  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11028  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11029  __ret; \
11030 })
11031 #endif
11032 
11033 #ifdef __LITTLE_ENDIAN__
11034 #define vld2q_dup_f32(__p0) __extension__ ({ \
11035  float32x4x2_t __ret; \
11036  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
11037  __ret; \
11038 })
11039 #else
11040 #define vld2q_dup_f32(__p0) __extension__ ({ \
11041  float32x4x2_t __ret; \
11042  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
11043  \
11044  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11045  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11046  __ret; \
11047 })
11048 #endif
11049 
11050 #ifdef __LITTLE_ENDIAN__
11051 #define vld2q_dup_s32(__p0) __extension__ ({ \
11052  int32x4x2_t __ret; \
11053  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
11054  __ret; \
11055 })
11056 #else
11057 #define vld2q_dup_s32(__p0) __extension__ ({ \
11058  int32x4x2_t __ret; \
11059  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
11060  \
11061  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11062  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11063  __ret; \
11064 })
11065 #endif
11066 
11067 #ifdef __LITTLE_ENDIAN__
11068 #define vld2q_dup_s64(__p0) __extension__ ({ \
11069  int64x2x2_t __ret; \
11070  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
11071  __ret; \
11072 })
11073 #else
11074 #define vld2q_dup_s64(__p0) __extension__ ({ \
11075  int64x2x2_t __ret; \
11076  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
11077  \
11078  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11079  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11080  __ret; \
11081 })
11082 #endif
11083 
11084 #ifdef __LITTLE_ENDIAN__
11085 #define vld2q_dup_s16(__p0) __extension__ ({ \
11086  int16x8x2_t __ret; \
11087  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
11088  __ret; \
11089 })
11090 #else
11091 #define vld2q_dup_s16(__p0) __extension__ ({ \
11092  int16x8x2_t __ret; \
11093  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
11094  \
11095  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11096  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11097  __ret; \
11098 })
11099 #endif
11100 
11101 #ifdef __LITTLE_ENDIAN__
11102 #define vld2_dup_u8(__p0) __extension__ ({ \
11103  uint8x8x2_t __ret; \
11104  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
11105  __ret; \
11106 })
11107 #else
11108 #define vld2_dup_u8(__p0) __extension__ ({ \
11109  uint8x8x2_t __ret; \
11110  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
11111  \
11112  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11113  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11114  __ret; \
11115 })
11116 #endif
11117 
11118 #ifdef __LITTLE_ENDIAN__
11119 #define vld2_dup_u32(__p0) __extension__ ({ \
11120  uint32x2x2_t __ret; \
11121  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
11122  __ret; \
11123 })
11124 #else
11125 #define vld2_dup_u32(__p0) __extension__ ({ \
11126  uint32x2x2_t __ret; \
11127  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
11128  \
11129  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11130  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11131  __ret; \
11132 })
11133 #endif
11134 
11135 #define vld2_dup_u64(__p0) __extension__ ({ \
11136  uint64x1x2_t __ret; \
11137  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
11138  __ret; \
11139 })
11140 #ifdef __LITTLE_ENDIAN__
11141 #define vld2_dup_u16(__p0) __extension__ ({ \
11142  uint16x4x2_t __ret; \
11143  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
11144  __ret; \
11145 })
11146 #else
11147 #define vld2_dup_u16(__p0) __extension__ ({ \
11148  uint16x4x2_t __ret; \
11149  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
11150  \
11151  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11152  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11153  __ret; \
11154 })
11155 #endif
11156 
11157 #ifdef __LITTLE_ENDIAN__
11158 #define vld2_dup_s8(__p0) __extension__ ({ \
11159  int8x8x2_t __ret; \
11160  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
11161  __ret; \
11162 })
11163 #else
11164 #define vld2_dup_s8(__p0) __extension__ ({ \
11165  int8x8x2_t __ret; \
11166  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
11167  \
11168  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11169  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11170  __ret; \
11171 })
11172 #endif
11173 
11174 #ifdef __LITTLE_ENDIAN__
11175 #define vld2_dup_f32(__p0) __extension__ ({ \
11176  float32x2x2_t __ret; \
11177  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
11178  __ret; \
11179 })
11180 #else
11181 #define vld2_dup_f32(__p0) __extension__ ({ \
11182  float32x2x2_t __ret; \
11183  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
11184  \
11185  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11186  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11187  __ret; \
11188 })
11189 #endif
11190 
11191 #ifdef __LITTLE_ENDIAN__
11192 #define vld2_dup_s32(__p0) __extension__ ({ \
11193  int32x2x2_t __ret; \
11194  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
11195  __ret; \
11196 })
11197 #else
11198 #define vld2_dup_s32(__p0) __extension__ ({ \
11199  int32x2x2_t __ret; \
11200  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
11201  \
11202  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11203  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11204  __ret; \
11205 })
11206 #endif
11207 
11208 #define vld2_dup_s64(__p0) __extension__ ({ \
11209  int64x1x2_t __ret; \
11210  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
11211  __ret; \
11212 })
11213 #ifdef __LITTLE_ENDIAN__
11214 #define vld2_dup_s16(__p0) __extension__ ({ \
11215  int16x4x2_t __ret; \
11216  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
11217  __ret; \
11218 })
11219 #else
11220 #define vld2_dup_s16(__p0) __extension__ ({ \
11221  int16x4x2_t __ret; \
11222  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
11223  \
11224  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11225  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11226  __ret; \
11227 })
11228 #endif
11229 
11230 #ifdef __LITTLE_ENDIAN__
11231 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11232  poly8x8x2_t __s1 = __p1; \
11233  poly8x8x2_t __ret; \
11234  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
11235  __ret; \
11236 })
11237 #else
11238 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11239  poly8x8x2_t __s1 = __p1; \
11240  poly8x8x2_t __rev1; \
11241  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11242  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11243  poly8x8x2_t __ret; \
11244  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
11245  \
11246  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11247  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11248  __ret; \
11249 })
11250 #endif
11251 
11252 #ifdef __LITTLE_ENDIAN__
11253 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11254  poly16x4x2_t __s1 = __p1; \
11255  poly16x4x2_t __ret; \
11256  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
11257  __ret; \
11258 })
11259 #else
11260 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11261  poly16x4x2_t __s1 = __p1; \
11262  poly16x4x2_t __rev1; \
11263  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11264  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11265  poly16x4x2_t __ret; \
11266  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
11267  \
11268  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11269  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11270  __ret; \
11271 })
11272 #endif
11273 
11274 #ifdef __LITTLE_ENDIAN__
11275 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11276  poly16x8x2_t __s1 = __p1; \
11277  poly16x8x2_t __ret; \
11278  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
11279  __ret; \
11280 })
11281 #else
11282 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11283  poly16x8x2_t __s1 = __p1; \
11284  poly16x8x2_t __rev1; \
11285  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11286  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11287  poly16x8x2_t __ret; \
11288  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
11289  \
11290  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11291  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11292  __ret; \
11293 })
11294 #endif
11295 
11296 #ifdef __LITTLE_ENDIAN__
11297 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11298  uint32x4x2_t __s1 = __p1; \
11299  uint32x4x2_t __ret; \
11300  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
11301  __ret; \
11302 })
11303 #else
11304 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11305  uint32x4x2_t __s1 = __p1; \
11306  uint32x4x2_t __rev1; \
11307  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11308  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11309  uint32x4x2_t __ret; \
11310  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
11311  \
11312  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11313  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11314  __ret; \
11315 })
11316 #endif
11317 
11318 #ifdef __LITTLE_ENDIAN__
11319 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11320  uint16x8x2_t __s1 = __p1; \
11321  uint16x8x2_t __ret; \
11322  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
11323  __ret; \
11324 })
11325 #else
11326 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11327  uint16x8x2_t __s1 = __p1; \
11328  uint16x8x2_t __rev1; \
11329  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11330  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11331  uint16x8x2_t __ret; \
11332  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
11333  \
11334  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11335  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11336  __ret; \
11337 })
11338 #endif
11339 
11340 #ifdef __LITTLE_ENDIAN__
11341 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11342  float32x4x2_t __s1 = __p1; \
11343  float32x4x2_t __ret; \
11344  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 41); \
11345  __ret; \
11346 })
11347 #else
11348 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11349  float32x4x2_t __s1 = __p1; \
11350  float32x4x2_t __rev1; \
11351  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11352  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11353  float32x4x2_t __ret; \
11354  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 41); \
11355  \
11356  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11357  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11358  __ret; \
11359 })
11360 #endif
11361 
11362 #ifdef __LITTLE_ENDIAN__
11363 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11364  int32x4x2_t __s1 = __p1; \
11365  int32x4x2_t __ret; \
11366  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 34); \
11367  __ret; \
11368 })
11369 #else
11370 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11371  int32x4x2_t __s1 = __p1; \
11372  int32x4x2_t __rev1; \
11373  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11374  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11375  int32x4x2_t __ret; \
11376  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 34); \
11377  \
11378  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11379  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11380  __ret; \
11381 })
11382 #endif
11383 
11384 #ifdef __LITTLE_ENDIAN__
11385 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11386  int16x8x2_t __s1 = __p1; \
11387  int16x8x2_t __ret; \
11388  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 33); \
11389  __ret; \
11390 })
11391 #else
11392 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11393  int16x8x2_t __s1 = __p1; \
11394  int16x8x2_t __rev1; \
11395  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11396  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11397  int16x8x2_t __ret; \
11398  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 33); \
11399  \
11400  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11401  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11402  __ret; \
11403 })
11404 #endif
11405 
11406 #ifdef __LITTLE_ENDIAN__
11407 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11408  uint8x8x2_t __s1 = __p1; \
11409  uint8x8x2_t __ret; \
11410  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
11411  __ret; \
11412 })
11413 #else
11414 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11415  uint8x8x2_t __s1 = __p1; \
11416  uint8x8x2_t __rev1; \
11417  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11418  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11419  uint8x8x2_t __ret; \
11420  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
11421  \
11422  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11423  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11424  __ret; \
11425 })
11426 #endif
11427 
11428 #ifdef __LITTLE_ENDIAN__
11429 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11430  uint32x2x2_t __s1 = __p1; \
11431  uint32x2x2_t __ret; \
11432  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
11433  __ret; \
11434 })
11435 #else
11436 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11437  uint32x2x2_t __s1 = __p1; \
11438  uint32x2x2_t __rev1; \
11439  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11440  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11441  uint32x2x2_t __ret; \
11442  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
11443  \
11444  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11445  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11446  __ret; \
11447 })
11448 #endif
11449 
11450 #ifdef __LITTLE_ENDIAN__
11451 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11452  uint16x4x2_t __s1 = __p1; \
11453  uint16x4x2_t __ret; \
11454  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
11455  __ret; \
11456 })
11457 #else
11458 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11459  uint16x4x2_t __s1 = __p1; \
11460  uint16x4x2_t __rev1; \
11461  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11462  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11463  uint16x4x2_t __ret; \
11464  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
11465  \
11466  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11467  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11468  __ret; \
11469 })
11470 #endif
11471 
11472 #ifdef __LITTLE_ENDIAN__
11473 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11474  int8x8x2_t __s1 = __p1; \
11475  int8x8x2_t __ret; \
11476  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
11477  __ret; \
11478 })
11479 #else
11480 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11481  int8x8x2_t __s1 = __p1; \
11482  int8x8x2_t __rev1; \
11483  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11484  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11485  int8x8x2_t __ret; \
11486  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
11487  \
11488  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11489  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11490  __ret; \
11491 })
11492 #endif
11493 
11494 #ifdef __LITTLE_ENDIAN__
11495 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11496  float32x2x2_t __s1 = __p1; \
11497  float32x2x2_t __ret; \
11498  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 9); \
11499  __ret; \
11500 })
11501 #else
11502 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11503  float32x2x2_t __s1 = __p1; \
11504  float32x2x2_t __rev1; \
11505  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11506  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11507  float32x2x2_t __ret; \
11508  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 9); \
11509  \
11510  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11511  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11512  __ret; \
11513 })
11514 #endif
11515 
11516 #ifdef __LITTLE_ENDIAN__
11517 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11518  int32x2x2_t __s1 = __p1; \
11519  int32x2x2_t __ret; \
11520  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 2); \
11521  __ret; \
11522 })
11523 #else
11524 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11525  int32x2x2_t __s1 = __p1; \
11526  int32x2x2_t __rev1; \
11527  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11528  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11529  int32x2x2_t __ret; \
11530  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 2); \
11531  \
11532  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11533  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11534  __ret; \
11535 })
11536 #endif
11537 
11538 #ifdef __LITTLE_ENDIAN__
11539 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11540  int16x4x2_t __s1 = __p1; \
11541  int16x4x2_t __ret; \
11542  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 1); \
11543  __ret; \
11544 })
11545 #else
11546 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11547  int16x4x2_t __s1 = __p1; \
11548  int16x4x2_t __rev1; \
11549  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11550  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11551  int16x4x2_t __ret; \
11552  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 1); \
11553  \
11554  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11555  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11556  __ret; \
11557 })
11558 #endif
11559 
11560 #ifdef __LITTLE_ENDIAN__
11561 #define vld3_p8(__p0) __extension__ ({ \
11562  poly8x8x3_t __ret; \
11563  __builtin_neon_vld3_v(&__ret, __p0, 4); \
11564  __ret; \
11565 })
11566 #else
11567 #define vld3_p8(__p0) __extension__ ({ \
11568  poly8x8x3_t __ret; \
11569  __builtin_neon_vld3_v(&__ret, __p0, 4); \
11570  \
11571  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11572  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11573  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11574  __ret; \
11575 })
11576 #endif
11577 
11578 #ifdef __LITTLE_ENDIAN__
11579 #define vld3_p16(__p0) __extension__ ({ \
11580  poly16x4x3_t __ret; \
11581  __builtin_neon_vld3_v(&__ret, __p0, 5); \
11582  __ret; \
11583 })
11584 #else
11585 #define vld3_p16(__p0) __extension__ ({ \
11586  poly16x4x3_t __ret; \
11587  __builtin_neon_vld3_v(&__ret, __p0, 5); \
11588  \
11589  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11590  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11591  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11592  __ret; \
11593 })
11594 #endif
11595 
11596 #ifdef __LITTLE_ENDIAN__
11597 #define vld3q_p8(__p0) __extension__ ({ \
11598  poly8x16x3_t __ret; \
11599  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
11600  __ret; \
11601 })
11602 #else
11603 #define vld3q_p8(__p0) __extension__ ({ \
11604  poly8x16x3_t __ret; \
11605  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
11606  \
11607  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11608  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11609  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11610  __ret; \
11611 })
11612 #endif
11613 
11614 #ifdef __LITTLE_ENDIAN__
11615 #define vld3q_p16(__p0) __extension__ ({ \
11616  poly16x8x3_t __ret; \
11617  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
11618  __ret; \
11619 })
11620 #else
11621 #define vld3q_p16(__p0) __extension__ ({ \
11622  poly16x8x3_t __ret; \
11623  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
11624  \
11625  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11626  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11627  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11628  __ret; \
11629 })
11630 #endif
11631 
11632 #ifdef __LITTLE_ENDIAN__
11633 #define vld3q_u8(__p0) __extension__ ({ \
11634  uint8x16x3_t __ret; \
11635  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
11636  __ret; \
11637 })
11638 #else
11639 #define vld3q_u8(__p0) __extension__ ({ \
11640  uint8x16x3_t __ret; \
11641  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
11642  \
11643  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11644  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11645  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11646  __ret; \
11647 })
11648 #endif
11649 
11650 #ifdef __LITTLE_ENDIAN__
11651 #define vld3q_u32(__p0) __extension__ ({ \
11652  uint32x4x3_t __ret; \
11653  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
11654  __ret; \
11655 })
11656 #else
11657 #define vld3q_u32(__p0) __extension__ ({ \
11658  uint32x4x3_t __ret; \
11659  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
11660  \
11661  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11662  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11663  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11664  __ret; \
11665 })
11666 #endif
11667 
11668 #ifdef __LITTLE_ENDIAN__
11669 #define vld3q_u16(__p0) __extension__ ({ \
11670  uint16x8x3_t __ret; \
11671  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
11672  __ret; \
11673 })
11674 #else
11675 #define vld3q_u16(__p0) __extension__ ({ \
11676  uint16x8x3_t __ret; \
11677  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
11678  \
11679  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11680  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11681  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11682  __ret; \
11683 })
11684 #endif
11685 
11686 #ifdef __LITTLE_ENDIAN__
11687 #define vld3q_s8(__p0) __extension__ ({ \
11688  int8x16x3_t __ret; \
11689  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
11690  __ret; \
11691 })
11692 #else
11693 #define vld3q_s8(__p0) __extension__ ({ \
11694  int8x16x3_t __ret; \
11695  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
11696  \
11697  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11698  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11699  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11700  __ret; \
11701 })
11702 #endif
11703 
11704 #ifdef __LITTLE_ENDIAN__
11705 #define vld3q_f32(__p0) __extension__ ({ \
11706  float32x4x3_t __ret; \
11707  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
11708  __ret; \
11709 })
11710 #else
11711 #define vld3q_f32(__p0) __extension__ ({ \
11712  float32x4x3_t __ret; \
11713  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
11714  \
11715  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11716  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11717  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11718  __ret; \
11719 })
11720 #endif
11721 
11722 #ifdef __LITTLE_ENDIAN__
11723 #define vld3q_s32(__p0) __extension__ ({ \
11724  int32x4x3_t __ret; \
11725  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
11726  __ret; \
11727 })
11728 #else
11729 #define vld3q_s32(__p0) __extension__ ({ \
11730  int32x4x3_t __ret; \
11731  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
11732  \
11733  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11734  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11735  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11736  __ret; \
11737 })
11738 #endif
11739 
11740 #ifdef __LITTLE_ENDIAN__
11741 #define vld3q_s16(__p0) __extension__ ({ \
11742  int16x8x3_t __ret; \
11743  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
11744  __ret; \
11745 })
11746 #else
11747 #define vld3q_s16(__p0) __extension__ ({ \
11748  int16x8x3_t __ret; \
11749  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
11750  \
11751  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11752  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11753  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11754  __ret; \
11755 })
11756 #endif
11757 
11758 #ifdef __LITTLE_ENDIAN__
11759 #define vld3_u8(__p0) __extension__ ({ \
11760  uint8x8x3_t __ret; \
11761  __builtin_neon_vld3_v(&__ret, __p0, 16); \
11762  __ret; \
11763 })
11764 #else
11765 #define vld3_u8(__p0) __extension__ ({ \
11766  uint8x8x3_t __ret; \
11767  __builtin_neon_vld3_v(&__ret, __p0, 16); \
11768  \
11769  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11770  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11771  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11772  __ret; \
11773 })
11774 #endif
11775 
11776 #ifdef __LITTLE_ENDIAN__
11777 #define vld3_u32(__p0) __extension__ ({ \
11778  uint32x2x3_t __ret; \
11779  __builtin_neon_vld3_v(&__ret, __p0, 18); \
11780  __ret; \
11781 })
11782 #else
11783 #define vld3_u32(__p0) __extension__ ({ \
11784  uint32x2x3_t __ret; \
11785  __builtin_neon_vld3_v(&__ret, __p0, 18); \
11786  \
11787  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11788  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11789  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11790  __ret; \
11791 })
11792 #endif
11793 
11794 #define vld3_u64(__p0) __extension__ ({ \
11795  uint64x1x3_t __ret; \
11796  __builtin_neon_vld3_v(&__ret, __p0, 19); \
11797  __ret; \
11798 })
11799 #ifdef __LITTLE_ENDIAN__
11800 #define vld3_u16(__p0) __extension__ ({ \
11801  uint16x4x3_t __ret; \
11802  __builtin_neon_vld3_v(&__ret, __p0, 17); \
11803  __ret; \
11804 })
11805 #else
11806 #define vld3_u16(__p0) __extension__ ({ \
11807  uint16x4x3_t __ret; \
11808  __builtin_neon_vld3_v(&__ret, __p0, 17); \
11809  \
11810  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11811  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11812  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11813  __ret; \
11814 })
11815 #endif
11816 
11817 #ifdef __LITTLE_ENDIAN__
11818 #define vld3_s8(__p0) __extension__ ({ \
11819  int8x8x3_t __ret; \
11820  __builtin_neon_vld3_v(&__ret, __p0, 0); \
11821  __ret; \
11822 })
11823 #else
11824 #define vld3_s8(__p0) __extension__ ({ \
11825  int8x8x3_t __ret; \
11826  __builtin_neon_vld3_v(&__ret, __p0, 0); \
11827  \
11828  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11829  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11830  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11831  __ret; \
11832 })
11833 #endif
11834 
11835 #ifdef __LITTLE_ENDIAN__
11836 #define vld3_f32(__p0) __extension__ ({ \
11837  float32x2x3_t __ret; \
11838  __builtin_neon_vld3_v(&__ret, __p0, 9); \
11839  __ret; \
11840 })
11841 #else
11842 #define vld3_f32(__p0) __extension__ ({ \
11843  float32x2x3_t __ret; \
11844  __builtin_neon_vld3_v(&__ret, __p0, 9); \
11845  \
11846  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11847  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11848  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11849  __ret; \
11850 })
11851 #endif
11852 
11853 #ifdef __LITTLE_ENDIAN__
11854 #define vld3_s32(__p0) __extension__ ({ \
11855  int32x2x3_t __ret; \
11856  __builtin_neon_vld3_v(&__ret, __p0, 2); \
11857  __ret; \
11858 })
11859 #else
11860 #define vld3_s32(__p0) __extension__ ({ \
11861  int32x2x3_t __ret; \
11862  __builtin_neon_vld3_v(&__ret, __p0, 2); \
11863  \
11864  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11865  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11866  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11867  __ret; \
11868 })
11869 #endif
11870 
11871 #define vld3_s64(__p0) __extension__ ({ \
11872  int64x1x3_t __ret; \
11873  __builtin_neon_vld3_v(&__ret, __p0, 3); \
11874  __ret; \
11875 })
11876 #ifdef __LITTLE_ENDIAN__
11877 #define vld3_s16(__p0) __extension__ ({ \
11878  int16x4x3_t __ret; \
11879  __builtin_neon_vld3_v(&__ret, __p0, 1); \
11880  __ret; \
11881 })
11882 #else
11883 #define vld3_s16(__p0) __extension__ ({ \
11884  int16x4x3_t __ret; \
11885  __builtin_neon_vld3_v(&__ret, __p0, 1); \
11886  \
11887  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11888  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11889  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11890  __ret; \
11891 })
11892 #endif
11893 
11894 #ifdef __LITTLE_ENDIAN__
11895 #define vld3_dup_p8(__p0) __extension__ ({ \
11896  poly8x8x3_t __ret; \
11897  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
11898  __ret; \
11899 })
11900 #else
11901 #define vld3_dup_p8(__p0) __extension__ ({ \
11902  poly8x8x3_t __ret; \
11903  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
11904  \
11905  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11906  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11907  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11908  __ret; \
11909 })
11910 #endif
11911 
11912 #ifdef __LITTLE_ENDIAN__
11913 #define vld3_dup_p16(__p0) __extension__ ({ \
11914  poly16x4x3_t __ret; \
11915  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
11916  __ret; \
11917 })
11918 #else
11919 #define vld3_dup_p16(__p0) __extension__ ({ \
11920  poly16x4x3_t __ret; \
11921  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
11922  \
11923  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11924  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11925  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11926  __ret; \
11927 })
11928 #endif
11929 
11930 #ifdef __LITTLE_ENDIAN__
11931 #define vld3q_dup_p8(__p0) __extension__ ({ \
11932  poly8x16x3_t __ret; \
11933  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
11934  __ret; \
11935 })
11936 #else
11937 #define vld3q_dup_p8(__p0) __extension__ ({ \
11938  poly8x16x3_t __ret; \
11939  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
11940  \
11941  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11942  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11943  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11944  __ret; \
11945 })
11946 #endif
11947 
11948 #ifdef __LITTLE_ENDIAN__
11949 #define vld3q_dup_p16(__p0) __extension__ ({ \
11950  poly16x8x3_t __ret; \
11951  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
11952  __ret; \
11953 })
11954 #else
11955 #define vld3q_dup_p16(__p0) __extension__ ({ \
11956  poly16x8x3_t __ret; \
11957  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
11958  \
11959  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11960  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11961  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11962  __ret; \
11963 })
11964 #endif
11965 
11966 #ifdef __LITTLE_ENDIAN__
11967 #define vld3q_dup_u8(__p0) __extension__ ({ \
11968  uint8x16x3_t __ret; \
11969  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
11970  __ret; \
11971 })
11972 #else
11973 #define vld3q_dup_u8(__p0) __extension__ ({ \
11974  uint8x16x3_t __ret; \
11975  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
11976  \
11977  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11978  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11979  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
11980  __ret; \
11981 })
11982 #endif
11983 
11984 #ifdef __LITTLE_ENDIAN__
11985 #define vld3q_dup_u32(__p0) __extension__ ({ \
11986  uint32x4x3_t __ret; \
11987  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
11988  __ret; \
11989 })
11990 #else
11991 #define vld3q_dup_u32(__p0) __extension__ ({ \
11992  uint32x4x3_t __ret; \
11993  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
11994  \
11995  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11996  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11997  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11998  __ret; \
11999 })
12000 #endif
12001 
12002 #ifdef __LITTLE_ENDIAN__
12003 #define vld3q_dup_u64(__p0) __extension__ ({ \
12004  uint64x2x3_t __ret; \
12005  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
12006  __ret; \
12007 })
12008 #else
12009 #define vld3q_dup_u64(__p0) __extension__ ({ \
12010  uint64x2x3_t __ret; \
12011  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
12012  \
12013  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12014  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12015  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12016  __ret; \
12017 })
12018 #endif
12019 
12020 #ifdef __LITTLE_ENDIAN__
12021 #define vld3q_dup_u16(__p0) __extension__ ({ \
12022  uint16x8x3_t __ret; \
12023  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
12024  __ret; \
12025 })
12026 #else
12027 #define vld3q_dup_u16(__p0) __extension__ ({ \
12028  uint16x8x3_t __ret; \
12029  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
12030  \
12031  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12032  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12033  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12034  __ret; \
12035 })
12036 #endif
12037 
12038 #ifdef __LITTLE_ENDIAN__
12039 #define vld3q_dup_s8(__p0) __extension__ ({ \
12040  int8x16x3_t __ret; \
12041  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
12042  __ret; \
12043 })
12044 #else
12045 #define vld3q_dup_s8(__p0) __extension__ ({ \
12046  int8x16x3_t __ret; \
12047  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
12048  \
12049  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12050  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12051  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12052  __ret; \
12053 })
12054 #endif
12055 
12056 #ifdef __LITTLE_ENDIAN__
12057 #define vld3q_dup_f32(__p0) __extension__ ({ \
12058  float32x4x3_t __ret; \
12059  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
12060  __ret; \
12061 })
12062 #else
12063 #define vld3q_dup_f32(__p0) __extension__ ({ \
12064  float32x4x3_t __ret; \
12065  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
12066  \
12067  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12068  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12069  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12070  __ret; \
12071 })
12072 #endif
12073 
12074 #ifdef __LITTLE_ENDIAN__
12075 #define vld3q_dup_s32(__p0) __extension__ ({ \
12076  int32x4x3_t __ret; \
12077  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
12078  __ret; \
12079 })
12080 #else
12081 #define vld3q_dup_s32(__p0) __extension__ ({ \
12082  int32x4x3_t __ret; \
12083  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
12084  \
12085  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12086  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12087  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12088  __ret; \
12089 })
12090 #endif
12091 
12092 #ifdef __LITTLE_ENDIAN__
12093 #define vld3q_dup_s64(__p0) __extension__ ({ \
12094  int64x2x3_t __ret; \
12095  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
12096  __ret; \
12097 })
12098 #else
12099 #define vld3q_dup_s64(__p0) __extension__ ({ \
12100  int64x2x3_t __ret; \
12101  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
12102  \
12103  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12104  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12105  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12106  __ret; \
12107 })
12108 #endif
12109 
12110 #ifdef __LITTLE_ENDIAN__
12111 #define vld3q_dup_s16(__p0) __extension__ ({ \
12112  int16x8x3_t __ret; \
12113  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
12114  __ret; \
12115 })
12116 #else
12117 #define vld3q_dup_s16(__p0) __extension__ ({ \
12118  int16x8x3_t __ret; \
12119  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
12120  \
12121  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12122  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12123  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12124  __ret; \
12125 })
12126 #endif
12127 
12128 #ifdef __LITTLE_ENDIAN__
12129 #define vld3_dup_u8(__p0) __extension__ ({ \
12130  uint8x8x3_t __ret; \
12131  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
12132  __ret; \
12133 })
12134 #else
12135 #define vld3_dup_u8(__p0) __extension__ ({ \
12136  uint8x8x3_t __ret; \
12137  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
12138  \
12139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12141  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12142  __ret; \
12143 })
12144 #endif
12145 
12146 #ifdef __LITTLE_ENDIAN__
12147 #define vld3_dup_u32(__p0) __extension__ ({ \
12148  uint32x2x3_t __ret; \
12149  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
12150  __ret; \
12151 })
12152 #else
12153 #define vld3_dup_u32(__p0) __extension__ ({ \
12154  uint32x2x3_t __ret; \
12155  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
12156  \
12157  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12158  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12159  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12160  __ret; \
12161 })
12162 #endif
12163 
12164 #define vld3_dup_u64(__p0) __extension__ ({ \
12165  uint64x1x3_t __ret; \
12166  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
12167  __ret; \
12168 })
12169 #ifdef __LITTLE_ENDIAN__
12170 #define vld3_dup_u16(__p0) __extension__ ({ \
12171  uint16x4x3_t __ret; \
12172  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
12173  __ret; \
12174 })
12175 #else
12176 #define vld3_dup_u16(__p0) __extension__ ({ \
12177  uint16x4x3_t __ret; \
12178  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
12179  \
12180  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12181  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12182  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12183  __ret; \
12184 })
12185 #endif
12186 
12187 #ifdef __LITTLE_ENDIAN__
12188 #define vld3_dup_s8(__p0) __extension__ ({ \
12189  int8x8x3_t __ret; \
12190  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
12191  __ret; \
12192 })
12193 #else
12194 #define vld3_dup_s8(__p0) __extension__ ({ \
12195  int8x8x3_t __ret; \
12196  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
12197  \
12198  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12199  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12200  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12201  __ret; \
12202 })
12203 #endif
12204 
12205 #ifdef __LITTLE_ENDIAN__
12206 #define vld3_dup_f32(__p0) __extension__ ({ \
12207  float32x2x3_t __ret; \
12208  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
12209  __ret; \
12210 })
12211 #else
12212 #define vld3_dup_f32(__p0) __extension__ ({ \
12213  float32x2x3_t __ret; \
12214  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
12215  \
12216  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12217  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12218  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12219  __ret; \
12220 })
12221 #endif
12222 
12223 #ifdef __LITTLE_ENDIAN__
12224 #define vld3_dup_s32(__p0) __extension__ ({ \
12225  int32x2x3_t __ret; \
12226  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
12227  __ret; \
12228 })
12229 #else
12230 #define vld3_dup_s32(__p0) __extension__ ({ \
12231  int32x2x3_t __ret; \
12232  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
12233  \
12234  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12235  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12236  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12237  __ret; \
12238 })
12239 #endif
12240 
12241 #define vld3_dup_s64(__p0) __extension__ ({ \
12242  int64x1x3_t __ret; \
12243  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
12244  __ret; \
12245 })
12246 #ifdef __LITTLE_ENDIAN__
12247 #define vld3_dup_s16(__p0) __extension__ ({ \
12248  int16x4x3_t __ret; \
12249  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
12250  __ret; \
12251 })
12252 #else
12253 #define vld3_dup_s16(__p0) __extension__ ({ \
12254  int16x4x3_t __ret; \
12255  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
12256  \
12257  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12258  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12259  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12260  __ret; \
12261 })
12262 #endif
12263 
12264 #ifdef __LITTLE_ENDIAN__
12265 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
12266  poly8x8x3_t __s1 = __p1; \
12267  poly8x8x3_t __ret; \
12268  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
12269  __ret; \
12270 })
12271 #else
12272 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
12273  poly8x8x3_t __s1 = __p1; \
12274  poly8x8x3_t __rev1; \
12275  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12276  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12277  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12278  poly8x8x3_t __ret; \
12279  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
12280  \
12281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12283  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12284  __ret; \
12285 })
12286 #endif
12287 
12288 #ifdef __LITTLE_ENDIAN__
12289 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12290  poly16x4x3_t __s1 = __p1; \
12291  poly16x4x3_t __ret; \
12292  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
12293  __ret; \
12294 })
12295 #else
12296 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12297  poly16x4x3_t __s1 = __p1; \
12298  poly16x4x3_t __rev1; \
12299  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12300  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12301  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12302  poly16x4x3_t __ret; \
12303  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
12304  \
12305  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12306  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12307  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12308  __ret; \
12309 })
12310 #endif
12311 
12312 #ifdef __LITTLE_ENDIAN__
12313 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12314  poly16x8x3_t __s1 = __p1; \
12315  poly16x8x3_t __ret; \
12316  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
12317  __ret; \
12318 })
12319 #else
12320 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
12321  poly16x8x3_t __s1 = __p1; \
12322  poly16x8x3_t __rev1; \
12323  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12324  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12325  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12326  poly16x8x3_t __ret; \
12327  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
12328  \
12329  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12330  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12331  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12332  __ret; \
12333 })
12334 #endif
12335 
12336 #ifdef __LITTLE_ENDIAN__
12337 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12338  uint32x4x3_t __s1 = __p1; \
12339  uint32x4x3_t __ret; \
12340  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
12341  __ret; \
12342 })
12343 #else
12344 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12345  uint32x4x3_t __s1 = __p1; \
12346  uint32x4x3_t __rev1; \
12347  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12348  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12349  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12350  uint32x4x3_t __ret; \
12351  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
12352  \
12353  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12354  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12355  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12356  __ret; \
12357 })
12358 #endif
12359 
12360 #ifdef __LITTLE_ENDIAN__
12361 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12362  uint16x8x3_t __s1 = __p1; \
12363  uint16x8x3_t __ret; \
12364  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
12365  __ret; \
12366 })
12367 #else
12368 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12369  uint16x8x3_t __s1 = __p1; \
12370  uint16x8x3_t __rev1; \
12371  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12372  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12373  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12374  uint16x8x3_t __ret; \
12375  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
12376  \
12377  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12378  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12379  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12380  __ret; \
12381 })
12382 #endif
12383 
12384 #ifdef __LITTLE_ENDIAN__
12385 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12386  float32x4x3_t __s1 = __p1; \
12387  float32x4x3_t __ret; \
12388  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 41); \
12389  __ret; \
12390 })
12391 #else
12392 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12393  float32x4x3_t __s1 = __p1; \
12394  float32x4x3_t __rev1; \
12395  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12396  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12397  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12398  float32x4x3_t __ret; \
12399  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 41); \
12400  \
12401  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12402  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12403  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12404  __ret; \
12405 })
12406 #endif
12407 
12408 #ifdef __LITTLE_ENDIAN__
12409 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12410  int32x4x3_t __s1 = __p1; \
12411  int32x4x3_t __ret; \
12412  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 34); \
12413  __ret; \
12414 })
12415 #else
12416 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12417  int32x4x3_t __s1 = __p1; \
12418  int32x4x3_t __rev1; \
12419  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12420  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12421  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12422  int32x4x3_t __ret; \
12423  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 34); \
12424  \
12425  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12426  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12427  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12428  __ret; \
12429 })
12430 #endif
12431 
12432 #ifdef __LITTLE_ENDIAN__
12433 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12434  int16x8x3_t __s1 = __p1; \
12435  int16x8x3_t __ret; \
12436  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 33); \
12437  __ret; \
12438 })
12439 #else
12440 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12441  int16x8x3_t __s1 = __p1; \
12442  int16x8x3_t __rev1; \
12443  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12444  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12445  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12446  int16x8x3_t __ret; \
12447  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 33); \
12448  \
12449  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12450  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12451  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12452  __ret; \
12453 })
12454 #endif
12455 
12456 #ifdef __LITTLE_ENDIAN__
12457 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
12458  uint8x8x3_t __s1 = __p1; \
12459  uint8x8x3_t __ret; \
12460  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
12461  __ret; \
12462 })
12463 #else
12464 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
12465  uint8x8x3_t __s1 = __p1; \
12466  uint8x8x3_t __rev1; \
12467  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12468  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12469  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12470  uint8x8x3_t __ret; \
12471  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
12472  \
12473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12476  __ret; \
12477 })
12478 #endif
12479 
12480 #ifdef __LITTLE_ENDIAN__
12481 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12482  uint32x2x3_t __s1 = __p1; \
12483  uint32x2x3_t __ret; \
12484  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
12485  __ret; \
12486 })
12487 #else
12488 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
12489  uint32x2x3_t __s1 = __p1; \
12490  uint32x2x3_t __rev1; \
12491  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12492  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12493  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12494  uint32x2x3_t __ret; \
12495  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
12496  \
12497  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12498  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12499  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12500  __ret; \
12501 })
12502 #endif
12503 
12504 #ifdef __LITTLE_ENDIAN__
12505 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12506  uint16x4x3_t __s1 = __p1; \
12507  uint16x4x3_t __ret; \
12508  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
12509  __ret; \
12510 })
12511 #else
12512 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
12513  uint16x4x3_t __s1 = __p1; \
12514  uint16x4x3_t __rev1; \
12515  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12516  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12517  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12518  uint16x4x3_t __ret; \
12519  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
12520  \
12521  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12522  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12523  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12524  __ret; \
12525 })
12526 #endif
12527 
12528 #ifdef __LITTLE_ENDIAN__
12529 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
12530  int8x8x3_t __s1 = __p1; \
12531  int8x8x3_t __ret; \
12532  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
12533  __ret; \
12534 })
12535 #else
12536 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
12537  int8x8x3_t __s1 = __p1; \
12538  int8x8x3_t __rev1; \
12539  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12540  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12541  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12542  int8x8x3_t __ret; \
12543  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
12544  \
12545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12547  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12548  __ret; \
12549 })
12550 #endif
12551 
12552 #ifdef __LITTLE_ENDIAN__
12553 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12554  float32x2x3_t __s1 = __p1; \
12555  float32x2x3_t __ret; \
12556  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 9); \
12557  __ret; \
12558 })
12559 #else
12560 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
12561  float32x2x3_t __s1 = __p1; \
12562  float32x2x3_t __rev1; \
12563  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12564  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12565  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12566  float32x2x3_t __ret; \
12567  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 9); \
12568  \
12569  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12570  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12571  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12572  __ret; \
12573 })
12574 #endif
12575 
12576 #ifdef __LITTLE_ENDIAN__
12577 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12578  int32x2x3_t __s1 = __p1; \
12579  int32x2x3_t __ret; \
12580  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 2); \
12581  __ret; \
12582 })
12583 #else
12584 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
12585  int32x2x3_t __s1 = __p1; \
12586  int32x2x3_t __rev1; \
12587  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
12588  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
12589  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
12590  int32x2x3_t __ret; \
12591  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 2); \
12592  \
12593  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12594  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12595  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12596  __ret; \
12597 })
12598 #endif
12599 
12600 #ifdef __LITTLE_ENDIAN__
12601 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12602  int16x4x3_t __s1 = __p1; \
12603  int16x4x3_t __ret; \
12604  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 1); \
12605  __ret; \
12606 })
12607 #else
12608 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
12609  int16x4x3_t __s1 = __p1; \
12610  int16x4x3_t __rev1; \
12611  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
12612  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
12613  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
12614  int16x4x3_t __ret; \
12615  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 1); \
12616  \
12617  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12618  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12619  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12620  __ret; \
12621 })
12622 #endif
12623 
12624 #ifdef __LITTLE_ENDIAN__
12625 #define vld4_p8(__p0) __extension__ ({ \
12626  poly8x8x4_t __ret; \
12627  __builtin_neon_vld4_v(&__ret, __p0, 4); \
12628  __ret; \
12629 })
12630 #else
12631 #define vld4_p8(__p0) __extension__ ({ \
12632  poly8x8x4_t __ret; \
12633  __builtin_neon_vld4_v(&__ret, __p0, 4); \
12634  \
12635  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12636  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12637  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12638  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12639  __ret; \
12640 })
12641 #endif
12642 
12643 #ifdef __LITTLE_ENDIAN__
12644 #define vld4_p16(__p0) __extension__ ({ \
12645  poly16x4x4_t __ret; \
12646  __builtin_neon_vld4_v(&__ret, __p0, 5); \
12647  __ret; \
12648 })
12649 #else
12650 #define vld4_p16(__p0) __extension__ ({ \
12651  poly16x4x4_t __ret; \
12652  __builtin_neon_vld4_v(&__ret, __p0, 5); \
12653  \
12654  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12655  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12656  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12657  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12658  __ret; \
12659 })
12660 #endif
12661 
12662 #ifdef __LITTLE_ENDIAN__
12663 #define vld4q_p8(__p0) __extension__ ({ \
12664  poly8x16x4_t __ret; \
12665  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
12666  __ret; \
12667 })
12668 #else
12669 #define vld4q_p8(__p0) __extension__ ({ \
12670  poly8x16x4_t __ret; \
12671  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
12672  \
12673  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12674  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12675  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12676  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12677  __ret; \
12678 })
12679 #endif
12680 
12681 #ifdef __LITTLE_ENDIAN__
12682 #define vld4q_p16(__p0) __extension__ ({ \
12683  poly16x8x4_t __ret; \
12684  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
12685  __ret; \
12686 })
12687 #else
12688 #define vld4q_p16(__p0) __extension__ ({ \
12689  poly16x8x4_t __ret; \
12690  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
12691  \
12692  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12693  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12694  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12695  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12696  __ret; \
12697 })
12698 #endif
12699 
12700 #ifdef __LITTLE_ENDIAN__
12701 #define vld4q_u8(__p0) __extension__ ({ \
12702  uint8x16x4_t __ret; \
12703  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
12704  __ret; \
12705 })
12706 #else
12707 #define vld4q_u8(__p0) __extension__ ({ \
12708  uint8x16x4_t __ret; \
12709  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
12710  \
12711  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12712  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12713  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12714  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12715  __ret; \
12716 })
12717 #endif
12718 
12719 #ifdef __LITTLE_ENDIAN__
12720 #define vld4q_u32(__p0) __extension__ ({ \
12721  uint32x4x4_t __ret; \
12722  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
12723  __ret; \
12724 })
12725 #else
12726 #define vld4q_u32(__p0) __extension__ ({ \
12727  uint32x4x4_t __ret; \
12728  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
12729  \
12730  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12731  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12732  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12733  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12734  __ret; \
12735 })
12736 #endif
12737 
12738 #ifdef __LITTLE_ENDIAN__
12739 #define vld4q_u16(__p0) __extension__ ({ \
12740  uint16x8x4_t __ret; \
12741  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
12742  __ret; \
12743 })
12744 #else
12745 #define vld4q_u16(__p0) __extension__ ({ \
12746  uint16x8x4_t __ret; \
12747  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
12748  \
12749  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12750  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12751  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12752  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12753  __ret; \
12754 })
12755 #endif
12756 
12757 #ifdef __LITTLE_ENDIAN__
12758 #define vld4q_s8(__p0) __extension__ ({ \
12759  int8x16x4_t __ret; \
12760  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
12761  __ret; \
12762 })
12763 #else
12764 #define vld4q_s8(__p0) __extension__ ({ \
12765  int8x16x4_t __ret; \
12766  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
12767  \
12768  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12769  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12770  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12771  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
12772  __ret; \
12773 })
12774 #endif
12775 
12776 #ifdef __LITTLE_ENDIAN__
12777 #define vld4q_f32(__p0) __extension__ ({ \
12778  float32x4x4_t __ret; \
12779  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
12780  __ret; \
12781 })
12782 #else
12783 #define vld4q_f32(__p0) __extension__ ({ \
12784  float32x4x4_t __ret; \
12785  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
12786  \
12787  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12788  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12789  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12790  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12791  __ret; \
12792 })
12793 #endif
12794 
12795 #ifdef __LITTLE_ENDIAN__
12796 #define vld4q_s32(__p0) __extension__ ({ \
12797  int32x4x4_t __ret; \
12798  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
12799  __ret; \
12800 })
12801 #else
12802 #define vld4q_s32(__p0) __extension__ ({ \
12803  int32x4x4_t __ret; \
12804  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
12805  \
12806  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12807  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12808  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12809  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12810  __ret; \
12811 })
12812 #endif
12813 
12814 #ifdef __LITTLE_ENDIAN__
12815 #define vld4q_s16(__p0) __extension__ ({ \
12816  int16x8x4_t __ret; \
12817  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
12818  __ret; \
12819 })
12820 #else
12821 #define vld4q_s16(__p0) __extension__ ({ \
12822  int16x8x4_t __ret; \
12823  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
12824  \
12825  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12826  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12827  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12828  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12829  __ret; \
12830 })
12831 #endif
12832 
12833 #ifdef __LITTLE_ENDIAN__
12834 #define vld4_u8(__p0) __extension__ ({ \
12835  uint8x8x4_t __ret; \
12836  __builtin_neon_vld4_v(&__ret, __p0, 16); \
12837  __ret; \
12838 })
12839 #else
12840 #define vld4_u8(__p0) __extension__ ({ \
12841  uint8x8x4_t __ret; \
12842  __builtin_neon_vld4_v(&__ret, __p0, 16); \
12843  \
12844  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12845  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12846  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12847  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12848  __ret; \
12849 })
12850 #endif
12851 
12852 #ifdef __LITTLE_ENDIAN__
12853 #define vld4_u32(__p0) __extension__ ({ \
12854  uint32x2x4_t __ret; \
12855  __builtin_neon_vld4_v(&__ret, __p0, 18); \
12856  __ret; \
12857 })
12858 #else
12859 #define vld4_u32(__p0) __extension__ ({ \
12860  uint32x2x4_t __ret; \
12861  __builtin_neon_vld4_v(&__ret, __p0, 18); \
12862  \
12863  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12864  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12865  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12866  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12867  __ret; \
12868 })
12869 #endif
12870 
12871 #define vld4_u64(__p0) __extension__ ({ \
12872  uint64x1x4_t __ret; \
12873  __builtin_neon_vld4_v(&__ret, __p0, 19); \
12874  __ret; \
12875 })
12876 #ifdef __LITTLE_ENDIAN__
12877 #define vld4_u16(__p0) __extension__ ({ \
12878  uint16x4x4_t __ret; \
12879  __builtin_neon_vld4_v(&__ret, __p0, 17); \
12880  __ret; \
12881 })
12882 #else
12883 #define vld4_u16(__p0) __extension__ ({ \
12884  uint16x4x4_t __ret; \
12885  __builtin_neon_vld4_v(&__ret, __p0, 17); \
12886  \
12887  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12888  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12889  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12890  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12891  __ret; \
12892 })
12893 #endif
12894 
12895 #ifdef __LITTLE_ENDIAN__
12896 #define vld4_s8(__p0) __extension__ ({ \
12897  int8x8x4_t __ret; \
12898  __builtin_neon_vld4_v(&__ret, __p0, 0); \
12899  __ret; \
12900 })
12901 #else
12902 #define vld4_s8(__p0) __extension__ ({ \
12903  int8x8x4_t __ret; \
12904  __builtin_neon_vld4_v(&__ret, __p0, 0); \
12905  \
12906  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12907  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12908  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12909  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12910  __ret; \
12911 })
12912 #endif
12913 
12914 #ifdef __LITTLE_ENDIAN__
12915 #define vld4_f32(__p0) __extension__ ({ \
12916  float32x2x4_t __ret; \
12917  __builtin_neon_vld4_v(&__ret, __p0, 9); \
12918  __ret; \
12919 })
12920 #else
12921 #define vld4_f32(__p0) __extension__ ({ \
12922  float32x2x4_t __ret; \
12923  __builtin_neon_vld4_v(&__ret, __p0, 9); \
12924  \
12925  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12926  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12927  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12928  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12929  __ret; \
12930 })
12931 #endif
12932 
12933 #ifdef __LITTLE_ENDIAN__
12934 #define vld4_s32(__p0) __extension__ ({ \
12935  int32x2x4_t __ret; \
12936  __builtin_neon_vld4_v(&__ret, __p0, 2); \
12937  __ret; \
12938 })
12939 #else
12940 #define vld4_s32(__p0) __extension__ ({ \
12941  int32x2x4_t __ret; \
12942  __builtin_neon_vld4_v(&__ret, __p0, 2); \
12943  \
12944  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
12945  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
12946  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
12947  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
12948  __ret; \
12949 })
12950 #endif
12951 
12952 #define vld4_s64(__p0) __extension__ ({ \
12953  int64x1x4_t __ret; \
12954  __builtin_neon_vld4_v(&__ret, __p0, 3); \
12955  __ret; \
12956 })
12957 #ifdef __LITTLE_ENDIAN__
12958 #define vld4_s16(__p0) __extension__ ({ \
12959  int16x4x4_t __ret; \
12960  __builtin_neon_vld4_v(&__ret, __p0, 1); \
12961  __ret; \
12962 })
12963 #else
12964 #define vld4_s16(__p0) __extension__ ({ \
12965  int16x4x4_t __ret; \
12966  __builtin_neon_vld4_v(&__ret, __p0, 1); \
12967  \
12968  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
12969  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
12970  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
12971  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
12972  __ret; \
12973 })
12974 #endif
12975 
12976 #ifdef __LITTLE_ENDIAN__
12977 #define vld4_dup_p8(__p0) __extension__ ({ \
12978  poly8x8x4_t __ret; \
12979  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
12980  __ret; \
12981 })
12982 #else
12983 #define vld4_dup_p8(__p0) __extension__ ({ \
12984  poly8x8x4_t __ret; \
12985  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
12986  \
12987  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
12988  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
12989  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
12990  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
12991  __ret; \
12992 })
12993 #endif
12994 
12995 #ifdef __LITTLE_ENDIAN__
12996 #define vld4_dup_p16(__p0) __extension__ ({ \
12997  poly16x4x4_t __ret; \
12998  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
12999  __ret; \
13000 })
13001 #else
13002 #define vld4_dup_p16(__p0) __extension__ ({ \
13003  poly16x4x4_t __ret; \
13004  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
13005  \
13006  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13007  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13008  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13009  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13010  __ret; \
13011 })
13012 #endif
13013 
13014 #ifdef __LITTLE_ENDIAN__
13015 #define vld4q_dup_p8(__p0) __extension__ ({ \
13016  poly8x16x4_t __ret; \
13017  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
13018  __ret; \
13019 })
13020 #else
13021 #define vld4q_dup_p8(__p0) __extension__ ({ \
13022  poly8x16x4_t __ret; \
13023  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
13024  \
13025  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13026  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13027  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13028  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13029  __ret; \
13030 })
13031 #endif
13032 
13033 #ifdef __LITTLE_ENDIAN__
13034 #define vld4q_dup_p16(__p0) __extension__ ({ \
13035  poly16x8x4_t __ret; \
13036  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
13037  __ret; \
13038 })
13039 #else
13040 #define vld4q_dup_p16(__p0) __extension__ ({ \
13041  poly16x8x4_t __ret; \
13042  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
13043  \
13044  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13045  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13046  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13047  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13048  __ret; \
13049 })
13050 #endif
13051 
13052 #ifdef __LITTLE_ENDIAN__
13053 #define vld4q_dup_u8(__p0) __extension__ ({ \
13054  uint8x16x4_t __ret; \
13055  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
13056  __ret; \
13057 })
13058 #else
13059 #define vld4q_dup_u8(__p0) __extension__ ({ \
13060  uint8x16x4_t __ret; \
13061  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
13062  \
13063  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13064  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13065  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13066  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13067  __ret; \
13068 })
13069 #endif
13070 
13071 #ifdef __LITTLE_ENDIAN__
13072 #define vld4q_dup_u32(__p0) __extension__ ({ \
13073  uint32x4x4_t __ret; \
13074  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
13075  __ret; \
13076 })
13077 #else
13078 #define vld4q_dup_u32(__p0) __extension__ ({ \
13079  uint32x4x4_t __ret; \
13080  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
13081  \
13082  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13083  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13084  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13085  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13086  __ret; \
13087 })
13088 #endif
13089 
13090 #ifdef __LITTLE_ENDIAN__
13091 #define vld4q_dup_u64(__p0) __extension__ ({ \
13092  uint64x2x4_t __ret; \
13093  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
13094  __ret; \
13095 })
13096 #else
13097 #define vld4q_dup_u64(__p0) __extension__ ({ \
13098  uint64x2x4_t __ret; \
13099  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
13100  \
13101  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13102  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13103  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13104  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13105  __ret; \
13106 })
13107 #endif
13108 
13109 #ifdef __LITTLE_ENDIAN__
13110 #define vld4q_dup_u16(__p0) __extension__ ({ \
13111  uint16x8x4_t __ret; \
13112  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
13113  __ret; \
13114 })
13115 #else
13116 #define vld4q_dup_u16(__p0) __extension__ ({ \
13117  uint16x8x4_t __ret; \
13118  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
13119  \
13120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13123  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13124  __ret; \
13125 })
13126 #endif
13127 
13128 #ifdef __LITTLE_ENDIAN__
13129 #define vld4q_dup_s8(__p0) __extension__ ({ \
13130  int8x16x4_t __ret; \
13131  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
13132  __ret; \
13133 })
13134 #else
13135 #define vld4q_dup_s8(__p0) __extension__ ({ \
13136  int8x16x4_t __ret; \
13137  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
13138  \
13139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13141  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13142  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
13143  __ret; \
13144 })
13145 #endif
13146 
13147 #ifdef __LITTLE_ENDIAN__
13148 #define vld4q_dup_f32(__p0) __extension__ ({ \
13149  float32x4x4_t __ret; \
13150  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
13151  __ret; \
13152 })
13153 #else
13154 #define vld4q_dup_f32(__p0) __extension__ ({ \
13155  float32x4x4_t __ret; \
13156  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
13157  \
13158  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13159  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13160  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13161  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13162  __ret; \
13163 })
13164 #endif
13165 
13166 #ifdef __LITTLE_ENDIAN__
13167 #define vld4q_dup_s32(__p0) __extension__ ({ \
13168  int32x4x4_t __ret; \
13169  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
13170  __ret; \
13171 })
13172 #else
13173 #define vld4q_dup_s32(__p0) __extension__ ({ \
13174  int32x4x4_t __ret; \
13175  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
13176  \
13177  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13178  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13179  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13180  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13181  __ret; \
13182 })
13183 #endif
13184 
13185 #ifdef __LITTLE_ENDIAN__
13186 #define vld4q_dup_s64(__p0) __extension__ ({ \
13187  int64x2x4_t __ret; \
13188  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
13189  __ret; \
13190 })
13191 #else
13192 #define vld4q_dup_s64(__p0) __extension__ ({ \
13193  int64x2x4_t __ret; \
13194  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
13195  \
13196  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13197  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13198  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13199  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13200  __ret; \
13201 })
13202 #endif
13203 
13204 #ifdef __LITTLE_ENDIAN__
13205 #define vld4q_dup_s16(__p0) __extension__ ({ \
13206  int16x8x4_t __ret; \
13207  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
13208  __ret; \
13209 })
13210 #else
13211 #define vld4q_dup_s16(__p0) __extension__ ({ \
13212  int16x8x4_t __ret; \
13213  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
13214  \
13215  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13216  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13217  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13218  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13219  __ret; \
13220 })
13221 #endif
13222 
13223 #ifdef __LITTLE_ENDIAN__
13224 #define vld4_dup_u8(__p0) __extension__ ({ \
13225  uint8x8x4_t __ret; \
13226  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
13227  __ret; \
13228 })
13229 #else
13230 #define vld4_dup_u8(__p0) __extension__ ({ \
13231  uint8x8x4_t __ret; \
13232  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
13233  \
13234  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13235  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13236  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13237  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13238  __ret; \
13239 })
13240 #endif
13241 
13242 #ifdef __LITTLE_ENDIAN__
13243 #define vld4_dup_u32(__p0) __extension__ ({ \
13244  uint32x2x4_t __ret; \
13245  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
13246  __ret; \
13247 })
13248 #else
13249 #define vld4_dup_u32(__p0) __extension__ ({ \
13250  uint32x2x4_t __ret; \
13251  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
13252  \
13253  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13254  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13255  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13256  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13257  __ret; \
13258 })
13259 #endif
13260 
13261 #define vld4_dup_u64(__p0) __extension__ ({ \
13262  uint64x1x4_t __ret; \
13263  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
13264  __ret; \
13265 })
13266 #ifdef __LITTLE_ENDIAN__
13267 #define vld4_dup_u16(__p0) __extension__ ({ \
13268  uint16x4x4_t __ret; \
13269  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
13270  __ret; \
13271 })
13272 #else
13273 #define vld4_dup_u16(__p0) __extension__ ({ \
13274  uint16x4x4_t __ret; \
13275  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
13276  \
13277  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13278  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13279  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13280  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13281  __ret; \
13282 })
13283 #endif
13284 
13285 #ifdef __LITTLE_ENDIAN__
13286 #define vld4_dup_s8(__p0) __extension__ ({ \
13287  int8x8x4_t __ret; \
13288  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
13289  __ret; \
13290 })
13291 #else
13292 #define vld4_dup_s8(__p0) __extension__ ({ \
13293  int8x8x4_t __ret; \
13294  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
13295  \
13296  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13297  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13298  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13299  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13300  __ret; \
13301 })
13302 #endif
13303 
13304 #ifdef __LITTLE_ENDIAN__
13305 #define vld4_dup_f32(__p0) __extension__ ({ \
13306  float32x2x4_t __ret; \
13307  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
13308  __ret; \
13309 })
13310 #else
13311 #define vld4_dup_f32(__p0) __extension__ ({ \
13312  float32x2x4_t __ret; \
13313  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
13314  \
13315  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13316  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13317  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13318  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13319  __ret; \
13320 })
13321 #endif
13322 
13323 #ifdef __LITTLE_ENDIAN__
13324 #define vld4_dup_s32(__p0) __extension__ ({ \
13325  int32x2x4_t __ret; \
13326  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
13327  __ret; \
13328 })
13329 #else
13330 #define vld4_dup_s32(__p0) __extension__ ({ \
13331  int32x2x4_t __ret; \
13332  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
13333  \
13334  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13335  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13336  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13337  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13338  __ret; \
13339 })
13340 #endif
13341 
13342 #define vld4_dup_s64(__p0) __extension__ ({ \
13343  int64x1x4_t __ret; \
13344  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
13345  __ret; \
13346 })
13347 #ifdef __LITTLE_ENDIAN__
13348 #define vld4_dup_s16(__p0) __extension__ ({ \
13349  int16x4x4_t __ret; \
13350  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
13351  __ret; \
13352 })
13353 #else
13354 #define vld4_dup_s16(__p0) __extension__ ({ \
13355  int16x4x4_t __ret; \
13356  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
13357  \
13358  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13359  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13360  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13361  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13362  __ret; \
13363 })
13364 #endif
13365 
13366 #ifdef __LITTLE_ENDIAN__
13367 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
13368  poly8x8x4_t __s1 = __p1; \
13369  poly8x8x4_t __ret; \
13370  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
13371  __ret; \
13372 })
13373 #else
13374 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
13375  poly8x8x4_t __s1 = __p1; \
13376  poly8x8x4_t __rev1; \
13377  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13378  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13379  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13380  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13381  poly8x8x4_t __ret; \
13382  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
13383  \
13384  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13385  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13386  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13387  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13388  __ret; \
13389 })
13390 #endif
13391 
13392 #ifdef __LITTLE_ENDIAN__
13393 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13394  poly16x4x4_t __s1 = __p1; \
13395  poly16x4x4_t __ret; \
13396  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
13397  __ret; \
13398 })
13399 #else
13400 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13401  poly16x4x4_t __s1 = __p1; \
13402  poly16x4x4_t __rev1; \
13403  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13404  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13405  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13406  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13407  poly16x4x4_t __ret; \
13408  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
13409  \
13410  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13411  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13412  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13413  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13414  __ret; \
13415 })
13416 #endif
13417 
13418 #ifdef __LITTLE_ENDIAN__
13419 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13420  poly16x8x4_t __s1 = __p1; \
13421  poly16x8x4_t __ret; \
13422  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
13423  __ret; \
13424 })
13425 #else
13426 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
13427  poly16x8x4_t __s1 = __p1; \
13428  poly16x8x4_t __rev1; \
13429  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13430  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13431  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13432  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13433  poly16x8x4_t __ret; \
13434  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
13435  \
13436  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13437  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13438  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13439  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13440  __ret; \
13441 })
13442 #endif
13443 
13444 #ifdef __LITTLE_ENDIAN__
13445 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13446  uint32x4x4_t __s1 = __p1; \
13447  uint32x4x4_t __ret; \
13448  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
13449  __ret; \
13450 })
13451 #else
13452 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13453  uint32x4x4_t __s1 = __p1; \
13454  uint32x4x4_t __rev1; \
13455  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13456  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13457  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13458  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13459  uint32x4x4_t __ret; \
13460  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
13461  \
13462  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13463  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13464  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13465  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13466  __ret; \
13467 })
13468 #endif
13469 
13470 #ifdef __LITTLE_ENDIAN__
13471 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13472  uint16x8x4_t __s1 = __p1; \
13473  uint16x8x4_t __ret; \
13474  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
13475  __ret; \
13476 })
13477 #else
13478 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13479  uint16x8x4_t __s1 = __p1; \
13480  uint16x8x4_t __rev1; \
13481  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13482  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13483  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13484  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13485  uint16x8x4_t __ret; \
13486  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
13487  \
13488  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13489  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13490  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13491  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13492  __ret; \
13493 })
13494 #endif
13495 
13496 #ifdef __LITTLE_ENDIAN__
13497 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13498  float32x4x4_t __s1 = __p1; \
13499  float32x4x4_t __ret; \
13500  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 41); \
13501  __ret; \
13502 })
13503 #else
13504 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13505  float32x4x4_t __s1 = __p1; \
13506  float32x4x4_t __rev1; \
13507  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13508  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13509  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13510  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13511  float32x4x4_t __ret; \
13512  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 41); \
13513  \
13514  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13515  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13516  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13517  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13518  __ret; \
13519 })
13520 #endif
13521 
13522 #ifdef __LITTLE_ENDIAN__
13523 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13524  int32x4x4_t __s1 = __p1; \
13525  int32x4x4_t __ret; \
13526  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 34); \
13527  __ret; \
13528 })
13529 #else
13530 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13531  int32x4x4_t __s1 = __p1; \
13532  int32x4x4_t __rev1; \
13533  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13534  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13535  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13536  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13537  int32x4x4_t __ret; \
13538  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 34); \
13539  \
13540  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13541  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13542  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13543  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13544  __ret; \
13545 })
13546 #endif
13547 
13548 #ifdef __LITTLE_ENDIAN__
13549 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13550  int16x8x4_t __s1 = __p1; \
13551  int16x8x4_t __ret; \
13552  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 33); \
13553  __ret; \
13554 })
13555 #else
13556 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13557  int16x8x4_t __s1 = __p1; \
13558  int16x8x4_t __rev1; \
13559  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13560  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13561  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13562  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13563  int16x8x4_t __ret; \
13564  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 33); \
13565  \
13566  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13567  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13568  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13569  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13570  __ret; \
13571 })
13572 #endif
13573 
13574 #ifdef __LITTLE_ENDIAN__
13575 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
13576  uint8x8x4_t __s1 = __p1; \
13577  uint8x8x4_t __ret; \
13578  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
13579  __ret; \
13580 })
13581 #else
13582 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
13583  uint8x8x4_t __s1 = __p1; \
13584  uint8x8x4_t __rev1; \
13585  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13586  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13587  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13588  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13589  uint8x8x4_t __ret; \
13590  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
13591  \
13592  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13593  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13594  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13595  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13596  __ret; \
13597 })
13598 #endif
13599 
13600 #ifdef __LITTLE_ENDIAN__
13601 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13602  uint32x2x4_t __s1 = __p1; \
13603  uint32x2x4_t __ret; \
13604  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
13605  __ret; \
13606 })
13607 #else
13608 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
13609  uint32x2x4_t __s1 = __p1; \
13610  uint32x2x4_t __rev1; \
13611  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13612  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13613  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13614  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13615  uint32x2x4_t __ret; \
13616  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
13617  \
13618  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13619  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13620  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13621  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13622  __ret; \
13623 })
13624 #endif
13625 
13626 #ifdef __LITTLE_ENDIAN__
13627 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13628  uint16x4x4_t __s1 = __p1; \
13629  uint16x4x4_t __ret; \
13630  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
13631  __ret; \
13632 })
13633 #else
13634 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
13635  uint16x4x4_t __s1 = __p1; \
13636  uint16x4x4_t __rev1; \
13637  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13638  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13639  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13640  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13641  uint16x4x4_t __ret; \
13642  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
13643  \
13644  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13645  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13646  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13647  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13648  __ret; \
13649 })
13650 #endif
13651 
13652 #ifdef __LITTLE_ENDIAN__
13653 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
13654  int8x8x4_t __s1 = __p1; \
13655  int8x8x4_t __ret; \
13656  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
13657  __ret; \
13658 })
13659 #else
13660 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
13661  int8x8x4_t __s1 = __p1; \
13662  int8x8x4_t __rev1; \
13663  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13664  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13665  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13666  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13667  int8x8x4_t __ret; \
13668  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
13669  \
13670  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
13671  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
13672  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
13673  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
13674  __ret; \
13675 })
13676 #endif
13677 
13678 #ifdef __LITTLE_ENDIAN__
13679 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13680  float32x2x4_t __s1 = __p1; \
13681  float32x2x4_t __ret; \
13682  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 9); \
13683  __ret; \
13684 })
13685 #else
13686 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
13687  float32x2x4_t __s1 = __p1; \
13688  float32x2x4_t __rev1; \
13689  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13690  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13691  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13692  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13693  float32x2x4_t __ret; \
13694  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 9); \
13695  \
13696  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13697  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13698  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13699  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13700  __ret; \
13701 })
13702 #endif
13703 
13704 #ifdef __LITTLE_ENDIAN__
13705 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13706  int32x2x4_t __s1 = __p1; \
13707  int32x2x4_t __ret; \
13708  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 2); \
13709  __ret; \
13710 })
13711 #else
13712 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
13713  int32x2x4_t __s1 = __p1; \
13714  int32x2x4_t __rev1; \
13715  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
13716  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
13717  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
13718  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
13719  int32x2x4_t __ret; \
13720  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 2); \
13721  \
13722  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
13723  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
13724  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
13725  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
13726  __ret; \
13727 })
13728 #endif
13729 
13730 #ifdef __LITTLE_ENDIAN__
13731 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13732  int16x4x4_t __s1 = __p1; \
13733  int16x4x4_t __ret; \
13734  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 1); \
13735  __ret; \
13736 })
13737 #else
13738 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
13739  int16x4x4_t __s1 = __p1; \
13740  int16x4x4_t __rev1; \
13741  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
13742  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
13743  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
13744  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
13745  int16x4x4_t __ret; \
13746  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 1); \
13747  \
13748  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
13749  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
13750  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
13751  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
13752  __ret; \
13753 })
13754 #endif
13755 
13756 #ifdef __LITTLE_ENDIAN__
13757 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13758  uint8x16_t __ret;
13759  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
13760  return __ret;
13761 }
13762 #else
13763 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13764  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13765  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13766  uint8x16_t __ret;
13767  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
13768  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13769  return __ret;
13770 }
13771 #endif
13772 
13773 #ifdef __LITTLE_ENDIAN__
13774 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13775  uint32x4_t __ret;
13776  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
13777  return __ret;
13778 }
13779 #else
13780 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
13781  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13782  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13783  uint32x4_t __ret;
13784  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
13785  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13786  return __ret;
13787 }
13788 #endif
13789 
13790 #ifdef __LITTLE_ENDIAN__
13791 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13792  uint16x8_t __ret;
13793  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
13794  return __ret;
13795 }
13796 #else
13797 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
13798  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13799  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13800  uint16x8_t __ret;
13801  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
13802  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13803  return __ret;
13804 }
13805 #endif
13806 
13807 #ifdef __LITTLE_ENDIAN__
13808 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
13809  int8x16_t __ret;
13810  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
13811  return __ret;
13812 }
13813 #else
13814 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
13815  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13816  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13817  int8x16_t __ret;
13818  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
13819  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13820  return __ret;
13821 }
13822 #endif
13823 
13824 #ifdef __LITTLE_ENDIAN__
13825 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
13826  float32x4_t __ret;
13827  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
13828  return __ret;
13829 }
13830 #else
13831 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
13832  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13833  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13834  float32x4_t __ret;
13835  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
13836  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13837  return __ret;
13838 }
13839 #endif
13840 
13841 #ifdef __LITTLE_ENDIAN__
13842 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
13843  int32x4_t __ret;
13844  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
13845  return __ret;
13846 }
13847 #else
13848 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
13849  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13850  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13851  int32x4_t __ret;
13852  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
13853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13854  return __ret;
13855 }
13856 #endif
13857 
13858 #ifdef __LITTLE_ENDIAN__
13859 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
13860  int16x8_t __ret;
13861  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
13862  return __ret;
13863 }
13864 #else
13865 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
13866  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13867  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13868  int16x8_t __ret;
13869  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
13870  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13871  return __ret;
13872 }
13873 #endif
13874 
13875 #ifdef __LITTLE_ENDIAN__
13876 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
13877  uint8x8_t __ret;
13878  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
13879  return __ret;
13880 }
13881 #else
13882 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
13883  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13884  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13885  uint8x8_t __ret;
13886  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
13887  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13888  return __ret;
13889 }
13890 #endif
13891 
13892 #ifdef __LITTLE_ENDIAN__
13893 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
13894  uint32x2_t __ret;
13895  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
13896  return __ret;
13897 }
13898 #else
13899 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
13900  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13901  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13902  uint32x2_t __ret;
13903  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
13904  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13905  return __ret;
13906 }
13907 #endif
13908 
13909 #ifdef __LITTLE_ENDIAN__
13910 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
13911  uint16x4_t __ret;
13912  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
13913  return __ret;
13914 }
13915 #else
13916 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
13917  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13918  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13919  uint16x4_t __ret;
13920  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
13921  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13922  return __ret;
13923 }
13924 #endif
13925 
13926 #ifdef __LITTLE_ENDIAN__
13927 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
13928  int8x8_t __ret;
13929  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
13930  return __ret;
13931 }
13932 #else
13933 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
13934  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13935  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13936  int8x8_t __ret;
13937  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
13938  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13939  return __ret;
13940 }
13941 #endif
13942 
13943 #ifdef __LITTLE_ENDIAN__
13944 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
13945  float32x2_t __ret;
13946  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
13947  return __ret;
13948 }
13949 #else
13950 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
13951  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13952  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13953  float32x2_t __ret;
13954  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
13955  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13956  return __ret;
13957 }
13958 #endif
13959 
13960 #ifdef __LITTLE_ENDIAN__
13961 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
13962  int32x2_t __ret;
13963  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
13964  return __ret;
13965 }
13966 #else
13967 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
13968  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13969  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13970  int32x2_t __ret;
13971  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
13972  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13973  return __ret;
13974 }
13975 #endif
13976 
13977 #ifdef __LITTLE_ENDIAN__
13978 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
13979  int16x4_t __ret;
13980  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
13981  return __ret;
13982 }
13983 #else
13984 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
13985  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13986  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13987  int16x4_t __ret;
13988  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
13989  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13990  return __ret;
13991 }
13992 #endif
13993 
13994 #ifdef __LITTLE_ENDIAN__
13995 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
13996  uint8x16_t __ret;
13997  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
13998  return __ret;
13999 }
14000 #else
14001 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14002  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14003  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14004  uint8x16_t __ret;
14005  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
14006  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14007  return __ret;
14008 }
14009 #endif
14010 
14011 #ifdef __LITTLE_ENDIAN__
14012 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14013  uint32x4_t __ret;
14014  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
14015  return __ret;
14016 }
14017 #else
14018 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14019  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14020  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14021  uint32x4_t __ret;
14022  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
14023  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14024  return __ret;
14025 }
14026 #endif
14027 
14028 #ifdef __LITTLE_ENDIAN__
14029 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14030  uint16x8_t __ret;
14031  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
14032  return __ret;
14033 }
14034 #else
14035 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14036  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14037  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14038  uint16x8_t __ret;
14039  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
14040  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14041  return __ret;
14042 }
14043 #endif
14044 
14045 #ifdef __LITTLE_ENDIAN__
14046 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
14047  int8x16_t __ret;
14048  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
14049  return __ret;
14050 }
14051 #else
14052 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
14053  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14054  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14055  int8x16_t __ret;
14056  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
14057  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14058  return __ret;
14059 }
14060 #endif
14061 
14062 #ifdef __LITTLE_ENDIAN__
14063 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
14064  float32x4_t __ret;
14065  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
14066  return __ret;
14067 }
14068 #else
14069 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
14070  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14071  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14072  float32x4_t __ret;
14073  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
14074  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14075  return __ret;
14076 }
14077 #endif
14078 
14079 #ifdef __LITTLE_ENDIAN__
14080 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
14081  int32x4_t __ret;
14082  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
14083  return __ret;
14084 }
14085 #else
14086 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
14087  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14088  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14089  int32x4_t __ret;
14090  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
14091  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14092  return __ret;
14093 }
14094 #endif
14095 
14096 #ifdef __LITTLE_ENDIAN__
14097 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
14098  int16x8_t __ret;
14099  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
14100  return __ret;
14101 }
14102 #else
14103 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
14104  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14105  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14106  int16x8_t __ret;
14107  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
14108  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14109  return __ret;
14110 }
14111 #endif
14112 
14113 #ifdef __LITTLE_ENDIAN__
14114 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
14115  uint8x8_t __ret;
14116  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
14117  return __ret;
14118 }
14119 #else
14120 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
14121  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14122  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14123  uint8x8_t __ret;
14124  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
14125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14126  return __ret;
14127 }
14128 #endif
14129 
14130 #ifdef __LITTLE_ENDIAN__
14131 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
14132  uint32x2_t __ret;
14133  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
14134  return __ret;
14135 }
14136 #else
14137 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
14138  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14139  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14140  uint32x2_t __ret;
14141  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
14142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14143  return __ret;
14144 }
14145 #endif
14146 
14147 #ifdef __LITTLE_ENDIAN__
14148 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
14149  uint16x4_t __ret;
14150  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
14151  return __ret;
14152 }
14153 #else
14154 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
14155  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14156  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14157  uint16x4_t __ret;
14158  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
14159  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14160  return __ret;
14161 }
14162 #endif
14163 
14164 #ifdef __LITTLE_ENDIAN__
14165 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
14166  int8x8_t __ret;
14167  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
14168  return __ret;
14169 }
14170 #else
14171 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
14172  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14173  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14174  int8x8_t __ret;
14175  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
14176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14177  return __ret;
14178 }
14179 #endif
14180 
14181 #ifdef __LITTLE_ENDIAN__
14182 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
14183  float32x2_t __ret;
14184  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
14185  return __ret;
14186 }
14187 #else
14188 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
14189  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14190  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14191  float32x2_t __ret;
14192  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
14193  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14194  return __ret;
14195 }
14196 #endif
14197 
14198 #ifdef __LITTLE_ENDIAN__
14199 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
14200  int32x2_t __ret;
14201  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
14202  return __ret;
14203 }
14204 #else
14205 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
14206  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14207  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14208  int32x2_t __ret;
14209  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
14210  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14211  return __ret;
14212 }
14213 #endif
14214 
14215 #ifdef __LITTLE_ENDIAN__
14216 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
14217  int16x4_t __ret;
14218  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
14219  return __ret;
14220 }
14221 #else
14222 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
14223  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14224  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14225  int16x4_t __ret;
14226  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
14227  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14228  return __ret;
14229 }
14230 #endif
14231 
14232 #ifdef __LITTLE_ENDIAN__
14233 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14234  uint8x16_t __ret;
14235  __ret = __p0 + __p1 * __p2;
14236  return __ret;
14237 }
14238 #else
14239 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14240  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14241  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14242  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14243  uint8x16_t __ret;
14244  __ret = __rev0 + __rev1 * __rev2;
14245  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14246  return __ret;
14247 }
14248 #endif
14249 
14250 #ifdef __LITTLE_ENDIAN__
14251 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14252  uint32x4_t __ret;
14253  __ret = __p0 + __p1 * __p2;
14254  return __ret;
14255 }
14256 #else
14257 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14258  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14259  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14260  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14261  uint32x4_t __ret;
14262  __ret = __rev0 + __rev1 * __rev2;
14263  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14264  return __ret;
14265 }
14266 #endif
14267 
14268 #ifdef __LITTLE_ENDIAN__
14269 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14270  uint16x8_t __ret;
14271  __ret = __p0 + __p1 * __p2;
14272  return __ret;
14273 }
14274 #else
14275 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14276  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14277  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14278  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14279  uint16x8_t __ret;
14280  __ret = __rev0 + __rev1 * __rev2;
14281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14282  return __ret;
14283 }
14284 #endif
14285 
14286 #ifdef __LITTLE_ENDIAN__
14287 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14288  int8x16_t __ret;
14289  __ret = __p0 + __p1 * __p2;
14290  return __ret;
14291 }
14292 #else
14293 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14294  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14295  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14296  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14297  int8x16_t __ret;
14298  __ret = __rev0 + __rev1 * __rev2;
14299  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14300  return __ret;
14301 }
14302 #endif
14303 
14304 #ifdef __LITTLE_ENDIAN__
14305 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14306  float32x4_t __ret;
14307  __ret = __p0 + __p1 * __p2;
14308  return __ret;
14309 }
14310 #else
14311 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14312  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14313  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14314  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14315  float32x4_t __ret;
14316  __ret = __rev0 + __rev1 * __rev2;
14317  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14318  return __ret;
14319 }
14320 #endif
14321 
14322 #ifdef __LITTLE_ENDIAN__
14323 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14324  int32x4_t __ret;
14325  __ret = __p0 + __p1 * __p2;
14326  return __ret;
14327 }
14328 #else
14329 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14330  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14331  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14332  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14333  int32x4_t __ret;
14334  __ret = __rev0 + __rev1 * __rev2;
14335  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14336  return __ret;
14337 }
14338 #endif
14339 
14340 #ifdef __LITTLE_ENDIAN__
14341 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14342  int16x8_t __ret;
14343  __ret = __p0 + __p1 * __p2;
14344  return __ret;
14345 }
14346 #else
14347 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
14348  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14349  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14350  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14351  int16x8_t __ret;
14352  __ret = __rev0 + __rev1 * __rev2;
14353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14354  return __ret;
14355 }
14356 #endif
14357 
14358 #ifdef __LITTLE_ENDIAN__
14359 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14360  uint8x8_t __ret;
14361  __ret = __p0 + __p1 * __p2;
14362  return __ret;
14363 }
14364 #else
14365 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
14366  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14367  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14368  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14369  uint8x8_t __ret;
14370  __ret = __rev0 + __rev1 * __rev2;
14371  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14372  return __ret;
14373 }
14374 #endif
14375 
14376 #ifdef __LITTLE_ENDIAN__
14377 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14378  uint32x2_t __ret;
14379  __ret = __p0 + __p1 * __p2;
14380  return __ret;
14381 }
14382 #else
14383 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
14384  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14385  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14386  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14387  uint32x2_t __ret;
14388  __ret = __rev0 + __rev1 * __rev2;
14389  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14390  return __ret;
14391 }
14392 #endif
14393 
14394 #ifdef __LITTLE_ENDIAN__
14395 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14396  uint16x4_t __ret;
14397  __ret = __p0 + __p1 * __p2;
14398  return __ret;
14399 }
14400 #else
14401 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
14402  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14403  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14404  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14405  uint16x4_t __ret;
14406  __ret = __rev0 + __rev1 * __rev2;
14407  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14408  return __ret;
14409 }
14410 #endif
14411 
14412 #ifdef __LITTLE_ENDIAN__
14413 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14414  int8x8_t __ret;
14415  __ret = __p0 + __p1 * __p2;
14416  return __ret;
14417 }
14418 #else
14419 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
14420  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14421  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14422  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14423  int8x8_t __ret;
14424  __ret = __rev0 + __rev1 * __rev2;
14425  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14426  return __ret;
14427 }
14428 #endif
14429 
14430 #ifdef __LITTLE_ENDIAN__
14431 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14432  float32x2_t __ret;
14433  __ret = __p0 + __p1 * __p2;
14434  return __ret;
14435 }
14436 #else
14437 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
14438  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14439  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14440  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14441  float32x2_t __ret;
14442  __ret = __rev0 + __rev1 * __rev2;
14443  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14444  return __ret;
14445 }
14446 #endif
14447 
14448 #ifdef __LITTLE_ENDIAN__
14449 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
14450  int32x2_t __ret;
14451  __ret = __p0 + __p1 * __p2;
14452  return __ret;
14453 }
14454 #else
14455 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
14456  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14457  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14458  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
14459  int32x2_t __ret;
14460  __ret = __rev0 + __rev1 * __rev2;
14461  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14462  return __ret;
14463 }
14464 #endif
14465 
14466 #ifdef __LITTLE_ENDIAN__
14467 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
14468  int16x4_t __ret;
14469  __ret = __p0 + __p1 * __p2;
14470  return __ret;
14471 }
14472 #else
14473 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
14474  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14475  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14476  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14477  int16x4_t __ret;
14478  __ret = __rev0 + __rev1 * __rev2;
14479  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14480  return __ret;
14481 }
14482 #endif
14483 
14484 #ifdef __LITTLE_ENDIAN__
14485 #define vmlaq_lane_u32(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
14486  uint32x4_t __s0_42 = __p0_42; \
14487  uint32x4_t __s1_42 = __p1_42; \
14488  uint32x2_t __s2_42 = __p2_42; \
14489  uint32x4_t __ret_42; \
14490  __ret_42 = __s0_42 + __s1_42 * splatq_lane_u32(__s2_42, __p3_42); \
14491  __ret_42; \
14492 })
14493 #else
14494 #define vmlaq_lane_u32(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
14495  uint32x4_t __s0_43 = __p0_43; \
14496  uint32x4_t __s1_43 = __p1_43; \
14497  uint32x2_t __s2_43 = __p2_43; \
14498  uint32x4_t __rev0_43; __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
14499  uint32x4_t __rev1_43; __rev1_43 = __builtin_shufflevector(__s1_43, __s1_43, 3, 2, 1, 0); \
14500  uint32x2_t __rev2_43; __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 1, 0); \
14501  uint32x4_t __ret_43; \
14502  __ret_43 = __rev0_43 + __rev1_43 * __noswap_splatq_lane_u32(__rev2_43, __p3_43); \
14503  __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
14504  __ret_43; \
14505 })
14506 #endif
14507 
14508 #ifdef __LITTLE_ENDIAN__
14509 #define vmlaq_lane_u16(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
14510  uint16x8_t __s0_44 = __p0_44; \
14511  uint16x8_t __s1_44 = __p1_44; \
14512  uint16x4_t __s2_44 = __p2_44; \
14513  uint16x8_t __ret_44; \
14514  __ret_44 = __s0_44 + __s1_44 * splatq_lane_u16(__s2_44, __p3_44); \
14515  __ret_44; \
14516 })
14517 #else
14518 #define vmlaq_lane_u16(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
14519  uint16x8_t __s0_45 = __p0_45; \
14520  uint16x8_t __s1_45 = __p1_45; \
14521  uint16x4_t __s2_45 = __p2_45; \
14522  uint16x8_t __rev0_45; __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 7, 6, 5, 4, 3, 2, 1, 0); \
14523  uint16x8_t __rev1_45; __rev1_45 = __builtin_shufflevector(__s1_45, __s1_45, 7, 6, 5, 4, 3, 2, 1, 0); \
14524  uint16x4_t __rev2_45; __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 3, 2, 1, 0); \
14525  uint16x8_t __ret_45; \
14526  __ret_45 = __rev0_45 + __rev1_45 * __noswap_splatq_lane_u16(__rev2_45, __p3_45); \
14527  __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 7, 6, 5, 4, 3, 2, 1, 0); \
14528  __ret_45; \
14529 })
14530 #endif
14531 
14532 #ifdef __LITTLE_ENDIAN__
14533 #define vmlaq_lane_f32(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
14534  float32x4_t __s0_46 = __p0_46; \
14535  float32x4_t __s1_46 = __p1_46; \
14536  float32x2_t __s2_46 = __p2_46; \
14537  float32x4_t __ret_46; \
14538  __ret_46 = __s0_46 + __s1_46 * splatq_lane_f32(__s2_46, __p3_46); \
14539  __ret_46; \
14540 })
14541 #else
14542 #define vmlaq_lane_f32(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
14543  float32x4_t __s0_47 = __p0_47; \
14544  float32x4_t __s1_47 = __p1_47; \
14545  float32x2_t __s2_47 = __p2_47; \
14546  float32x4_t __rev0_47; __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 3, 2, 1, 0); \
14547  float32x4_t __rev1_47; __rev1_47 = __builtin_shufflevector(__s1_47, __s1_47, 3, 2, 1, 0); \
14548  float32x2_t __rev2_47; __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 1, 0); \
14549  float32x4_t __ret_47; \
14550  __ret_47 = __rev0_47 + __rev1_47 * __noswap_splatq_lane_f32(__rev2_47, __p3_47); \
14551  __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 3, 2, 1, 0); \
14552  __ret_47; \
14553 })
14554 #endif
14555 
14556 #ifdef __LITTLE_ENDIAN__
14557 #define vmlaq_lane_s32(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
14558  int32x4_t __s0_48 = __p0_48; \
14559  int32x4_t __s1_48 = __p1_48; \
14560  int32x2_t __s2_48 = __p2_48; \
14561  int32x4_t __ret_48; \
14562  __ret_48 = __s0_48 + __s1_48 * splatq_lane_s32(__s2_48, __p3_48); \
14563  __ret_48; \
14564 })
14565 #else
14566 #define vmlaq_lane_s32(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
14567  int32x4_t __s0_49 = __p0_49; \
14568  int32x4_t __s1_49 = __p1_49; \
14569  int32x2_t __s2_49 = __p2_49; \
14570  int32x4_t __rev0_49; __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 3, 2, 1, 0); \
14571  int32x4_t __rev1_49; __rev1_49 = __builtin_shufflevector(__s1_49, __s1_49, 3, 2, 1, 0); \
14572  int32x2_t __rev2_49; __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 1, 0); \
14573  int32x4_t __ret_49; \
14574  __ret_49 = __rev0_49 + __rev1_49 * __noswap_splatq_lane_s32(__rev2_49, __p3_49); \
14575  __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 3, 2, 1, 0); \
14576  __ret_49; \
14577 })
14578 #endif
14579 
14580 #ifdef __LITTLE_ENDIAN__
14581 #define vmlaq_lane_s16(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
14582  int16x8_t __s0_50 = __p0_50; \
14583  int16x8_t __s1_50 = __p1_50; \
14584  int16x4_t __s2_50 = __p2_50; \
14585  int16x8_t __ret_50; \
14586  __ret_50 = __s0_50 + __s1_50 * splatq_lane_s16(__s2_50, __p3_50); \
14587  __ret_50; \
14588 })
14589 #else
14590 #define vmlaq_lane_s16(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
14591  int16x8_t __s0_51 = __p0_51; \
14592  int16x8_t __s1_51 = __p1_51; \
14593  int16x4_t __s2_51 = __p2_51; \
14594  int16x8_t __rev0_51; __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 7, 6, 5, 4, 3, 2, 1, 0); \
14595  int16x8_t __rev1_51; __rev1_51 = __builtin_shufflevector(__s1_51, __s1_51, 7, 6, 5, 4, 3, 2, 1, 0); \
14596  int16x4_t __rev2_51; __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
14597  int16x8_t __ret_51; \
14598  __ret_51 = __rev0_51 + __rev1_51 * __noswap_splatq_lane_s16(__rev2_51, __p3_51); \
14599  __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 7, 6, 5, 4, 3, 2, 1, 0); \
14600  __ret_51; \
14601 })
14602 #endif
14603 
14604 #ifdef __LITTLE_ENDIAN__
14605 #define vmla_lane_u32(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
14606  uint32x2_t __s0_52 = __p0_52; \
14607  uint32x2_t __s1_52 = __p1_52; \
14608  uint32x2_t __s2_52 = __p2_52; \
14609  uint32x2_t __ret_52; \
14610  __ret_52 = __s0_52 + __s1_52 * splat_lane_u32(__s2_52, __p3_52); \
14611  __ret_52; \
14612 })
14613 #else
14614 #define vmla_lane_u32(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
14615  uint32x2_t __s0_53 = __p0_53; \
14616  uint32x2_t __s1_53 = __p1_53; \
14617  uint32x2_t __s2_53 = __p2_53; \
14618  uint32x2_t __rev0_53; __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
14619  uint32x2_t __rev1_53; __rev1_53 = __builtin_shufflevector(__s1_53, __s1_53, 1, 0); \
14620  uint32x2_t __rev2_53; __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
14621  uint32x2_t __ret_53; \
14622  __ret_53 = __rev0_53 + __rev1_53 * __noswap_splat_lane_u32(__rev2_53, __p3_53); \
14623  __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
14624  __ret_53; \
14625 })
14626 #endif
14627 
14628 #ifdef __LITTLE_ENDIAN__
14629 #define vmla_lane_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
14630  uint16x4_t __s0_54 = __p0_54; \
14631  uint16x4_t __s1_54 = __p1_54; \
14632  uint16x4_t __s2_54 = __p2_54; \
14633  uint16x4_t __ret_54; \
14634  __ret_54 = __s0_54 + __s1_54 * splat_lane_u16(__s2_54, __p3_54); \
14635  __ret_54; \
14636 })
14637 #else
14638 #define vmla_lane_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
14639  uint16x4_t __s0_55 = __p0_55; \
14640  uint16x4_t __s1_55 = __p1_55; \
14641  uint16x4_t __s2_55 = __p2_55; \
14642  uint16x4_t __rev0_55; __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 3, 2, 1, 0); \
14643  uint16x4_t __rev1_55; __rev1_55 = __builtin_shufflevector(__s1_55, __s1_55, 3, 2, 1, 0); \
14644  uint16x4_t __rev2_55; __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 3, 2, 1, 0); \
14645  uint16x4_t __ret_55; \
14646  __ret_55 = __rev0_55 + __rev1_55 * __noswap_splat_lane_u16(__rev2_55, __p3_55); \
14647  __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 3, 2, 1, 0); \
14648  __ret_55; \
14649 })
14650 #endif
14651 
14652 #ifdef __LITTLE_ENDIAN__
14653 #define vmla_lane_f32(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
14654  float32x2_t __s0_56 = __p0_56; \
14655  float32x2_t __s1_56 = __p1_56; \
14656  float32x2_t __s2_56 = __p2_56; \
14657  float32x2_t __ret_56; \
14658  __ret_56 = __s0_56 + __s1_56 * splat_lane_f32(__s2_56, __p3_56); \
14659  __ret_56; \
14660 })
14661 #else
14662 #define vmla_lane_f32(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
14663  float32x2_t __s0_57 = __p0_57; \
14664  float32x2_t __s1_57 = __p1_57; \
14665  float32x2_t __s2_57 = __p2_57; \
14666  float32x2_t __rev0_57; __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 1, 0); \
14667  float32x2_t __rev1_57; __rev1_57 = __builtin_shufflevector(__s1_57, __s1_57, 1, 0); \
14668  float32x2_t __rev2_57; __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 1, 0); \
14669  float32x2_t __ret_57; \
14670  __ret_57 = __rev0_57 + __rev1_57 * __noswap_splat_lane_f32(__rev2_57, __p3_57); \
14671  __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 1, 0); \
14672  __ret_57; \
14673 })
14674 #endif
14675 
14676 #ifdef __LITTLE_ENDIAN__
14677 #define vmla_lane_s32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
14678  int32x2_t __s0_58 = __p0_58; \
14679  int32x2_t __s1_58 = __p1_58; \
14680  int32x2_t __s2_58 = __p2_58; \
14681  int32x2_t __ret_58; \
14682  __ret_58 = __s0_58 + __s1_58 * splat_lane_s32(__s2_58, __p3_58); \
14683  __ret_58; \
14684 })
14685 #else
14686 #define vmla_lane_s32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
14687  int32x2_t __s0_59 = __p0_59; \
14688  int32x2_t __s1_59 = __p1_59; \
14689  int32x2_t __s2_59 = __p2_59; \
14690  int32x2_t __rev0_59; __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 1, 0); \
14691  int32x2_t __rev1_59; __rev1_59 = __builtin_shufflevector(__s1_59, __s1_59, 1, 0); \
14692  int32x2_t __rev2_59; __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 1, 0); \
14693  int32x2_t __ret_59; \
14694  __ret_59 = __rev0_59 + __rev1_59 * __noswap_splat_lane_s32(__rev2_59, __p3_59); \
14695  __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 1, 0); \
14696  __ret_59; \
14697 })
14698 #endif
14699 
14700 #ifdef __LITTLE_ENDIAN__
14701 #define vmla_lane_s16(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
14702  int16x4_t __s0_60 = __p0_60; \
14703  int16x4_t __s1_60 = __p1_60; \
14704  int16x4_t __s2_60 = __p2_60; \
14705  int16x4_t __ret_60; \
14706  __ret_60 = __s0_60 + __s1_60 * splat_lane_s16(__s2_60, __p3_60); \
14707  __ret_60; \
14708 })
14709 #else
14710 #define vmla_lane_s16(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
14711  int16x4_t __s0_61 = __p0_61; \
14712  int16x4_t __s1_61 = __p1_61; \
14713  int16x4_t __s2_61 = __p2_61; \
14714  int16x4_t __rev0_61; __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
14715  int16x4_t __rev1_61; __rev1_61 = __builtin_shufflevector(__s1_61, __s1_61, 3, 2, 1, 0); \
14716  int16x4_t __rev2_61; __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
14717  int16x4_t __ret_61; \
14718  __ret_61 = __rev0_61 + __rev1_61 * __noswap_splat_lane_s16(__rev2_61, __p3_61); \
14719  __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
14720  __ret_61; \
14721 })
14722 #endif
14723 
14724 #ifdef __LITTLE_ENDIAN__
14725 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
14726  uint32x4_t __ret;
14727  __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
14728  return __ret;
14729 }
14730 #else
14731 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
14732  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14733  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14734  uint32x4_t __ret;
14735  __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
14736  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14737  return __ret;
14738 }
14739 #endif
14740 
14741 #ifdef __LITTLE_ENDIAN__
14742 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
14743  uint16x8_t __ret;
14744  __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14745  return __ret;
14746 }
14747 #else
14748 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
14749  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14750  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14751  uint16x8_t __ret;
14752  __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14753  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14754  return __ret;
14755 }
14756 #endif
14757 
14758 #ifdef __LITTLE_ENDIAN__
14759 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
14760  float32x4_t __ret;
14761  __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
14762  return __ret;
14763 }
14764 #else
14765 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
14766  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14767  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14768  float32x4_t __ret;
14769  __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
14770  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14771  return __ret;
14772 }
14773 #endif
14774 
14775 #ifdef __LITTLE_ENDIAN__
14776 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
14777  int32x4_t __ret;
14778  __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
14779  return __ret;
14780 }
14781 #else
14782 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
14783  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14784  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14785  int32x4_t __ret;
14786  __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
14787  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14788  return __ret;
14789 }
14790 #endif
14791 
14792 #ifdef __LITTLE_ENDIAN__
14793 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
14794  int16x8_t __ret;
14795  __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14796  return __ret;
14797 }
14798 #else
14799 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
14800  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14801  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14802  int16x8_t __ret;
14803  __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
14804  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14805  return __ret;
14806 }
14807 #endif
14808 
14809 #ifdef __LITTLE_ENDIAN__
14810 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
14811  uint32x2_t __ret;
14812  __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
14813  return __ret;
14814 }
14815 #else
14816 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
14817  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14818  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14819  uint32x2_t __ret;
14820  __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
14821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14822  return __ret;
14823 }
14824 #endif
14825 
14826 #ifdef __LITTLE_ENDIAN__
14827 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
14828  uint16x4_t __ret;
14829  __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
14830  return __ret;
14831 }
14832 #else
14833 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
14834  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14835  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14836  uint16x4_t __ret;
14837  __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
14838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14839  return __ret;
14840 }
14841 #endif
14842 
14843 #ifdef __LITTLE_ENDIAN__
14844 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
14845  float32x2_t __ret;
14846  __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
14847  return __ret;
14848 }
14849 #else
14850 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
14851  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14852  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14853  float32x2_t __ret;
14854  __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
14855  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14856  return __ret;
14857 }
14858 #endif
14859 
14860 #ifdef __LITTLE_ENDIAN__
14861 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
14862  int32x2_t __ret;
14863  __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
14864  return __ret;
14865 }
14866 #else
14867 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
14868  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14869  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14870  int32x2_t __ret;
14871  __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
14872  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14873  return __ret;
14874 }
14875 #endif
14876 
14877 #ifdef __LITTLE_ENDIAN__
14878 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
14879  int16x4_t __ret;
14880  __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
14881  return __ret;
14882 }
14883 #else
14884 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
14885  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14886  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14887  int16x4_t __ret;
14888  __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
14889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14890  return __ret;
14891 }
14892 #endif
14893 
14894 #ifdef __LITTLE_ENDIAN__
14895 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14896  uint8x16_t __ret;
14897  __ret = __p0 - __p1 * __p2;
14898  return __ret;
14899 }
14900 #else
14901 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
14902  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14903  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14904  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14905  uint8x16_t __ret;
14906  __ret = __rev0 - __rev1 * __rev2;
14907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14908  return __ret;
14909 }
14910 #endif
14911 
14912 #ifdef __LITTLE_ENDIAN__
14913 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14914  uint32x4_t __ret;
14915  __ret = __p0 - __p1 * __p2;
14916  return __ret;
14917 }
14918 #else
14919 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
14920  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14921  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14922  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14923  uint32x4_t __ret;
14924  __ret = __rev0 - __rev1 * __rev2;
14925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14926  return __ret;
14927 }
14928 #endif
14929 
14930 #ifdef __LITTLE_ENDIAN__
14931 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14932  uint16x8_t __ret;
14933  __ret = __p0 - __p1 * __p2;
14934  return __ret;
14935 }
14936 #else
14937 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
14938  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14939  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14940  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
14941  uint16x8_t __ret;
14942  __ret = __rev0 - __rev1 * __rev2;
14943  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14944  return __ret;
14945 }
14946 #endif
14947 
14948 #ifdef __LITTLE_ENDIAN__
14949 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14950  int8x16_t __ret;
14951  __ret = __p0 - __p1 * __p2;
14952  return __ret;
14953 }
14954 #else
14955 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
14956  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14957  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14958  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14959  int8x16_t __ret;
14960  __ret = __rev0 - __rev1 * __rev2;
14961  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14962  return __ret;
14963 }
14964 #endif
14965 
14966 #ifdef __LITTLE_ENDIAN__
14967 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14968  float32x4_t __ret;
14969  __ret = __p0 - __p1 * __p2;
14970  return __ret;
14971 }
14972 #else
14973 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
14974  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14975  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14976  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14977  float32x4_t __ret;
14978  __ret = __rev0 - __rev1 * __rev2;
14979  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14980  return __ret;
14981 }
14982 #endif
14983 
14984 #ifdef __LITTLE_ENDIAN__
14985 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14986  int32x4_t __ret;
14987  __ret = __p0 - __p1 * __p2;
14988  return __ret;
14989 }
14990 #else
14991 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
14992  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14993  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14994  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
14995  int32x4_t __ret;
14996  __ret = __rev0 - __rev1 * __rev2;
14997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14998  return __ret;
14999 }
15000 #endif
15001 
15002 #ifdef __LITTLE_ENDIAN__
15003 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
15004  int16x8_t __ret;
15005  __ret = __p0 - __p1 * __p2;
15006  return __ret;
15007 }
15008 #else
15009 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
15010  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15011  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15012  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
15013  int16x8_t __ret;
15014  __ret = __rev0 - __rev1 * __rev2;
15015  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15016  return __ret;
15017 }
15018 #endif
15019 
15020 #ifdef __LITTLE_ENDIAN__
15021 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
15022  uint8x8_t __ret;
15023  __ret = __p0 - __p1 * __p2;
15024  return __ret;
15025 }
15026 #else
15027 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
15028  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15029  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15030  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
15031  uint8x8_t __ret;
15032  __ret = __rev0 - __rev1 * __rev2;
15033  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15034  return __ret;
15035 }
15036 #endif
15037 
15038 #ifdef __LITTLE_ENDIAN__
15039 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
15040  uint32x2_t __ret;
15041  __ret = __p0 - __p1 * __p2;
15042  return __ret;
15043 }
15044 #else
15045 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
15046  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15047  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15048  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
15049  uint32x2_t __ret;
15050  __ret = __rev0 - __rev1 * __rev2;
15051  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15052  return __ret;
15053 }
15054 #endif
15055 
15056 #ifdef __LITTLE_ENDIAN__
15057 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
15058  uint16x4_t __ret;
15059  __ret = __p0 - __p1 * __p2;
15060  return __ret;
15061 }
15062 #else
15063 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
15064  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15065  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15066  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
15067  uint16x4_t __ret;
15068  __ret = __rev0 - __rev1 * __rev2;
15069  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15070  return __ret;
15071 }
15072 #endif
15073 
15074 #ifdef __LITTLE_ENDIAN__
15075 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
15076  int8x8_t __ret;
15077  __ret = __p0 - __p1 * __p2;
15078  return __ret;
15079 }
15080 #else
15081 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
15082  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15083  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15084  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
15085  int8x8_t __ret;
15086  __ret = __rev0 - __rev1 * __rev2;
15087  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15088  return __ret;
15089 }
15090 #endif
15091 
15092 #ifdef __LITTLE_ENDIAN__
15093 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
15094  float32x2_t __ret;
15095  __ret = __p0 - __p1 * __p2;
15096  return __ret;
15097 }
15098 #else
15099 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
15100  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15101  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15102  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
15103  float32x2_t __ret;
15104  __ret = __rev0 - __rev1 * __rev2;
15105  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15106  return __ret;
15107 }
15108 #endif
15109 
15110 #ifdef __LITTLE_ENDIAN__
15111 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
15112  int32x2_t __ret;
15113  __ret = __p0 - __p1 * __p2;
15114  return __ret;
15115 }
15116 #else
15117 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
15118  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15119  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15120  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
15121  int32x2_t __ret;
15122  __ret = __rev0 - __rev1 * __rev2;
15123  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15124  return __ret;
15125 }
15126 #endif
15127 
15128 #ifdef __LITTLE_ENDIAN__
15129 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
15130  int16x4_t __ret;
15131  __ret = __p0 - __p1 * __p2;
15132  return __ret;
15133 }
15134 #else
15135 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
15136  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15137  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15138  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
15139  int16x4_t __ret;
15140  __ret = __rev0 - __rev1 * __rev2;
15141  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15142  return __ret;
15143 }
15144 #endif
15145 
15146 #ifdef __LITTLE_ENDIAN__
15147 #define vmlsq_lane_u32(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
15148  uint32x4_t __s0_62 = __p0_62; \
15149  uint32x4_t __s1_62 = __p1_62; \
15150  uint32x2_t __s2_62 = __p2_62; \
15151  uint32x4_t __ret_62; \
15152  __ret_62 = __s0_62 - __s1_62 * splatq_lane_u32(__s2_62, __p3_62); \
15153  __ret_62; \
15154 })
15155 #else
15156 #define vmlsq_lane_u32(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
15157  uint32x4_t __s0_63 = __p0_63; \
15158  uint32x4_t __s1_63 = __p1_63; \
15159  uint32x2_t __s2_63 = __p2_63; \
15160  uint32x4_t __rev0_63; __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 3, 2, 1, 0); \
15161  uint32x4_t __rev1_63; __rev1_63 = __builtin_shufflevector(__s1_63, __s1_63, 3, 2, 1, 0); \
15162  uint32x2_t __rev2_63; __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
15163  uint32x4_t __ret_63; \
15164  __ret_63 = __rev0_63 - __rev1_63 * __noswap_splatq_lane_u32(__rev2_63, __p3_63); \
15165  __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 3, 2, 1, 0); \
15166  __ret_63; \
15167 })
15168 #endif
15169 
15170 #ifdef __LITTLE_ENDIAN__
15171 #define vmlsq_lane_u16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
15172  uint16x8_t __s0_64 = __p0_64; \
15173  uint16x8_t __s1_64 = __p1_64; \
15174  uint16x4_t __s2_64 = __p2_64; \
15175  uint16x8_t __ret_64; \
15176  __ret_64 = __s0_64 - __s1_64 * splatq_lane_u16(__s2_64, __p3_64); \
15177  __ret_64; \
15178 })
15179 #else
15180 #define vmlsq_lane_u16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
15181  uint16x8_t __s0_65 = __p0_65; \
15182  uint16x8_t __s1_65 = __p1_65; \
15183  uint16x4_t __s2_65 = __p2_65; \
15184  uint16x8_t __rev0_65; __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
15185  uint16x8_t __rev1_65; __rev1_65 = __builtin_shufflevector(__s1_65, __s1_65, 7, 6, 5, 4, 3, 2, 1, 0); \
15186  uint16x4_t __rev2_65; __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 3, 2, 1, 0); \
15187  uint16x8_t __ret_65; \
15188  __ret_65 = __rev0_65 - __rev1_65 * __noswap_splatq_lane_u16(__rev2_65, __p3_65); \
15189  __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
15190  __ret_65; \
15191 })
15192 #endif
15193 
15194 #ifdef __LITTLE_ENDIAN__
15195 #define vmlsq_lane_f32(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
15196  float32x4_t __s0_66 = __p0_66; \
15197  float32x4_t __s1_66 = __p1_66; \
15198  float32x2_t __s2_66 = __p2_66; \
15199  float32x4_t __ret_66; \
15200  __ret_66 = __s0_66 - __s1_66 * splatq_lane_f32(__s2_66, __p3_66); \
15201  __ret_66; \
15202 })
15203 #else
15204 #define vmlsq_lane_f32(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
15205  float32x4_t __s0_67 = __p0_67; \
15206  float32x4_t __s1_67 = __p1_67; \
15207  float32x2_t __s2_67 = __p2_67; \
15208  float32x4_t __rev0_67; __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 3, 2, 1, 0); \
15209  float32x4_t __rev1_67; __rev1_67 = __builtin_shufflevector(__s1_67, __s1_67, 3, 2, 1, 0); \
15210  float32x2_t __rev2_67; __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 1, 0); \
15211  float32x4_t __ret_67; \
15212  __ret_67 = __rev0_67 - __rev1_67 * __noswap_splatq_lane_f32(__rev2_67, __p3_67); \
15213  __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 3, 2, 1, 0); \
15214  __ret_67; \
15215 })
15216 #endif
15217 
15218 #ifdef __LITTLE_ENDIAN__
15219 #define vmlsq_lane_s32(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
15220  int32x4_t __s0_68 = __p0_68; \
15221  int32x4_t __s1_68 = __p1_68; \
15222  int32x2_t __s2_68 = __p2_68; \
15223  int32x4_t __ret_68; \
15224  __ret_68 = __s0_68 - __s1_68 * splatq_lane_s32(__s2_68, __p3_68); \
15225  __ret_68; \
15226 })
15227 #else
15228 #define vmlsq_lane_s32(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
15229  int32x4_t __s0_69 = __p0_69; \
15230  int32x4_t __s1_69 = __p1_69; \
15231  int32x2_t __s2_69 = __p2_69; \
15232  int32x4_t __rev0_69; __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
15233  int32x4_t __rev1_69; __rev1_69 = __builtin_shufflevector(__s1_69, __s1_69, 3, 2, 1, 0); \
15234  int32x2_t __rev2_69; __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 1, 0); \
15235  int32x4_t __ret_69; \
15236  __ret_69 = __rev0_69 - __rev1_69 * __noswap_splatq_lane_s32(__rev2_69, __p3_69); \
15237  __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
15238  __ret_69; \
15239 })
15240 #endif
15241 
15242 #ifdef __LITTLE_ENDIAN__
15243 #define vmlsq_lane_s16(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
15244  int16x8_t __s0_70 = __p0_70; \
15245  int16x8_t __s1_70 = __p1_70; \
15246  int16x4_t __s2_70 = __p2_70; \
15247  int16x8_t __ret_70; \
15248  __ret_70 = __s0_70 - __s1_70 * splatq_lane_s16(__s2_70, __p3_70); \
15249  __ret_70; \
15250 })
15251 #else
15252 #define vmlsq_lane_s16(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
15253  int16x8_t __s0_71 = __p0_71; \
15254  int16x8_t __s1_71 = __p1_71; \
15255  int16x4_t __s2_71 = __p2_71; \
15256  int16x8_t __rev0_71; __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
15257  int16x8_t __rev1_71; __rev1_71 = __builtin_shufflevector(__s1_71, __s1_71, 7, 6, 5, 4, 3, 2, 1, 0); \
15258  int16x4_t __rev2_71; __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 3, 2, 1, 0); \
15259  int16x8_t __ret_71; \
15260  __ret_71 = __rev0_71 - __rev1_71 * __noswap_splatq_lane_s16(__rev2_71, __p3_71); \
15261  __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
15262  __ret_71; \
15263 })
15264 #endif
15265 
15266 #ifdef __LITTLE_ENDIAN__
15267 #define vmls_lane_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
15268  uint32x2_t __s0_72 = __p0_72; \
15269  uint32x2_t __s1_72 = __p1_72; \
15270  uint32x2_t __s2_72 = __p2_72; \
15271  uint32x2_t __ret_72; \
15272  __ret_72 = __s0_72 - __s1_72 * splat_lane_u32(__s2_72, __p3_72); \
15273  __ret_72; \
15274 })
15275 #else
15276 #define vmls_lane_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
15277  uint32x2_t __s0_73 = __p0_73; \
15278  uint32x2_t __s1_73 = __p1_73; \
15279  uint32x2_t __s2_73 = __p2_73; \
15280  uint32x2_t __rev0_73; __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
15281  uint32x2_t __rev1_73; __rev1_73 = __builtin_shufflevector(__s1_73, __s1_73, 1, 0); \
15282  uint32x2_t __rev2_73; __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 1, 0); \
15283  uint32x2_t __ret_73; \
15284  __ret_73 = __rev0_73 - __rev1_73 * __noswap_splat_lane_u32(__rev2_73, __p3_73); \
15285  __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
15286  __ret_73; \
15287 })
15288 #endif
15289 
15290 #ifdef __LITTLE_ENDIAN__
15291 #define vmls_lane_u16(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
15292  uint16x4_t __s0_74 = __p0_74; \
15293  uint16x4_t __s1_74 = __p1_74; \
15294  uint16x4_t __s2_74 = __p2_74; \
15295  uint16x4_t __ret_74; \
15296  __ret_74 = __s0_74 - __s1_74 * splat_lane_u16(__s2_74, __p3_74); \
15297  __ret_74; \
15298 })
15299 #else
15300 #define vmls_lane_u16(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
15301  uint16x4_t __s0_75 = __p0_75; \
15302  uint16x4_t __s1_75 = __p1_75; \
15303  uint16x4_t __s2_75 = __p2_75; \
15304  uint16x4_t __rev0_75; __rev0_75 = __builtin_shufflevector(__s0_75, __s0_75, 3, 2, 1, 0); \
15305  uint16x4_t __rev1_75; __rev1_75 = __builtin_shufflevector(__s1_75, __s1_75, 3, 2, 1, 0); \
15306  uint16x4_t __rev2_75; __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 3, 2, 1, 0); \
15307  uint16x4_t __ret_75; \
15308  __ret_75 = __rev0_75 - __rev1_75 * __noswap_splat_lane_u16(__rev2_75, __p3_75); \
15309  __ret_75 = __builtin_shufflevector(__ret_75, __ret_75, 3, 2, 1, 0); \
15310  __ret_75; \
15311 })
15312 #endif
15313 
15314 #ifdef __LITTLE_ENDIAN__
15315 #define vmls_lane_f32(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
15316  float32x2_t __s0_76 = __p0_76; \
15317  float32x2_t __s1_76 = __p1_76; \
15318  float32x2_t __s2_76 = __p2_76; \
15319  float32x2_t __ret_76; \
15320  __ret_76 = __s0_76 - __s1_76 * splat_lane_f32(__s2_76, __p3_76); \
15321  __ret_76; \
15322 })
15323 #else
15324 #define vmls_lane_f32(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
15325  float32x2_t __s0_77 = __p0_77; \
15326  float32x2_t __s1_77 = __p1_77; \
15327  float32x2_t __s2_77 = __p2_77; \
15328  float32x2_t __rev0_77; __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 1, 0); \
15329  float32x2_t __rev1_77; __rev1_77 = __builtin_shufflevector(__s1_77, __s1_77, 1, 0); \
15330  float32x2_t __rev2_77; __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 1, 0); \
15331  float32x2_t __ret_77; \
15332  __ret_77 = __rev0_77 - __rev1_77 * __noswap_splat_lane_f32(__rev2_77, __p3_77); \
15333  __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 1, 0); \
15334  __ret_77; \
15335 })
15336 #endif
15337 
15338 #ifdef __LITTLE_ENDIAN__
15339 #define vmls_lane_s32(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
15340  int32x2_t __s0_78 = __p0_78; \
15341  int32x2_t __s1_78 = __p1_78; \
15342  int32x2_t __s2_78 = __p2_78; \
15343  int32x2_t __ret_78; \
15344  __ret_78 = __s0_78 - __s1_78 * splat_lane_s32(__s2_78, __p3_78); \
15345  __ret_78; \
15346 })
15347 #else
15348 #define vmls_lane_s32(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
15349  int32x2_t __s0_79 = __p0_79; \
15350  int32x2_t __s1_79 = __p1_79; \
15351  int32x2_t __s2_79 = __p2_79; \
15352  int32x2_t __rev0_79; __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 1, 0); \
15353  int32x2_t __rev1_79; __rev1_79 = __builtin_shufflevector(__s1_79, __s1_79, 1, 0); \
15354  int32x2_t __rev2_79; __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 1, 0); \
15355  int32x2_t __ret_79; \
15356  __ret_79 = __rev0_79 - __rev1_79 * __noswap_splat_lane_s32(__rev2_79, __p3_79); \
15357  __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 1, 0); \
15358  __ret_79; \
15359 })
15360 #endif
15361 
15362 #ifdef __LITTLE_ENDIAN__
15363 #define vmls_lane_s16(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
15364  int16x4_t __s0_80 = __p0_80; \
15365  int16x4_t __s1_80 = __p1_80; \
15366  int16x4_t __s2_80 = __p2_80; \
15367  int16x4_t __ret_80; \
15368  __ret_80 = __s0_80 - __s1_80 * splat_lane_s16(__s2_80, __p3_80); \
15369  __ret_80; \
15370 })
15371 #else
15372 #define vmls_lane_s16(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
15373  int16x4_t __s0_81 = __p0_81; \
15374  int16x4_t __s1_81 = __p1_81; \
15375  int16x4_t __s2_81 = __p2_81; \
15376  int16x4_t __rev0_81; __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 3, 2, 1, 0); \
15377  int16x4_t __rev1_81; __rev1_81 = __builtin_shufflevector(__s1_81, __s1_81, 3, 2, 1, 0); \
15378  int16x4_t __rev2_81; __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
15379  int16x4_t __ret_81; \
15380  __ret_81 = __rev0_81 - __rev1_81 * __noswap_splat_lane_s16(__rev2_81, __p3_81); \
15381  __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 3, 2, 1, 0); \
15382  __ret_81; \
15383 })
15384 #endif
15385 
15386 #ifdef __LITTLE_ENDIAN__
15387 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
15388  uint32x4_t __ret;
15389  __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
15390  return __ret;
15391 }
15392 #else
15393 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
15394  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15395  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15396  uint32x4_t __ret;
15397  __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
15398  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15399  return __ret;
15400 }
15401 #endif
15402 
15403 #ifdef __LITTLE_ENDIAN__
15404 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
15405  uint16x8_t __ret;
15406  __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15407  return __ret;
15408 }
15409 #else
15410 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
15411  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15412  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15413  uint16x8_t __ret;
15414  __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15415  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15416  return __ret;
15417 }
15418 #endif
15419 
15420 #ifdef __LITTLE_ENDIAN__
15421 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
15422  float32x4_t __ret;
15423  __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
15424  return __ret;
15425 }
15426 #else
15427 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
15428  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15429  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15430  float32x4_t __ret;
15431  __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
15432  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15433  return __ret;
15434 }
15435 #endif
15436 
15437 #ifdef __LITTLE_ENDIAN__
15438 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
15439  int32x4_t __ret;
15440  __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
15441  return __ret;
15442 }
15443 #else
15444 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
15445  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15446  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15447  int32x4_t __ret;
15448  __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
15449  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15450  return __ret;
15451 }
15452 #endif
15453 
15454 #ifdef __LITTLE_ENDIAN__
15455 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
15456  int16x8_t __ret;
15457  __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15458  return __ret;
15459 }
15460 #else
15461 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
15462  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15463  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15464  int16x8_t __ret;
15465  __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
15466  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15467  return __ret;
15468 }
15469 #endif
15470 
15471 #ifdef __LITTLE_ENDIAN__
15472 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
15473  uint32x2_t __ret;
15474  __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
15475  return __ret;
15476 }
15477 #else
15478 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
15479  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15480  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15481  uint32x2_t __ret;
15482  __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
15483  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15484  return __ret;
15485 }
15486 #endif
15487 
15488 #ifdef __LITTLE_ENDIAN__
15489 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
15490  uint16x4_t __ret;
15491  __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
15492  return __ret;
15493 }
15494 #else
15495 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
15496  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15497  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15498  uint16x4_t __ret;
15499  __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
15500  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15501  return __ret;
15502 }
15503 #endif
15504 
15505 #ifdef __LITTLE_ENDIAN__
15506 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
15507  float32x2_t __ret;
15508  __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
15509  return __ret;
15510 }
15511 #else
15512 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
15513  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15514  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15515  float32x2_t __ret;
15516  __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
15517  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15518  return __ret;
15519 }
15520 #endif
15521 
15522 #ifdef __LITTLE_ENDIAN__
15523 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
15524  int32x2_t __ret;
15525  __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
15526  return __ret;
15527 }
15528 #else
15529 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
15530  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15531  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15532  int32x2_t __ret;
15533  __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
15534  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15535  return __ret;
15536 }
15537 #endif
15538 
15539 #ifdef __LITTLE_ENDIAN__
15540 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
15541  int16x4_t __ret;
15542  __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
15543  return __ret;
15544 }
15545 #else
15546 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
15547  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15548  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15549  int16x4_t __ret;
15550  __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
15551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15552  return __ret;
15553 }
15554 #endif
15555 
15556 #ifdef __LITTLE_ENDIAN__
15557 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
15558  poly8x8_t __ret;
15559  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15560  return __ret;
15561 }
15562 #else
15563 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
15564  poly8x8_t __ret;
15565  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15566  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15567  return __ret;
15568 }
15569 #endif
15570 
15571 #ifdef __LITTLE_ENDIAN__
15572 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
15573  poly16x4_t __ret;
15574  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
15575  return __ret;
15576 }
15577 #else
15578 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
15579  poly16x4_t __ret;
15580  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
15581  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15582  return __ret;
15583 }
15584 #endif
15585 
15586 #ifdef __LITTLE_ENDIAN__
15587 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
15588  poly8x16_t __ret;
15589  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15590  return __ret;
15591 }
15592 #else
15593 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
15594  poly8x16_t __ret;
15595  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15596  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15597  return __ret;
15598 }
15599 #endif
15600 
15601 #ifdef __LITTLE_ENDIAN__
15602 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
15603  poly16x8_t __ret;
15604  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15605  return __ret;
15606 }
15607 #else
15608 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
15609  poly16x8_t __ret;
15610  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15611  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15612  return __ret;
15613 }
15614 #endif
15615 
15616 #ifdef __LITTLE_ENDIAN__
15617 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
15618  uint8x16_t __ret;
15619  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15620  return __ret;
15621 }
15622 #else
15623 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
15624  uint8x16_t __ret;
15625  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15626  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15627  return __ret;
15628 }
15629 #endif
15630 
15631 #ifdef __LITTLE_ENDIAN__
15632 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
15633  uint32x4_t __ret;
15634  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
15635  return __ret;
15636 }
15637 #else
15638 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
15639  uint32x4_t __ret;
15640  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
15641  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15642  return __ret;
15643 }
15644 #endif
15645 
15646 #ifdef __LITTLE_ENDIAN__
15647 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
15648  uint64x2_t __ret;
15649  __ret = (uint64x2_t) {__p0, __p0};
15650  return __ret;
15651 }
15652 #else
15653 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
15654  uint64x2_t __ret;
15655  __ret = (uint64x2_t) {__p0, __p0};
15656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15657  return __ret;
15658 }
15659 #endif
15660 
15661 #ifdef __LITTLE_ENDIAN__
15662 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
15663  uint16x8_t __ret;
15664  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15665  return __ret;
15666 }
15667 #else
15668 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
15669  uint16x8_t __ret;
15670  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15671  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15672  return __ret;
15673 }
15674 #endif
15675 
15676 #ifdef __LITTLE_ENDIAN__
15677 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
15678  int8x16_t __ret;
15679  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15680  return __ret;
15681 }
15682 #else
15683 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
15684  int8x16_t __ret;
15685  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15686  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15687  return __ret;
15688 }
15689 #endif
15690 
15691 #ifdef __LITTLE_ENDIAN__
15692 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
15693  float32x4_t __ret;
15694  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
15695  return __ret;
15696 }
15697 #else
15698 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
15699  float32x4_t __ret;
15700  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
15701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15702  return __ret;
15703 }
15704 #endif
15705 
15706 #ifdef __LITTLE_ENDIAN__
15707 #define vmovq_n_f16(__p0) __extension__ ({ \
15708  float16_t __s0 = __p0; \
15709  float16x8_t __ret; \
15710  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
15711  __ret; \
15712 })
15713 #else
15714 #define vmovq_n_f16(__p0) __extension__ ({ \
15715  float16_t __s0 = __p0; \
15716  float16x8_t __ret; \
15717  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
15718  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
15719  __ret; \
15720 })
15721 #endif
15722 
15723 #ifdef __LITTLE_ENDIAN__
15724 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
15725  int32x4_t __ret;
15726  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
15727  return __ret;
15728 }
15729 #else
15730 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
15731  int32x4_t __ret;
15732  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
15733  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15734  return __ret;
15735 }
15736 #endif
15737 
15738 #ifdef __LITTLE_ENDIAN__
15739 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
15740  int64x2_t __ret;
15741  __ret = (int64x2_t) {__p0, __p0};
15742  return __ret;
15743 }
15744 #else
15745 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
15746  int64x2_t __ret;
15747  __ret = (int64x2_t) {__p0, __p0};
15748  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15749  return __ret;
15750 }
15751 #endif
15752 
15753 #ifdef __LITTLE_ENDIAN__
15754 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
15755  int16x8_t __ret;
15756  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15757  return __ret;
15758 }
15759 #else
15760 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
15761  int16x8_t __ret;
15762  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15763  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15764  return __ret;
15765 }
15766 #endif
15767 
15768 #ifdef __LITTLE_ENDIAN__
15769 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
15770  uint8x8_t __ret;
15771  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15772  return __ret;
15773 }
15774 #else
15775 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
15776  uint8x8_t __ret;
15777  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15778  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15779  return __ret;
15780 }
15781 #endif
15782 
15783 #ifdef __LITTLE_ENDIAN__
15784 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
15785  uint32x2_t __ret;
15786  __ret = (uint32x2_t) {__p0, __p0};
15787  return __ret;
15788 }
15789 #else
15790 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
15791  uint32x2_t __ret;
15792  __ret = (uint32x2_t) {__p0, __p0};
15793  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15794  return __ret;
15795 }
15796 #endif
15797 
15798 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
15799  uint64x1_t __ret;
15800  __ret = (uint64x1_t) {__p0};
15801  return __ret;
15802 }
15803 #ifdef __LITTLE_ENDIAN__
15804 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
15805  uint16x4_t __ret;
15806  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
15807  return __ret;
15808 }
15809 #else
15810 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
15811  uint16x4_t __ret;
15812  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
15813  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15814  return __ret;
15815 }
15816 #endif
15817 
15818 #ifdef __LITTLE_ENDIAN__
15819 __ai int8x8_t vmov_n_s8(int8_t __p0) {
15820  int8x8_t __ret;
15821  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15822  return __ret;
15823 }
15824 #else
15825 __ai int8x8_t vmov_n_s8(int8_t __p0) {
15826  int8x8_t __ret;
15827  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
15828  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15829  return __ret;
15830 }
15831 #endif
15832 
15833 #ifdef __LITTLE_ENDIAN__
15834 __ai float32x2_t vmov_n_f32(float32_t __p0) {
15835  float32x2_t __ret;
15836  __ret = (float32x2_t) {__p0, __p0};
15837  return __ret;
15838 }
15839 #else
15840 __ai float32x2_t vmov_n_f32(float32_t __p0) {
15841  float32x2_t __ret;
15842  __ret = (float32x2_t) {__p0, __p0};
15843  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15844  return __ret;
15845 }
15846 #endif
15847 
15848 #ifdef __LITTLE_ENDIAN__
15849 #define vmov_n_f16(__p0) __extension__ ({ \
15850  float16_t __s0 = __p0; \
15851  float16x4_t __ret; \
15852  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
15853  __ret; \
15854 })
15855 #else
15856 #define vmov_n_f16(__p0) __extension__ ({ \
15857  float16_t __s0 = __p0; \
15858  float16x4_t __ret; \
15859  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
15860  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
15861  __ret; \
15862 })
15863 #endif
15864 
15865 #ifdef __LITTLE_ENDIAN__
15866 __ai int32x2_t vmov_n_s32(int32_t __p0) {
15867  int32x2_t __ret;
15868  __ret = (int32x2_t) {__p0, __p0};
15869  return __ret;
15870 }
15871 #else
15872 __ai int32x2_t vmov_n_s32(int32_t __p0) {
15873  int32x2_t __ret;
15874  __ret = (int32x2_t) {__p0, __p0};
15875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15876  return __ret;
15877 }
15878 #endif
15879 
15880 __ai int64x1_t vmov_n_s64(int64_t __p0) {
15881  int64x1_t __ret;
15882  __ret = (int64x1_t) {__p0};
15883  return __ret;
15884 }
15885 #ifdef __LITTLE_ENDIAN__
15886 __ai int16x4_t vmov_n_s16(int16_t __p0) {
15887  int16x4_t __ret;
15888  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
15889  return __ret;
15890 }
15891 #else
15892 __ai int16x4_t vmov_n_s16(int16_t __p0) {
15893  int16x4_t __ret;
15894  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
15895  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15896  return __ret;
15897 }
15898 #endif
15899 
15900 #ifdef __LITTLE_ENDIAN__
15901 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
15902  uint16x8_t __ret;
15903  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
15904  return __ret;
15905 }
15906 #else
15907 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
15908  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15909  uint16x8_t __ret;
15910  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
15911  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15912  return __ret;
15913 }
15914 __ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
15915  uint16x8_t __ret;
15916  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
15917  return __ret;
15918 }
15919 #endif
15920 
15921 #ifdef __LITTLE_ENDIAN__
15922 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
15923  uint64x2_t __ret;
15924  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
15925  return __ret;
15926 }
15927 #else
15928 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
15929  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15930  uint64x2_t __ret;
15931  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
15932  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15933  return __ret;
15934 }
15935 __ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
15936  uint64x2_t __ret;
15937  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
15938  return __ret;
15939 }
15940 #endif
15941 
15942 #ifdef __LITTLE_ENDIAN__
15943 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
15944  uint32x4_t __ret;
15945  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
15946  return __ret;
15947 }
15948 #else
15949 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
15950  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15951  uint32x4_t __ret;
15952  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
15953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15954  return __ret;
15955 }
15956 __ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
15957  uint32x4_t __ret;
15958  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
15959  return __ret;
15960 }
15961 #endif
15962 
15963 #ifdef __LITTLE_ENDIAN__
15964 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
15965  int16x8_t __ret;
15966  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
15967  return __ret;
15968 }
15969 #else
15970 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
15971  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15972  int16x8_t __ret;
15973  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
15974  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15975  return __ret;
15976 }
15977 __ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
15978  int16x8_t __ret;
15979  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
15980  return __ret;
15981 }
15982 #endif
15983 
15984 #ifdef __LITTLE_ENDIAN__
15985 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
15986  int64x2_t __ret;
15987  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
15988  return __ret;
15989 }
15990 #else
15991 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
15992  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15993  int64x2_t __ret;
15994  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
15995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15996  return __ret;
15997 }
15998 __ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
15999  int64x2_t __ret;
16000  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
16001  return __ret;
16002 }
16003 #endif
16004 
16005 #ifdef __LITTLE_ENDIAN__
16006 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
16007  int32x4_t __ret;
16008  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
16009  return __ret;
16010 }
16011 #else
16012 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
16013  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16014  int32x4_t __ret;
16015  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
16016  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16017  return __ret;
16018 }
16019 __ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
16020  int32x4_t __ret;
16021  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
16022  return __ret;
16023 }
16024 #endif
16025 
16026 #ifdef __LITTLE_ENDIAN__
16027 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
16028  uint16x4_t __ret;
16029  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
16030  return __ret;
16031 }
16032 #else
16033 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
16034  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16035  uint16x4_t __ret;
16036  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
16037  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16038  return __ret;
16039 }
16040 __ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
16041  uint16x4_t __ret;
16042  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
16043  return __ret;
16044 }
16045 #endif
16046 
16047 #ifdef __LITTLE_ENDIAN__
16048 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
16049  uint32x2_t __ret;
16050  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
16051  return __ret;
16052 }
16053 #else
16054 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
16055  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16056  uint32x2_t __ret;
16057  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
16058  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16059  return __ret;
16060 }
16061 __ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
16062  uint32x2_t __ret;
16063  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
16064  return __ret;
16065 }
16066 #endif
16067 
16068 #ifdef __LITTLE_ENDIAN__
16069 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
16070  uint8x8_t __ret;
16071  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
16072  return __ret;
16073 }
16074 #else
16075 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
16076  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16077  uint8x8_t __ret;
16078  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
16079  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16080  return __ret;
16081 }
16082 __ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
16083  uint8x8_t __ret;
16084  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
16085  return __ret;
16086 }
16087 #endif
16088 
16089 #ifdef __LITTLE_ENDIAN__
16090 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
16091  int16x4_t __ret;
16092  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
16093  return __ret;
16094 }
16095 #else
16096 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
16097  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16098  int16x4_t __ret;
16099  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
16100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16101  return __ret;
16102 }
16103 __ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
16104  int16x4_t __ret;
16105  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
16106  return __ret;
16107 }
16108 #endif
16109 
16110 #ifdef __LITTLE_ENDIAN__
16111 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
16112  int32x2_t __ret;
16113  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
16114  return __ret;
16115 }
16116 #else
16117 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
16118  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16119  int32x2_t __ret;
16120  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
16121  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16122  return __ret;
16123 }
16124 __ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
16125  int32x2_t __ret;
16126  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
16127  return __ret;
16128 }
16129 #endif
16130 
16131 #ifdef __LITTLE_ENDIAN__
16132 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
16133  int8x8_t __ret;
16134  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
16135  return __ret;
16136 }
16137 #else
16138 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
16139  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16140  int8x8_t __ret;
16141  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
16142  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16143  return __ret;
16144 }
16145 __ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
16146  int8x8_t __ret;
16147  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
16148  return __ret;
16149 }
16150 #endif
16151 
16152 #ifdef __LITTLE_ENDIAN__
16153 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16154  uint8x16_t __ret;
16155  __ret = __p0 * __p1;
16156  return __ret;
16157 }
16158 #else
16159 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16160  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16161  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16162  uint8x16_t __ret;
16163  __ret = __rev0 * __rev1;
16164  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16165  return __ret;
16166 }
16167 #endif
16168 
16169 #ifdef __LITTLE_ENDIAN__
16170 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16171  uint32x4_t __ret;
16172  __ret = __p0 * __p1;
16173  return __ret;
16174 }
16175 #else
16176 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16177  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16178  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16179  uint32x4_t __ret;
16180  __ret = __rev0 * __rev1;
16181  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16182  return __ret;
16183 }
16184 #endif
16185 
16186 #ifdef __LITTLE_ENDIAN__
16187 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16188  uint16x8_t __ret;
16189  __ret = __p0 * __p1;
16190  return __ret;
16191 }
16192 #else
16193 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16194  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16195  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16196  uint16x8_t __ret;
16197  __ret = __rev0 * __rev1;
16198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16199  return __ret;
16200 }
16201 #endif
16202 
16203 #ifdef __LITTLE_ENDIAN__
16204 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
16205  int8x16_t __ret;
16206  __ret = __p0 * __p1;
16207  return __ret;
16208 }
16209 #else
16210 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
16211  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16212  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16213  int8x16_t __ret;
16214  __ret = __rev0 * __rev1;
16215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16216  return __ret;
16217 }
16218 #endif
16219 
16220 #ifdef __LITTLE_ENDIAN__
16221 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
16222  float32x4_t __ret;
16223  __ret = __p0 * __p1;
16224  return __ret;
16225 }
16226 #else
16227 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
16228  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16229  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16230  float32x4_t __ret;
16231  __ret = __rev0 * __rev1;
16232  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16233  return __ret;
16234 }
16235 #endif
16236 
16237 #ifdef __LITTLE_ENDIAN__
16238 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
16239  int32x4_t __ret;
16240  __ret = __p0 * __p1;
16241  return __ret;
16242 }
16243 #else
16244 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
16245  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16246  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16247  int32x4_t __ret;
16248  __ret = __rev0 * __rev1;
16249  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16250  return __ret;
16251 }
16252 #endif
16253 
16254 #ifdef __LITTLE_ENDIAN__
16255 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
16256  int16x8_t __ret;
16257  __ret = __p0 * __p1;
16258  return __ret;
16259 }
16260 #else
16261 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
16262  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16263  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16264  int16x8_t __ret;
16265  __ret = __rev0 * __rev1;
16266  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16267  return __ret;
16268 }
16269 #endif
16270 
16271 #ifdef __LITTLE_ENDIAN__
16272 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
16273  uint8x8_t __ret;
16274  __ret = __p0 * __p1;
16275  return __ret;
16276 }
16277 #else
16278 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
16279  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16280  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16281  uint8x8_t __ret;
16282  __ret = __rev0 * __rev1;
16283  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16284  return __ret;
16285 }
16286 #endif
16287 
16288 #ifdef __LITTLE_ENDIAN__
16289 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
16290  uint32x2_t __ret;
16291  __ret = __p0 * __p1;
16292  return __ret;
16293 }
16294 #else
16295 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
16296  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16297  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16298  uint32x2_t __ret;
16299  __ret = __rev0 * __rev1;
16300  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16301  return __ret;
16302 }
16303 #endif
16304 
16305 #ifdef __LITTLE_ENDIAN__
16306 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
16307  uint16x4_t __ret;
16308  __ret = __p0 * __p1;
16309  return __ret;
16310 }
16311 #else
16312 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
16313  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16314  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16315  uint16x4_t __ret;
16316  __ret = __rev0 * __rev1;
16317  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16318  return __ret;
16319 }
16320 #endif
16321 
16322 #ifdef __LITTLE_ENDIAN__
16323 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
16324  int8x8_t __ret;
16325  __ret = __p0 * __p1;
16326  return __ret;
16327 }
16328 #else
16329 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
16330  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16331  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16332  int8x8_t __ret;
16333  __ret = __rev0 * __rev1;
16334  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16335  return __ret;
16336 }
16337 #endif
16338 
16339 #ifdef __LITTLE_ENDIAN__
16340 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
16341  float32x2_t __ret;
16342  __ret = __p0 * __p1;
16343  return __ret;
16344 }
16345 #else
16346 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
16347  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16348  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16349  float32x2_t __ret;
16350  __ret = __rev0 * __rev1;
16351  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16352  return __ret;
16353 }
16354 #endif
16355 
16356 #ifdef __LITTLE_ENDIAN__
16357 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
16358  int32x2_t __ret;
16359  __ret = __p0 * __p1;
16360  return __ret;
16361 }
16362 #else
16363 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
16364  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16365  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16366  int32x2_t __ret;
16367  __ret = __rev0 * __rev1;
16368  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16369  return __ret;
16370 }
16371 #endif
16372 
16373 #ifdef __LITTLE_ENDIAN__
16374 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
16375  int16x4_t __ret;
16376  __ret = __p0 * __p1;
16377  return __ret;
16378 }
16379 #else
16380 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
16381  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16382  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16383  int16x4_t __ret;
16384  __ret = __rev0 * __rev1;
16385  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16386  return __ret;
16387 }
16388 #endif
16389 
16390 #ifdef __LITTLE_ENDIAN__
16391 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
16392  poly8x8_t __ret;
16393  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
16394  return __ret;
16395 }
16396 #else
16397 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
16398  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16399  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16400  poly8x8_t __ret;
16401  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
16402  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16403  return __ret;
16404 }
16405 #endif
16406 
16407 #ifdef __LITTLE_ENDIAN__
16408 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
16409  poly8x16_t __ret;
16410  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
16411  return __ret;
16412 }
16413 #else
16414 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
16415  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16416  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16417  poly8x16_t __ret;
16418  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
16419  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16420  return __ret;
16421 }
16422 #endif
16423 
16424 #ifdef __LITTLE_ENDIAN__
16425 #define vmulq_lane_u32(__p0_82, __p1_82, __p2_82) __extension__ ({ \
16426  uint32x4_t __s0_82 = __p0_82; \
16427  uint32x2_t __s1_82 = __p1_82; \
16428  uint32x4_t __ret_82; \
16429  __ret_82 = __s0_82 * splatq_lane_u32(__s1_82, __p2_82); \
16430  __ret_82; \
16431 })
16432 #else
16433 #define vmulq_lane_u32(__p0_83, __p1_83, __p2_83) __extension__ ({ \
16434  uint32x4_t __s0_83 = __p0_83; \
16435  uint32x2_t __s1_83 = __p1_83; \
16436  uint32x4_t __rev0_83; __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 3, 2, 1, 0); \
16437  uint32x2_t __rev1_83; __rev1_83 = __builtin_shufflevector(__s1_83, __s1_83, 1, 0); \
16438  uint32x4_t __ret_83; \
16439  __ret_83 = __rev0_83 * __noswap_splatq_lane_u32(__rev1_83, __p2_83); \
16440  __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 3, 2, 1, 0); \
16441  __ret_83; \
16442 })
16443 #endif
16444 
16445 #ifdef __LITTLE_ENDIAN__
16446 #define vmulq_lane_u16(__p0_84, __p1_84, __p2_84) __extension__ ({ \
16447  uint16x8_t __s0_84 = __p0_84; \
16448  uint16x4_t __s1_84 = __p1_84; \
16449  uint16x8_t __ret_84; \
16450  __ret_84 = __s0_84 * splatq_lane_u16(__s1_84, __p2_84); \
16451  __ret_84; \
16452 })
16453 #else
16454 #define vmulq_lane_u16(__p0_85, __p1_85, __p2_85) __extension__ ({ \
16455  uint16x8_t __s0_85 = __p0_85; \
16456  uint16x4_t __s1_85 = __p1_85; \
16457  uint16x8_t __rev0_85; __rev0_85 = __builtin_shufflevector(__s0_85, __s0_85, 7, 6, 5, 4, 3, 2, 1, 0); \
16458  uint16x4_t __rev1_85; __rev1_85 = __builtin_shufflevector(__s1_85, __s1_85, 3, 2, 1, 0); \
16459  uint16x8_t __ret_85; \
16460  __ret_85 = __rev0_85 * __noswap_splatq_lane_u16(__rev1_85, __p2_85); \
16461  __ret_85 = __builtin_shufflevector(__ret_85, __ret_85, 7, 6, 5, 4, 3, 2, 1, 0); \
16462  __ret_85; \
16463 })
16464 #endif
16465 
16466 #ifdef __LITTLE_ENDIAN__
16467 #define vmulq_lane_f32(__p0_86, __p1_86, __p2_86) __extension__ ({ \
16468  float32x4_t __s0_86 = __p0_86; \
16469  float32x2_t __s1_86 = __p1_86; \
16470  float32x4_t __ret_86; \
16471  __ret_86 = __s0_86 * splatq_lane_f32(__s1_86, __p2_86); \
16472  __ret_86; \
16473 })
16474 #else
16475 #define vmulq_lane_f32(__p0_87, __p1_87, __p2_87) __extension__ ({ \
16476  float32x4_t __s0_87 = __p0_87; \
16477  float32x2_t __s1_87 = __p1_87; \
16478  float32x4_t __rev0_87; __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
16479  float32x2_t __rev1_87; __rev1_87 = __builtin_shufflevector(__s1_87, __s1_87, 1, 0); \
16480  float32x4_t __ret_87; \
16481  __ret_87 = __rev0_87 * __noswap_splatq_lane_f32(__rev1_87, __p2_87); \
16482  __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
16483  __ret_87; \
16484 })
16485 #endif
16486 
16487 #ifdef __LITTLE_ENDIAN__
16488 #define vmulq_lane_s32(__p0_88, __p1_88, __p2_88) __extension__ ({ \
16489  int32x4_t __s0_88 = __p0_88; \
16490  int32x2_t __s1_88 = __p1_88; \
16491  int32x4_t __ret_88; \
16492  __ret_88 = __s0_88 * splatq_lane_s32(__s1_88, __p2_88); \
16493  __ret_88; \
16494 })
16495 #else
16496 #define vmulq_lane_s32(__p0_89, __p1_89, __p2_89) __extension__ ({ \
16497  int32x4_t __s0_89 = __p0_89; \
16498  int32x2_t __s1_89 = __p1_89; \
16499  int32x4_t __rev0_89; __rev0_89 = __builtin_shufflevector(__s0_89, __s0_89, 3, 2, 1, 0); \
16500  int32x2_t __rev1_89; __rev1_89 = __builtin_shufflevector(__s1_89, __s1_89, 1, 0); \
16501  int32x4_t __ret_89; \
16502  __ret_89 = __rev0_89 * __noswap_splatq_lane_s32(__rev1_89, __p2_89); \
16503  __ret_89 = __builtin_shufflevector(__ret_89, __ret_89, 3, 2, 1, 0); \
16504  __ret_89; \
16505 })
16506 #endif
16507 
16508 #ifdef __LITTLE_ENDIAN__
16509 #define vmulq_lane_s16(__p0_90, __p1_90, __p2_90) __extension__ ({ \
16510  int16x8_t __s0_90 = __p0_90; \
16511  int16x4_t __s1_90 = __p1_90; \
16512  int16x8_t __ret_90; \
16513  __ret_90 = __s0_90 * splatq_lane_s16(__s1_90, __p2_90); \
16514  __ret_90; \
16515 })
16516 #else
16517 #define vmulq_lane_s16(__p0_91, __p1_91, __p2_91) __extension__ ({ \
16518  int16x8_t __s0_91 = __p0_91; \
16519  int16x4_t __s1_91 = __p1_91; \
16520  int16x8_t __rev0_91; __rev0_91 = __builtin_shufflevector(__s0_91, __s0_91, 7, 6, 5, 4, 3, 2, 1, 0); \
16521  int16x4_t __rev1_91; __rev1_91 = __builtin_shufflevector(__s1_91, __s1_91, 3, 2, 1, 0); \
16522  int16x8_t __ret_91; \
16523  __ret_91 = __rev0_91 * __noswap_splatq_lane_s16(__rev1_91, __p2_91); \
16524  __ret_91 = __builtin_shufflevector(__ret_91, __ret_91, 7, 6, 5, 4, 3, 2, 1, 0); \
16525  __ret_91; \
16526 })
16527 #endif
16528 
16529 #ifdef __LITTLE_ENDIAN__
16530 #define vmul_lane_u32(__p0_92, __p1_92, __p2_92) __extension__ ({ \
16531  uint32x2_t __s0_92 = __p0_92; \
16532  uint32x2_t __s1_92 = __p1_92; \
16533  uint32x2_t __ret_92; \
16534  __ret_92 = __s0_92 * splat_lane_u32(__s1_92, __p2_92); \
16535  __ret_92; \
16536 })
16537 #else
16538 #define vmul_lane_u32(__p0_93, __p1_93, __p2_93) __extension__ ({ \
16539  uint32x2_t __s0_93 = __p0_93; \
16540  uint32x2_t __s1_93 = __p1_93; \
16541  uint32x2_t __rev0_93; __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
16542  uint32x2_t __rev1_93; __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
16543  uint32x2_t __ret_93; \
16544  __ret_93 = __rev0_93 * __noswap_splat_lane_u32(__rev1_93, __p2_93); \
16545  __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
16546  __ret_93; \
16547 })
16548 #endif
16549 
16550 #ifdef __LITTLE_ENDIAN__
16551 #define vmul_lane_u16(__p0_94, __p1_94, __p2_94) __extension__ ({ \
16552  uint16x4_t __s0_94 = __p0_94; \
16553  uint16x4_t __s1_94 = __p1_94; \
16554  uint16x4_t __ret_94; \
16555  __ret_94 = __s0_94 * splat_lane_u16(__s1_94, __p2_94); \
16556  __ret_94; \
16557 })
16558 #else
16559 #define vmul_lane_u16(__p0_95, __p1_95, __p2_95) __extension__ ({ \
16560  uint16x4_t __s0_95 = __p0_95; \
16561  uint16x4_t __s1_95 = __p1_95; \
16562  uint16x4_t __rev0_95; __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
16563  uint16x4_t __rev1_95; __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
16564  uint16x4_t __ret_95; \
16565  __ret_95 = __rev0_95 * __noswap_splat_lane_u16(__rev1_95, __p2_95); \
16566  __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
16567  __ret_95; \
16568 })
16569 #endif
16570 
16571 #ifdef __LITTLE_ENDIAN__
16572 #define vmul_lane_f32(__p0_96, __p1_96, __p2_96) __extension__ ({ \
16573  float32x2_t __s0_96 = __p0_96; \
16574  float32x2_t __s1_96 = __p1_96; \
16575  float32x2_t __ret_96; \
16576  __ret_96 = __s0_96 * splat_lane_f32(__s1_96, __p2_96); \
16577  __ret_96; \
16578 })
16579 #else
16580 #define vmul_lane_f32(__p0_97, __p1_97, __p2_97) __extension__ ({ \
16581  float32x2_t __s0_97 = __p0_97; \
16582  float32x2_t __s1_97 = __p1_97; \
16583  float32x2_t __rev0_97; __rev0_97 = __builtin_shufflevector(__s0_97, __s0_97, 1, 0); \
16584  float32x2_t __rev1_97; __rev1_97 = __builtin_shufflevector(__s1_97, __s1_97, 1, 0); \
16585  float32x2_t __ret_97; \
16586  __ret_97 = __rev0_97 * __noswap_splat_lane_f32(__rev1_97, __p2_97); \
16587  __ret_97 = __builtin_shufflevector(__ret_97, __ret_97, 1, 0); \
16588  __ret_97; \
16589 })
16590 #endif
16591 
16592 #ifdef __LITTLE_ENDIAN__
16593 #define vmul_lane_s32(__p0_98, __p1_98, __p2_98) __extension__ ({ \
16594  int32x2_t __s0_98 = __p0_98; \
16595  int32x2_t __s1_98 = __p1_98; \
16596  int32x2_t __ret_98; \
16597  __ret_98 = __s0_98 * splat_lane_s32(__s1_98, __p2_98); \
16598  __ret_98; \
16599 })
16600 #else
16601 #define vmul_lane_s32(__p0_99, __p1_99, __p2_99) __extension__ ({ \
16602  int32x2_t __s0_99 = __p0_99; \
16603  int32x2_t __s1_99 = __p1_99; \
16604  int32x2_t __rev0_99; __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
16605  int32x2_t __rev1_99; __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
16606  int32x2_t __ret_99; \
16607  __ret_99 = __rev0_99 * __noswap_splat_lane_s32(__rev1_99, __p2_99); \
16608  __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
16609  __ret_99; \
16610 })
16611 #endif
16612 
16613 #ifdef __LITTLE_ENDIAN__
16614 #define vmul_lane_s16(__p0_100, __p1_100, __p2_100) __extension__ ({ \
16615  int16x4_t __s0_100 = __p0_100; \
16616  int16x4_t __s1_100 = __p1_100; \
16617  int16x4_t __ret_100; \
16618  __ret_100 = __s0_100 * splat_lane_s16(__s1_100, __p2_100); \
16619  __ret_100; \
16620 })
16621 #else
16622 #define vmul_lane_s16(__p0_101, __p1_101, __p2_101) __extension__ ({ \
16623  int16x4_t __s0_101 = __p0_101; \
16624  int16x4_t __s1_101 = __p1_101; \
16625  int16x4_t __rev0_101; __rev0_101 = __builtin_shufflevector(__s0_101, __s0_101, 3, 2, 1, 0); \
16626  int16x4_t __rev1_101; __rev1_101 = __builtin_shufflevector(__s1_101, __s1_101, 3, 2, 1, 0); \
16627  int16x4_t __ret_101; \
16628  __ret_101 = __rev0_101 * __noswap_splat_lane_s16(__rev1_101, __p2_101); \
16629  __ret_101 = __builtin_shufflevector(__ret_101, __ret_101, 3, 2, 1, 0); \
16630  __ret_101; \
16631 })
16632 #endif
16633 
16634 #ifdef __LITTLE_ENDIAN__
16635 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
16636  uint32x4_t __ret;
16637  __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
16638  return __ret;
16639 }
16640 #else
16641 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
16642  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16643  uint32x4_t __ret;
16644  __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
16645  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16646  return __ret;
16647 }
16648 #endif
16649 
16650 #ifdef __LITTLE_ENDIAN__
16651 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
16652  uint16x8_t __ret;
16653  __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16654  return __ret;
16655 }
16656 #else
16657 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
16658  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16659  uint16x8_t __ret;
16660  __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16661  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16662  return __ret;
16663 }
16664 #endif
16665 
16666 #ifdef __LITTLE_ENDIAN__
16667 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
16668  float32x4_t __ret;
16669  __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
16670  return __ret;
16671 }
16672 #else
16673 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
16674  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16675  float32x4_t __ret;
16676  __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
16677  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16678  return __ret;
16679 }
16680 #endif
16681 
16682 #ifdef __LITTLE_ENDIAN__
16683 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
16684  int32x4_t __ret;
16685  __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
16686  return __ret;
16687 }
16688 #else
16689 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
16690  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16691  int32x4_t __ret;
16692  __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
16693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16694  return __ret;
16695 }
16696 #endif
16697 
16698 #ifdef __LITTLE_ENDIAN__
16699 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
16700  int16x8_t __ret;
16701  __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16702  return __ret;
16703 }
16704 #else
16705 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
16706  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16707  int16x8_t __ret;
16708  __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
16709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16710  return __ret;
16711 }
16712 #endif
16713 
16714 #ifdef __LITTLE_ENDIAN__
16715 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
16716  uint32x2_t __ret;
16717  __ret = __p0 * (uint32x2_t) {__p1, __p1};
16718  return __ret;
16719 }
16720 #else
16721 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
16722  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16723  uint32x2_t __ret;
16724  __ret = __rev0 * (uint32x2_t) {__p1, __p1};
16725  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16726  return __ret;
16727 }
16728 #endif
16729 
16730 #ifdef __LITTLE_ENDIAN__
16731 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
16732  uint16x4_t __ret;
16733  __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
16734  return __ret;
16735 }
16736 #else
16737 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
16738  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16739  uint16x4_t __ret;
16740  __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
16741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16742  return __ret;
16743 }
16744 #endif
16745 
16746 #ifdef __LITTLE_ENDIAN__
16747 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
16748  float32x2_t __ret;
16749  __ret = __p0 * (float32x2_t) {__p1, __p1};
16750  return __ret;
16751 }
16752 #else
16753 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
16754  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16755  float32x2_t __ret;
16756  __ret = __rev0 * (float32x2_t) {__p1, __p1};
16757  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16758  return __ret;
16759 }
16760 #endif
16761 
16762 #ifdef __LITTLE_ENDIAN__
16763 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
16764  int32x2_t __ret;
16765  __ret = __p0 * (int32x2_t) {__p1, __p1};
16766  return __ret;
16767 }
16768 #else
16769 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
16770  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16771  int32x2_t __ret;
16772  __ret = __rev0 * (int32x2_t) {__p1, __p1};
16773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16774  return __ret;
16775 }
16776 #endif
16777 
16778 #ifdef __LITTLE_ENDIAN__
16779 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
16780  int16x4_t __ret;
16781  __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
16782  return __ret;
16783 }
16784 #else
16785 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
16786  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16787  int16x4_t __ret;
16788  __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
16789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16790  return __ret;
16791 }
16792 #endif
16793 
16794 #ifdef __LITTLE_ENDIAN__
16795 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16796  poly16x8_t __ret;
16797  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
16798  return __ret;
16799 }
16800 #else
16801 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16802  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16803  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16804  poly16x8_t __ret;
16805  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
16806  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16807  return __ret;
16808 }
16809 __ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
16810  poly16x8_t __ret;
16811  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
16812  return __ret;
16813 }
16814 #endif
16815 
16816 #ifdef __LITTLE_ENDIAN__
16817 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16818  uint16x8_t __ret;
16819  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
16820  return __ret;
16821 }
16822 #else
16823 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16824  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16825  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16826  uint16x8_t __ret;
16827  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
16828  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16829  return __ret;
16830 }
16831 __ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
16832  uint16x8_t __ret;
16833  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
16834  return __ret;
16835 }
16836 #endif
16837 
16838 #ifdef __LITTLE_ENDIAN__
16839 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16840  uint64x2_t __ret;
16841  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
16842  return __ret;
16843 }
16844 #else
16845 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16846  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16847  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16848  uint64x2_t __ret;
16849  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
16850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16851  return __ret;
16852 }
16853 __ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
16854  uint64x2_t __ret;
16855  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
16856  return __ret;
16857 }
16858 #endif
16859 
16860 #ifdef __LITTLE_ENDIAN__
16861 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16862  uint32x4_t __ret;
16863  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
16864  return __ret;
16865 }
16866 #else
16867 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16868  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16869  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16870  uint32x4_t __ret;
16871  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
16872  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16873  return __ret;
16874 }
16875 __ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
16876  uint32x4_t __ret;
16877  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
16878  return __ret;
16879 }
16880 #endif
16881 
16882 #ifdef __LITTLE_ENDIAN__
16883 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16884  int16x8_t __ret;
16885  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
16886  return __ret;
16887 }
16888 #else
16889 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16890  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16891  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16892  int16x8_t __ret;
16893  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
16894  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16895  return __ret;
16896 }
16897 __ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
16898  int16x8_t __ret;
16899  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
16900  return __ret;
16901 }
16902 #endif
16903 
16904 #ifdef __LITTLE_ENDIAN__
16905 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16906  int64x2_t __ret;
16907  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
16908  return __ret;
16909 }
16910 #else
16911 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16912  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16913  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16914  int64x2_t __ret;
16915  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
16916  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16917  return __ret;
16918 }
16919 __ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
16920  int64x2_t __ret;
16921  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
16922  return __ret;
16923 }
16924 #endif
16925 
16926 #ifdef __LITTLE_ENDIAN__
16927 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16928  int32x4_t __ret;
16929  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
16930  return __ret;
16931 }
16932 #else
16933 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16934  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16935  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16936  int32x4_t __ret;
16937  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
16938  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16939  return __ret;
16940 }
16941 __ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
16942  int32x4_t __ret;
16943  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
16944  return __ret;
16945 }
16946 #endif
16947 
16948 #ifdef __LITTLE_ENDIAN__
16949 #define vmull_lane_u32(__p0_102, __p1_102, __p2_102) __extension__ ({ \
16950  uint32x2_t __s0_102 = __p0_102; \
16951  uint32x2_t __s1_102 = __p1_102; \
16952  uint64x2_t __ret_102; \
16953  __ret_102 = vmull_u32(__s0_102, splat_lane_u32(__s1_102, __p2_102)); \
16954  __ret_102; \
16955 })
16956 #else
16957 #define vmull_lane_u32(__p0_103, __p1_103, __p2_103) __extension__ ({ \
16958  uint32x2_t __s0_103 = __p0_103; \
16959  uint32x2_t __s1_103 = __p1_103; \
16960  uint32x2_t __rev0_103; __rev0_103 = __builtin_shufflevector(__s0_103, __s0_103, 1, 0); \
16961  uint32x2_t __rev1_103; __rev1_103 = __builtin_shufflevector(__s1_103, __s1_103, 1, 0); \
16962  uint64x2_t __ret_103; \
16963  __ret_103 = __noswap_vmull_u32(__rev0_103, __noswap_splat_lane_u32(__rev1_103, __p2_103)); \
16964  __ret_103 = __builtin_shufflevector(__ret_103, __ret_103, 1, 0); \
16965  __ret_103; \
16966 })
16967 #endif
16968 
16969 #ifdef __LITTLE_ENDIAN__
16970 #define vmull_lane_u16(__p0_104, __p1_104, __p2_104) __extension__ ({ \
16971  uint16x4_t __s0_104 = __p0_104; \
16972  uint16x4_t __s1_104 = __p1_104; \
16973  uint32x4_t __ret_104; \
16974  __ret_104 = vmull_u16(__s0_104, splat_lane_u16(__s1_104, __p2_104)); \
16975  __ret_104; \
16976 })
16977 #else
16978 #define vmull_lane_u16(__p0_105, __p1_105, __p2_105) __extension__ ({ \
16979  uint16x4_t __s0_105 = __p0_105; \
16980  uint16x4_t __s1_105 = __p1_105; \
16981  uint16x4_t __rev0_105; __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 3, 2, 1, 0); \
16982  uint16x4_t __rev1_105; __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 3, 2, 1, 0); \
16983  uint32x4_t __ret_105; \
16984  __ret_105 = __noswap_vmull_u16(__rev0_105, __noswap_splat_lane_u16(__rev1_105, __p2_105)); \
16985  __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 3, 2, 1, 0); \
16986  __ret_105; \
16987 })
16988 #endif
16989 
16990 #ifdef __LITTLE_ENDIAN__
16991 #define vmull_lane_s32(__p0_106, __p1_106, __p2_106) __extension__ ({ \
16992  int32x2_t __s0_106 = __p0_106; \
16993  int32x2_t __s1_106 = __p1_106; \
16994  int64x2_t __ret_106; \
16995  __ret_106 = vmull_s32(__s0_106, splat_lane_s32(__s1_106, __p2_106)); \
16996  __ret_106; \
16997 })
16998 #else
16999 #define vmull_lane_s32(__p0_107, __p1_107, __p2_107) __extension__ ({ \
17000  int32x2_t __s0_107 = __p0_107; \
17001  int32x2_t __s1_107 = __p1_107; \
17002  int32x2_t __rev0_107; __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 1, 0); \
17003  int32x2_t __rev1_107; __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 1, 0); \
17004  int64x2_t __ret_107; \
17005  __ret_107 = __noswap_vmull_s32(__rev0_107, __noswap_splat_lane_s32(__rev1_107, __p2_107)); \
17006  __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 1, 0); \
17007  __ret_107; \
17008 })
17009 #endif
17010 
17011 #ifdef __LITTLE_ENDIAN__
17012 #define vmull_lane_s16(__p0_108, __p1_108, __p2_108) __extension__ ({ \
17013  int16x4_t __s0_108 = __p0_108; \
17014  int16x4_t __s1_108 = __p1_108; \
17015  int32x4_t __ret_108; \
17016  __ret_108 = vmull_s16(__s0_108, splat_lane_s16(__s1_108, __p2_108)); \
17017  __ret_108; \
17018 })
17019 #else
17020 #define vmull_lane_s16(__p0_109, __p1_109, __p2_109) __extension__ ({ \
17021  int16x4_t __s0_109 = __p0_109; \
17022  int16x4_t __s1_109 = __p1_109; \
17023  int16x4_t __rev0_109; __rev0_109 = __builtin_shufflevector(__s0_109, __s0_109, 3, 2, 1, 0); \
17024  int16x4_t __rev1_109; __rev1_109 = __builtin_shufflevector(__s1_109, __s1_109, 3, 2, 1, 0); \
17025  int32x4_t __ret_109; \
17026  __ret_109 = __noswap_vmull_s16(__rev0_109, __noswap_splat_lane_s16(__rev1_109, __p2_109)); \
17027  __ret_109 = __builtin_shufflevector(__ret_109, __ret_109, 3, 2, 1, 0); \
17028  __ret_109; \
17029 })
17030 #endif
17031 
17032 #ifdef __LITTLE_ENDIAN__
17033 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
17034  uint64x2_t __ret;
17035  __ret = vmull_u32(__p0, (uint32x2_t) {__p1, __p1});
17036  return __ret;
17037 }
17038 #else
17039 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
17040  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17041  uint64x2_t __ret;
17042  __ret = __noswap_vmull_u32(__rev0, (uint32x2_t) {__p1, __p1});
17043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17044  return __ret;
17045 }
17046 __ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
17047  uint64x2_t __ret;
17048  __ret = __noswap_vmull_u32(__p0, (uint32x2_t) {__p1, __p1});
17049  return __ret;
17050 }
17051 #endif
17052 
17053 #ifdef __LITTLE_ENDIAN__
17054 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
17055  uint32x4_t __ret;
17056  __ret = vmull_u16(__p0, (uint16x4_t) {__p1, __p1, __p1, __p1});
17057  return __ret;
17058 }
17059 #else
17060 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
17061  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17062  uint32x4_t __ret;
17063  __ret = __noswap_vmull_u16(__rev0, (uint16x4_t) {__p1, __p1, __p1, __p1});
17064  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17065  return __ret;
17066 }
17067 __ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
17068  uint32x4_t __ret;
17069  __ret = __noswap_vmull_u16(__p0, (uint16x4_t) {__p1, __p1, __p1, __p1});
17070  return __ret;
17071 }
17072 #endif
17073 
17074 #ifdef __LITTLE_ENDIAN__
17075 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
17076  int64x2_t __ret;
17077  __ret = vmull_s32(__p0, (int32x2_t) {__p1, __p1});
17078  return __ret;
17079 }
17080 #else
17081 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
17082  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17083  int64x2_t __ret;
17084  __ret = __noswap_vmull_s32(__rev0, (int32x2_t) {__p1, __p1});
17085  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17086  return __ret;
17087 }
17088 __ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
17089  int64x2_t __ret;
17090  __ret = __noswap_vmull_s32(__p0, (int32x2_t) {__p1, __p1});
17091  return __ret;
17092 }
17093 #endif
17094 
17095 #ifdef __LITTLE_ENDIAN__
17096 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
17097  int32x4_t __ret;
17098  __ret = vmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
17099  return __ret;
17100 }
17101 #else
17102 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
17103  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17104  int32x4_t __ret;
17105  __ret = __noswap_vmull_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
17106  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17107  return __ret;
17108 }
17109 __ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
17110  int32x4_t __ret;
17111  __ret = __noswap_vmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
17112  return __ret;
17113 }
17114 #endif
17115 
17116 #ifdef __LITTLE_ENDIAN__
17117 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
17118  poly8x8_t __ret;
17119  __ret = ~__p0;
17120  return __ret;
17121 }
17122 #else
17123 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
17124  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17125  poly8x8_t __ret;
17126  __ret = ~__rev0;
17127  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17128  return __ret;
17129 }
17130 #endif
17131 
17132 #ifdef __LITTLE_ENDIAN__
17133 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
17134  poly8x16_t __ret;
17135  __ret = ~__p0;
17136  return __ret;
17137 }
17138 #else
17139 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
17140  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17141  poly8x16_t __ret;
17142  __ret = ~__rev0;
17143  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17144  return __ret;
17145 }
17146 #endif
17147 
17148 #ifdef __LITTLE_ENDIAN__
17149 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
17150  uint8x16_t __ret;
17151  __ret = ~__p0;
17152  return __ret;
17153 }
17154 #else
17155 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
17156  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17157  uint8x16_t __ret;
17158  __ret = ~__rev0;
17159  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17160  return __ret;
17161 }
17162 #endif
17163 
17164 #ifdef __LITTLE_ENDIAN__
17165 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
17166  uint32x4_t __ret;
17167  __ret = ~__p0;
17168  return __ret;
17169 }
17170 #else
17171 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
17172  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17173  uint32x4_t __ret;
17174  __ret = ~__rev0;
17175  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17176  return __ret;
17177 }
17178 #endif
17179 
17180 #ifdef __LITTLE_ENDIAN__
17181 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
17182  uint16x8_t __ret;
17183  __ret = ~__p0;
17184  return __ret;
17185 }
17186 #else
17187 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
17188  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17189  uint16x8_t __ret;
17190  __ret = ~__rev0;
17191  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17192  return __ret;
17193 }
17194 #endif
17195 
17196 #ifdef __LITTLE_ENDIAN__
17197 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
17198  int8x16_t __ret;
17199  __ret = ~__p0;
17200  return __ret;
17201 }
17202 #else
17203 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
17204  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17205  int8x16_t __ret;
17206  __ret = ~__rev0;
17207  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17208  return __ret;
17209 }
17210 #endif
17211 
17212 #ifdef __LITTLE_ENDIAN__
17213 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
17214  int32x4_t __ret;
17215  __ret = ~__p0;
17216  return __ret;
17217 }
17218 #else
17219 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
17220  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17221  int32x4_t __ret;
17222  __ret = ~__rev0;
17223  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17224  return __ret;
17225 }
17226 #endif
17227 
17228 #ifdef __LITTLE_ENDIAN__
17229 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
17230  int16x8_t __ret;
17231  __ret = ~__p0;
17232  return __ret;
17233 }
17234 #else
17235 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
17236  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17237  int16x8_t __ret;
17238  __ret = ~__rev0;
17239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17240  return __ret;
17241 }
17242 #endif
17243 
17244 #ifdef __LITTLE_ENDIAN__
17245 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
17246  uint8x8_t __ret;
17247  __ret = ~__p0;
17248  return __ret;
17249 }
17250 #else
17251 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
17252  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17253  uint8x8_t __ret;
17254  __ret = ~__rev0;
17255  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17256  return __ret;
17257 }
17258 #endif
17259 
17260 #ifdef __LITTLE_ENDIAN__
17261 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
17262  uint32x2_t __ret;
17263  __ret = ~__p0;
17264  return __ret;
17265 }
17266 #else
17267 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
17268  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17269  uint32x2_t __ret;
17270  __ret = ~__rev0;
17271  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17272  return __ret;
17273 }
17274 #endif
17275 
17276 #ifdef __LITTLE_ENDIAN__
17277 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
17278  uint16x4_t __ret;
17279  __ret = ~__p0;
17280  return __ret;
17281 }
17282 #else
17283 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
17284  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17285  uint16x4_t __ret;
17286  __ret = ~__rev0;
17287  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17288  return __ret;
17289 }
17290 #endif
17291 
17292 #ifdef __LITTLE_ENDIAN__
17293 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
17294  int8x8_t __ret;
17295  __ret = ~__p0;
17296  return __ret;
17297 }
17298 #else
17299 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
17300  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17301  int8x8_t __ret;
17302  __ret = ~__rev0;
17303  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17304  return __ret;
17305 }
17306 #endif
17307 
17308 #ifdef __LITTLE_ENDIAN__
17309 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
17310  int32x2_t __ret;
17311  __ret = ~__p0;
17312  return __ret;
17313 }
17314 #else
17315 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
17316  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17317  int32x2_t __ret;
17318  __ret = ~__rev0;
17319  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17320  return __ret;
17321 }
17322 #endif
17323 
17324 #ifdef __LITTLE_ENDIAN__
17325 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
17326  int16x4_t __ret;
17327  __ret = ~__p0;
17328  return __ret;
17329 }
17330 #else
17331 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
17332  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17333  int16x4_t __ret;
17334  __ret = ~__rev0;
17335  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17336  return __ret;
17337 }
17338 #endif
17339 
17340 #ifdef __LITTLE_ENDIAN__
17341 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
17342  int8x16_t __ret;
17343  __ret = -__p0;
17344  return __ret;
17345 }
17346 #else
17347 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
17348  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17349  int8x16_t __ret;
17350  __ret = -__rev0;
17351  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17352  return __ret;
17353 }
17354 #endif
17355 
17356 #ifdef __LITTLE_ENDIAN__
17357 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
17358  float32x4_t __ret;
17359  __ret = -__p0;
17360  return __ret;
17361 }
17362 #else
17363 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
17364  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17365  float32x4_t __ret;
17366  __ret = -__rev0;
17367  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17368  return __ret;
17369 }
17370 #endif
17371 
17372 #ifdef __LITTLE_ENDIAN__
17373 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
17374  int32x4_t __ret;
17375  __ret = -__p0;
17376  return __ret;
17377 }
17378 #else
17379 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
17380  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17381  int32x4_t __ret;
17382  __ret = -__rev0;
17383  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17384  return __ret;
17385 }
17386 #endif
17387 
17388 #ifdef __LITTLE_ENDIAN__
17389 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
17390  int16x8_t __ret;
17391  __ret = -__p0;
17392  return __ret;
17393 }
17394 #else
17395 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
17396  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17397  int16x8_t __ret;
17398  __ret = -__rev0;
17399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17400  return __ret;
17401 }
17402 #endif
17403 
17404 #ifdef __LITTLE_ENDIAN__
17405 __ai int8x8_t vneg_s8(int8x8_t __p0) {
17406  int8x8_t __ret;
17407  __ret = -__p0;
17408  return __ret;
17409 }
17410 #else
17411 __ai int8x8_t vneg_s8(int8x8_t __p0) {
17412  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17413  int8x8_t __ret;
17414  __ret = -__rev0;
17415  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17416  return __ret;
17417 }
17418 #endif
17419 
17420 #ifdef __LITTLE_ENDIAN__
17421 __ai float32x2_t vneg_f32(float32x2_t __p0) {
17422  float32x2_t __ret;
17423  __ret = -__p0;
17424  return __ret;
17425 }
17426 #else
17427 __ai float32x2_t vneg_f32(float32x2_t __p0) {
17428  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17429  float32x2_t __ret;
17430  __ret = -__rev0;
17431  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17432  return __ret;
17433 }
17434 #endif
17435 
17436 #ifdef __LITTLE_ENDIAN__
17437 __ai int32x2_t vneg_s32(int32x2_t __p0) {
17438  int32x2_t __ret;
17439  __ret = -__p0;
17440  return __ret;
17441 }
17442 #else
17443 __ai int32x2_t vneg_s32(int32x2_t __p0) {
17444  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17445  int32x2_t __ret;
17446  __ret = -__rev0;
17447  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17448  return __ret;
17449 }
17450 #endif
17451 
17452 #ifdef __LITTLE_ENDIAN__
17453 __ai int16x4_t vneg_s16(int16x4_t __p0) {
17454  int16x4_t __ret;
17455  __ret = -__p0;
17456  return __ret;
17457 }
17458 #else
17459 __ai int16x4_t vneg_s16(int16x4_t __p0) {
17460  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17461  int16x4_t __ret;
17462  __ret = -__rev0;
17463  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17464  return __ret;
17465 }
17466 #endif
17467 
17468 #ifdef __LITTLE_ENDIAN__
17469 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17470  uint8x16_t __ret;
17471  __ret = __p0 | ~__p1;
17472  return __ret;
17473 }
17474 #else
17475 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17476  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17477  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17478  uint8x16_t __ret;
17479  __ret = __rev0 | ~__rev1;
17480  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17481  return __ret;
17482 }
17483 #endif
17484 
17485 #ifdef __LITTLE_ENDIAN__
17486 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17487  uint32x4_t __ret;
17488  __ret = __p0 | ~__p1;
17489  return __ret;
17490 }
17491 #else
17492 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17493  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17494  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17495  uint32x4_t __ret;
17496  __ret = __rev0 | ~__rev1;
17497  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17498  return __ret;
17499 }
17500 #endif
17501 
17502 #ifdef __LITTLE_ENDIAN__
17503 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17504  uint64x2_t __ret;
17505  __ret = __p0 | ~__p1;
17506  return __ret;
17507 }
17508 #else
17509 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17510  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17511  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17512  uint64x2_t __ret;
17513  __ret = __rev0 | ~__rev1;
17514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17515  return __ret;
17516 }
17517 #endif
17518 
17519 #ifdef __LITTLE_ENDIAN__
17520 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17521  uint16x8_t __ret;
17522  __ret = __p0 | ~__p1;
17523  return __ret;
17524 }
17525 #else
17526 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17527  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17528  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17529  uint16x8_t __ret;
17530  __ret = __rev0 | ~__rev1;
17531  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17532  return __ret;
17533 }
17534 #endif
17535 
17536 #ifdef __LITTLE_ENDIAN__
17537 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
17538  int8x16_t __ret;
17539  __ret = __p0 | ~__p1;
17540  return __ret;
17541 }
17542 #else
17543 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
17544  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17545  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17546  int8x16_t __ret;
17547  __ret = __rev0 | ~__rev1;
17548  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17549  return __ret;
17550 }
17551 #endif
17552 
17553 #ifdef __LITTLE_ENDIAN__
17554 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
17555  int32x4_t __ret;
17556  __ret = __p0 | ~__p1;
17557  return __ret;
17558 }
17559 #else
17560 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
17561  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17562  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17563  int32x4_t __ret;
17564  __ret = __rev0 | ~__rev1;
17565  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17566  return __ret;
17567 }
17568 #endif
17569 
17570 #ifdef __LITTLE_ENDIAN__
17571 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
17572  int64x2_t __ret;
17573  __ret = __p0 | ~__p1;
17574  return __ret;
17575 }
17576 #else
17577 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
17578  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17579  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17580  int64x2_t __ret;
17581  __ret = __rev0 | ~__rev1;
17582  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17583  return __ret;
17584 }
17585 #endif
17586 
17587 #ifdef __LITTLE_ENDIAN__
17588 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
17589  int16x8_t __ret;
17590  __ret = __p0 | ~__p1;
17591  return __ret;
17592 }
17593 #else
17594 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
17595  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17596  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17597  int16x8_t __ret;
17598  __ret = __rev0 | ~__rev1;
17599  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17600  return __ret;
17601 }
17602 #endif
17603 
17604 #ifdef __LITTLE_ENDIAN__
17605 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
17606  uint8x8_t __ret;
17607  __ret = __p0 | ~__p1;
17608  return __ret;
17609 }
17610 #else
17611 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
17612  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17613  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17614  uint8x8_t __ret;
17615  __ret = __rev0 | ~__rev1;
17616  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17617  return __ret;
17618 }
17619 #endif
17620 
17621 #ifdef __LITTLE_ENDIAN__
17622 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
17623  uint32x2_t __ret;
17624  __ret = __p0 | ~__p1;
17625  return __ret;
17626 }
17627 #else
17628 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
17629  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17630  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17631  uint32x2_t __ret;
17632  __ret = __rev0 | ~__rev1;
17633  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17634  return __ret;
17635 }
17636 #endif
17637 
17638 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
17639  uint64x1_t __ret;
17640  __ret = __p0 | ~__p1;
17641  return __ret;
17642 }
17643 #ifdef __LITTLE_ENDIAN__
17644 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
17645  uint16x4_t __ret;
17646  __ret = __p0 | ~__p1;
17647  return __ret;
17648 }
17649 #else
17650 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
17651  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17652  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17653  uint16x4_t __ret;
17654  __ret = __rev0 | ~__rev1;
17655  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17656  return __ret;
17657 }
17658 #endif
17659 
17660 #ifdef __LITTLE_ENDIAN__
17661 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
17662  int8x8_t __ret;
17663  __ret = __p0 | ~__p1;
17664  return __ret;
17665 }
17666 #else
17667 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
17668  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17669  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17670  int8x8_t __ret;
17671  __ret = __rev0 | ~__rev1;
17672  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17673  return __ret;
17674 }
17675 #endif
17676 
17677 #ifdef __LITTLE_ENDIAN__
17678 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
17679  int32x2_t __ret;
17680  __ret = __p0 | ~__p1;
17681  return __ret;
17682 }
17683 #else
17684 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
17685  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17686  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17687  int32x2_t __ret;
17688  __ret = __rev0 | ~__rev1;
17689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17690  return __ret;
17691 }
17692 #endif
17693 
17694 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
17695  int64x1_t __ret;
17696  __ret = __p0 | ~__p1;
17697  return __ret;
17698 }
17699 #ifdef __LITTLE_ENDIAN__
17700 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
17701  int16x4_t __ret;
17702  __ret = __p0 | ~__p1;
17703  return __ret;
17704 }
17705 #else
17706 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
17707  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17708  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17709  int16x4_t __ret;
17710  __ret = __rev0 | ~__rev1;
17711  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17712  return __ret;
17713 }
17714 #endif
17715 
17716 #ifdef __LITTLE_ENDIAN__
17717 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17718  uint8x16_t __ret;
17719  __ret = __p0 | __p1;
17720  return __ret;
17721 }
17722 #else
17723 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
17724  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17725  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17726  uint8x16_t __ret;
17727  __ret = __rev0 | __rev1;
17728  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17729  return __ret;
17730 }
17731 #endif
17732 
17733 #ifdef __LITTLE_ENDIAN__
17734 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17735  uint32x4_t __ret;
17736  __ret = __p0 | __p1;
17737  return __ret;
17738 }
17739 #else
17740 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
17741  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17742  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17743  uint32x4_t __ret;
17744  __ret = __rev0 | __rev1;
17745  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17746  return __ret;
17747 }
17748 #endif
17749 
17750 #ifdef __LITTLE_ENDIAN__
17751 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17752  uint64x2_t __ret;
17753  __ret = __p0 | __p1;
17754  return __ret;
17755 }
17756 #else
17757 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
17758  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17759  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17760  uint64x2_t __ret;
17761  __ret = __rev0 | __rev1;
17762  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17763  return __ret;
17764 }
17765 #endif
17766 
17767 #ifdef __LITTLE_ENDIAN__
17768 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17769  uint16x8_t __ret;
17770  __ret = __p0 | __p1;
17771  return __ret;
17772 }
17773 #else
17774 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
17775  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17776  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17777  uint16x8_t __ret;
17778  __ret = __rev0 | __rev1;
17779  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17780  return __ret;
17781 }
17782 #endif
17783 
17784 #ifdef __LITTLE_ENDIAN__
17785 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
17786  int8x16_t __ret;
17787  __ret = __p0 | __p1;
17788  return __ret;
17789 }
17790 #else
17791 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
17792  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17793  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17794  int8x16_t __ret;
17795  __ret = __rev0 | __rev1;
17796  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17797  return __ret;
17798 }
17799 #endif
17800 
17801 #ifdef __LITTLE_ENDIAN__
17802 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
17803  int32x4_t __ret;
17804  __ret = __p0 | __p1;
17805  return __ret;
17806 }
17807 #else
17808 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
17809  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17810  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17811  int32x4_t __ret;
17812  __ret = __rev0 | __rev1;
17813  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17814  return __ret;
17815 }
17816 #endif
17817 
17818 #ifdef __LITTLE_ENDIAN__
17819 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
17820  int64x2_t __ret;
17821  __ret = __p0 | __p1;
17822  return __ret;
17823 }
17824 #else
17825 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
17826  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17827  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17828  int64x2_t __ret;
17829  __ret = __rev0 | __rev1;
17830  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17831  return __ret;
17832 }
17833 #endif
17834 
17835 #ifdef __LITTLE_ENDIAN__
17836 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
17837  int16x8_t __ret;
17838  __ret = __p0 | __p1;
17839  return __ret;
17840 }
17841 #else
17842 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
17843  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17844  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17845  int16x8_t __ret;
17846  __ret = __rev0 | __rev1;
17847  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17848  return __ret;
17849 }
17850 #endif
17851 
17852 #ifdef __LITTLE_ENDIAN__
17853 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
17854  uint8x8_t __ret;
17855  __ret = __p0 | __p1;
17856  return __ret;
17857 }
17858 #else
17859 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
17860  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17861  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17862  uint8x8_t __ret;
17863  __ret = __rev0 | __rev1;
17864  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17865  return __ret;
17866 }
17867 #endif
17868 
17869 #ifdef __LITTLE_ENDIAN__
17870 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
17871  uint32x2_t __ret;
17872  __ret = __p0 | __p1;
17873  return __ret;
17874 }
17875 #else
17876 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
17877  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17878  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17879  uint32x2_t __ret;
17880  __ret = __rev0 | __rev1;
17881  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17882  return __ret;
17883 }
17884 #endif
17885 
17886 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
17887  uint64x1_t __ret;
17888  __ret = __p0 | __p1;
17889  return __ret;
17890 }
17891 #ifdef __LITTLE_ENDIAN__
17892 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
17893  uint16x4_t __ret;
17894  __ret = __p0 | __p1;
17895  return __ret;
17896 }
17897 #else
17898 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
17899  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17900  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17901  uint16x4_t __ret;
17902  __ret = __rev0 | __rev1;
17903  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17904  return __ret;
17905 }
17906 #endif
17907 
17908 #ifdef __LITTLE_ENDIAN__
17909 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
17910  int8x8_t __ret;
17911  __ret = __p0 | __p1;
17912  return __ret;
17913 }
17914 #else
17915 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
17916  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17917  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17918  int8x8_t __ret;
17919  __ret = __rev0 | __rev1;
17920  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17921  return __ret;
17922 }
17923 #endif
17924 
17925 #ifdef __LITTLE_ENDIAN__
17926 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
17927  int32x2_t __ret;
17928  __ret = __p0 | __p1;
17929  return __ret;
17930 }
17931 #else
17932 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
17933  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17934  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17935  int32x2_t __ret;
17936  __ret = __rev0 | __rev1;
17937  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17938  return __ret;
17939 }
17940 #endif
17941 
17942 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
17943  int64x1_t __ret;
17944  __ret = __p0 | __p1;
17945  return __ret;
17946 }
17947 #ifdef __LITTLE_ENDIAN__
17948 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
17949  int16x4_t __ret;
17950  __ret = __p0 | __p1;
17951  return __ret;
17952 }
17953 #else
17954 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
17955  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17956  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17957  int16x4_t __ret;
17958  __ret = __rev0 | __rev1;
17959  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17960  return __ret;
17961 }
17962 #endif
17963 
17964 #ifdef __LITTLE_ENDIAN__
17965 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
17966  uint16x8_t __ret;
17967  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
17968  return __ret;
17969 }
17970 #else
17971 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
17972  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17973  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17974  uint16x8_t __ret;
17975  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
17976  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17977  return __ret;
17978 }
17979 #endif
17980 
17981 #ifdef __LITTLE_ENDIAN__
17982 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
17983  uint64x2_t __ret;
17984  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
17985  return __ret;
17986 }
17987 #else
17988 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
17989  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17990  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17991  uint64x2_t __ret;
17992  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
17993  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17994  return __ret;
17995 }
17996 #endif
17997 
17998 #ifdef __LITTLE_ENDIAN__
17999 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
18000  uint32x4_t __ret;
18001  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18002  return __ret;
18003 }
18004 #else
18005 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
18006  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18007  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18008  uint32x4_t __ret;
18009  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18010  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18011  return __ret;
18012 }
18013 #endif
18014 
18015 #ifdef __LITTLE_ENDIAN__
18016 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
18017  int16x8_t __ret;
18018  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18019  return __ret;
18020 }
18021 #else
18022 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
18023  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18024  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18025  int16x8_t __ret;
18026  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18027  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18028  return __ret;
18029 }
18030 #endif
18031 
18032 #ifdef __LITTLE_ENDIAN__
18033 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
18034  int64x2_t __ret;
18035  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18036  return __ret;
18037 }
18038 #else
18039 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
18040  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18041  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18042  int64x2_t __ret;
18043  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18044  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18045  return __ret;
18046 }
18047 #endif
18048 
18049 #ifdef __LITTLE_ENDIAN__
18050 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
18051  int32x4_t __ret;
18052  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18053  return __ret;
18054 }
18055 #else
18056 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
18057  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18058  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18059  int32x4_t __ret;
18060  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18062  return __ret;
18063 }
18064 #endif
18065 
18066 #ifdef __LITTLE_ENDIAN__
18067 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
18068  uint16x4_t __ret;
18069  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18070  return __ret;
18071 }
18072 #else
18073 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
18074  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18075  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18076  uint16x4_t __ret;
18077  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18078  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18079  return __ret;
18080 }
18081 #endif
18082 
18083 #ifdef __LITTLE_ENDIAN__
18084 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
18085  uint64x1_t __ret;
18086  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18087  return __ret;
18088 }
18089 #else
18090 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
18091  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18092  uint64x1_t __ret;
18093  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
18094  return __ret;
18095 }
18096 #endif
18097 
18098 #ifdef __LITTLE_ENDIAN__
18099 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
18100  uint32x2_t __ret;
18101  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18102  return __ret;
18103 }
18104 #else
18105 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
18106  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18107  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18108  uint32x2_t __ret;
18109  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18110  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18111  return __ret;
18112 }
18113 #endif
18114 
18115 #ifdef __LITTLE_ENDIAN__
18116 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
18117  int16x4_t __ret;
18118  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18119  return __ret;
18120 }
18121 #else
18122 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
18123  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18124  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18125  int16x4_t __ret;
18126  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18127  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18128  return __ret;
18129 }
18130 #endif
18131 
18132 #ifdef __LITTLE_ENDIAN__
18133 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
18134  int64x1_t __ret;
18135  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18136  return __ret;
18137 }
18138 #else
18139 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
18140  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18141  int64x1_t __ret;
18142  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
18143  return __ret;
18144 }
18145 #endif
18146 
18147 #ifdef __LITTLE_ENDIAN__
18148 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
18149  int32x2_t __ret;
18150  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18151  return __ret;
18152 }
18153 #else
18154 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
18155  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18156  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18157  int32x2_t __ret;
18158  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18159  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18160  return __ret;
18161 }
18162 #endif
18163 
18164 #ifdef __LITTLE_ENDIAN__
18165 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18166  uint8x8_t __ret;
18167  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18168  return __ret;
18169 }
18170 #else
18171 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18172  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18173  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18174  uint8x8_t __ret;
18175  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18177  return __ret;
18178 }
18179 #endif
18180 
18181 #ifdef __LITTLE_ENDIAN__
18182 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18183  uint32x2_t __ret;
18184  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18185  return __ret;
18186 }
18187 #else
18188 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18189  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18190  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18191  uint32x2_t __ret;
18192  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18193  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18194  return __ret;
18195 }
18196 #endif
18197 
18198 #ifdef __LITTLE_ENDIAN__
18199 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18200  uint16x4_t __ret;
18201  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18202  return __ret;
18203 }
18204 #else
18205 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18206  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18207  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18208  uint16x4_t __ret;
18209  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18210  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18211  return __ret;
18212 }
18213 #endif
18214 
18215 #ifdef __LITTLE_ENDIAN__
18216 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
18217  int8x8_t __ret;
18218  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18219  return __ret;
18220 }
18221 #else
18222 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
18223  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18224  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18225  int8x8_t __ret;
18226  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18227  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18228  return __ret;
18229 }
18230 #endif
18231 
18232 #ifdef __LITTLE_ENDIAN__
18233 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
18234  float32x2_t __ret;
18235  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18236  return __ret;
18237 }
18238 #else
18239 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
18240  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18241  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18242  float32x2_t __ret;
18243  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18244  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18245  return __ret;
18246 }
18247 #endif
18248 
18249 #ifdef __LITTLE_ENDIAN__
18250 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
18251  int32x2_t __ret;
18252  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18253  return __ret;
18254 }
18255 #else
18256 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
18257  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18258  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18259  int32x2_t __ret;
18260  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18261  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18262  return __ret;
18263 }
18264 #endif
18265 
18266 #ifdef __LITTLE_ENDIAN__
18267 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
18268  int16x4_t __ret;
18269  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18270  return __ret;
18271 }
18272 #else
18273 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
18274  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18275  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18276  int16x4_t __ret;
18277  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18278  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18279  return __ret;
18280 }
18281 #endif
18282 
18283 #ifdef __LITTLE_ENDIAN__
18284 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
18285  uint16x8_t __ret;
18286  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
18287  return __ret;
18288 }
18289 #else
18290 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
18291  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18292  uint16x8_t __ret;
18293  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
18294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18295  return __ret;
18296 }
18297 #endif
18298 
18299 #ifdef __LITTLE_ENDIAN__
18300 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
18301  uint64x2_t __ret;
18302  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
18303  return __ret;
18304 }
18305 #else
18306 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
18307  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18308  uint64x2_t __ret;
18309  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
18310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18311  return __ret;
18312 }
18313 #endif
18314 
18315 #ifdef __LITTLE_ENDIAN__
18316 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
18317  uint32x4_t __ret;
18318  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
18319  return __ret;
18320 }
18321 #else
18322 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
18323  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18324  uint32x4_t __ret;
18325  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
18326  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18327  return __ret;
18328 }
18329 #endif
18330 
18331 #ifdef __LITTLE_ENDIAN__
18332 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
18333  int16x8_t __ret;
18334  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
18335  return __ret;
18336 }
18337 #else
18338 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
18339  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18340  int16x8_t __ret;
18341  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
18342  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18343  return __ret;
18344 }
18345 #endif
18346 
18347 #ifdef __LITTLE_ENDIAN__
18348 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
18349  int64x2_t __ret;
18350  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
18351  return __ret;
18352 }
18353 #else
18354 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
18355  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18356  int64x2_t __ret;
18357  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
18358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18359  return __ret;
18360 }
18361 #endif
18362 
18363 #ifdef __LITTLE_ENDIAN__
18364 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
18365  int32x4_t __ret;
18366  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
18367  return __ret;
18368 }
18369 #else
18370 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
18371  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18372  int32x4_t __ret;
18373  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
18374  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18375  return __ret;
18376 }
18377 #endif
18378 
18379 #ifdef __LITTLE_ENDIAN__
18380 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
18381  uint16x4_t __ret;
18382  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
18383  return __ret;
18384 }
18385 #else
18386 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
18387  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18388  uint16x4_t __ret;
18389  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
18390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18391  return __ret;
18392 }
18393 #endif
18394 
18395 #ifdef __LITTLE_ENDIAN__
18396 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
18397  uint64x1_t __ret;
18398  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
18399  return __ret;
18400 }
18401 #else
18402 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
18403  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18404  uint64x1_t __ret;
18405  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
18406  return __ret;
18407 }
18408 #endif
18409 
18410 #ifdef __LITTLE_ENDIAN__
18411 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
18412  uint32x2_t __ret;
18413  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
18414  return __ret;
18415 }
18416 #else
18417 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
18418  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18419  uint32x2_t __ret;
18420  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
18421  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18422  return __ret;
18423 }
18424 #endif
18425 
18426 #ifdef __LITTLE_ENDIAN__
18427 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
18428  int16x4_t __ret;
18429  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
18430  return __ret;
18431 }
18432 #else
18433 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
18434  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18435  int16x4_t __ret;
18436  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
18437  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18438  return __ret;
18439 }
18440 #endif
18441 
18442 #ifdef __LITTLE_ENDIAN__
18443 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
18444  int64x1_t __ret;
18445  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
18446  return __ret;
18447 }
18448 #else
18449 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
18450  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18451  int64x1_t __ret;
18452  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
18453  return __ret;
18454 }
18455 #endif
18456 
18457 #ifdef __LITTLE_ENDIAN__
18458 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
18459  int32x2_t __ret;
18460  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
18461  return __ret;
18462 }
18463 #else
18464 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
18465  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18466  int32x2_t __ret;
18467  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
18468  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18469  return __ret;
18470 }
18471 #endif
18472 
18473 #ifdef __LITTLE_ENDIAN__
18474 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
18475  uint8x8_t __ret;
18476  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18477  return __ret;
18478 }
18479 #else
18480 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
18481  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18482  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18483  uint8x8_t __ret;
18484  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18485  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18486  return __ret;
18487 }
18488 #endif
18489 
18490 #ifdef __LITTLE_ENDIAN__
18491 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
18492  uint32x2_t __ret;
18493  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18494  return __ret;
18495 }
18496 #else
18497 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
18498  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18499  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18500  uint32x2_t __ret;
18501  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18502  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18503  return __ret;
18504 }
18505 #endif
18506 
18507 #ifdef __LITTLE_ENDIAN__
18508 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
18509  uint16x4_t __ret;
18510  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18511  return __ret;
18512 }
18513 #else
18514 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
18515  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18516  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18517  uint16x4_t __ret;
18518  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18519  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18520  return __ret;
18521 }
18522 #endif
18523 
18524 #ifdef __LITTLE_ENDIAN__
18525 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
18526  int8x8_t __ret;
18527  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18528  return __ret;
18529 }
18530 #else
18531 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
18532  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18533  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18534  int8x8_t __ret;
18535  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18536  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18537  return __ret;
18538 }
18539 #endif
18540 
18541 #ifdef __LITTLE_ENDIAN__
18542 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
18543  float32x2_t __ret;
18544  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18545  return __ret;
18546 }
18547 #else
18548 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
18549  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18550  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18551  float32x2_t __ret;
18552  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18553  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18554  return __ret;
18555 }
18556 #endif
18557 
18558 #ifdef __LITTLE_ENDIAN__
18559 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
18560  int32x2_t __ret;
18561  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18562  return __ret;
18563 }
18564 #else
18565 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
18566  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18567  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18568  int32x2_t __ret;
18569  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18570  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18571  return __ret;
18572 }
18573 #endif
18574 
18575 #ifdef __LITTLE_ENDIAN__
18576 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
18577  int16x4_t __ret;
18578  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18579  return __ret;
18580 }
18581 #else
18582 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
18583  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18584  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18585  int16x4_t __ret;
18586  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18587  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18588  return __ret;
18589 }
18590 #endif
18591 
18592 #ifdef __LITTLE_ENDIAN__
18593 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
18594  uint8x8_t __ret;
18595  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18596  return __ret;
18597 }
18598 #else
18599 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
18600  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18601  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18602  uint8x8_t __ret;
18603  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18605  return __ret;
18606 }
18607 #endif
18608 
18609 #ifdef __LITTLE_ENDIAN__
18610 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
18611  uint32x2_t __ret;
18612  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18613  return __ret;
18614 }
18615 #else
18616 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
18617  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18618  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18619  uint32x2_t __ret;
18620  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18621  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18622  return __ret;
18623 }
18624 #endif
18625 
18626 #ifdef __LITTLE_ENDIAN__
18627 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
18628  uint16x4_t __ret;
18629  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18630  return __ret;
18631 }
18632 #else
18633 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
18634  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18635  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18636  uint16x4_t __ret;
18637  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18638  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18639  return __ret;
18640 }
18641 #endif
18642 
18643 #ifdef __LITTLE_ENDIAN__
18644 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
18645  int8x8_t __ret;
18646  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18647  return __ret;
18648 }
18649 #else
18650 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
18651  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18652  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18653  int8x8_t __ret;
18654  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18655  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18656  return __ret;
18657 }
18658 #endif
18659 
18660 #ifdef __LITTLE_ENDIAN__
18661 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
18662  float32x2_t __ret;
18663  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
18664  return __ret;
18665 }
18666 #else
18667 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
18668  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18669  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18670  float32x2_t __ret;
18671  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
18672  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18673  return __ret;
18674 }
18675 #endif
18676 
18677 #ifdef __LITTLE_ENDIAN__
18678 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
18679  int32x2_t __ret;
18680  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18681  return __ret;
18682 }
18683 #else
18684 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
18685  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18686  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18687  int32x2_t __ret;
18688  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18690  return __ret;
18691 }
18692 #endif
18693 
18694 #ifdef __LITTLE_ENDIAN__
18695 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
18696  int16x4_t __ret;
18697  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18698  return __ret;
18699 }
18700 #else
18701 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
18702  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18703  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18704  int16x4_t __ret;
18705  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18706  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18707  return __ret;
18708 }
18709 #endif
18710 
18711 #ifdef __LITTLE_ENDIAN__
18712 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
18713  int8x16_t __ret;
18714  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
18715  return __ret;
18716 }
18717 #else
18718 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
18719  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18720  int8x16_t __ret;
18721  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
18722  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18723  return __ret;
18724 }
18725 #endif
18726 
18727 #ifdef __LITTLE_ENDIAN__
18728 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
18729  int32x4_t __ret;
18730  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
18731  return __ret;
18732 }
18733 #else
18734 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
18735  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18736  int32x4_t __ret;
18737  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
18738  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18739  return __ret;
18740 }
18741 #endif
18742 
18743 #ifdef __LITTLE_ENDIAN__
18744 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
18745  int16x8_t __ret;
18746  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
18747  return __ret;
18748 }
18749 #else
18750 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
18751  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18752  int16x8_t __ret;
18753  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
18754  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18755  return __ret;
18756 }
18757 #endif
18758 
18759 #ifdef __LITTLE_ENDIAN__
18760 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
18761  int8x8_t __ret;
18762  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
18763  return __ret;
18764 }
18765 #else
18766 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
18767  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18768  int8x8_t __ret;
18769  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
18770  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18771  return __ret;
18772 }
18773 #endif
18774 
18775 #ifdef __LITTLE_ENDIAN__
18776 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
18777  int32x2_t __ret;
18778  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
18779  return __ret;
18780 }
18781 #else
18782 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
18783  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18784  int32x2_t __ret;
18785  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
18786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18787  return __ret;
18788 }
18789 #endif
18790 
18791 #ifdef __LITTLE_ENDIAN__
18792 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
18793  int16x4_t __ret;
18794  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
18795  return __ret;
18796 }
18797 #else
18798 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
18799  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18800  int16x4_t __ret;
18801  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
18802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18803  return __ret;
18804 }
18805 #endif
18806 
18807 #ifdef __LITTLE_ENDIAN__
18808 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
18809  uint8x16_t __ret;
18810  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18811  return __ret;
18812 }
18813 #else
18814 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
18815  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18816  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18817  uint8x16_t __ret;
18818  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18819  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18820  return __ret;
18821 }
18822 #endif
18823 
18824 #ifdef __LITTLE_ENDIAN__
18825 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
18826  uint32x4_t __ret;
18827  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18828  return __ret;
18829 }
18830 #else
18831 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
18832  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18833  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18834  uint32x4_t __ret;
18835  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18836  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18837  return __ret;
18838 }
18839 #endif
18840 
18841 #ifdef __LITTLE_ENDIAN__
18842 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
18843  uint64x2_t __ret;
18844  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18845  return __ret;
18846 }
18847 #else
18848 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
18849  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18850  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18851  uint64x2_t __ret;
18852  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18853  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18854  return __ret;
18855 }
18856 #endif
18857 
18858 #ifdef __LITTLE_ENDIAN__
18859 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
18860  uint16x8_t __ret;
18861  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18862  return __ret;
18863 }
18864 #else
18865 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
18866  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18867  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18868  uint16x8_t __ret;
18869  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18870  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18871  return __ret;
18872 }
18873 #endif
18874 
18875 #ifdef __LITTLE_ENDIAN__
18876 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
18877  int8x16_t __ret;
18878  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18879  return __ret;
18880 }
18881 #else
18882 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
18883  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18884  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18885  int8x16_t __ret;
18886  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18887  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18888  return __ret;
18889 }
18890 #endif
18891 
18892 #ifdef __LITTLE_ENDIAN__
18893 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
18894  int32x4_t __ret;
18895  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18896  return __ret;
18897 }
18898 #else
18899 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
18900  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18901  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18902  int32x4_t __ret;
18903  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18904  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18905  return __ret;
18906 }
18907 __ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
18908  int32x4_t __ret;
18909  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18910  return __ret;
18911 }
18912 #endif
18913 
18914 #ifdef __LITTLE_ENDIAN__
18915 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
18916  int64x2_t __ret;
18917  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18918  return __ret;
18919 }
18920 #else
18921 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
18922  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18923  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18924  int64x2_t __ret;
18925  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18927  return __ret;
18928 }
18929 #endif
18930 
18931 #ifdef __LITTLE_ENDIAN__
18932 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
18933  int16x8_t __ret;
18934  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18935  return __ret;
18936 }
18937 #else
18938 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
18939  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18940  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18941  int16x8_t __ret;
18942  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18943  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18944  return __ret;
18945 }
18946 __ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
18947  int16x8_t __ret;
18948  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18949  return __ret;
18950 }
18951 #endif
18952 
18953 #ifdef __LITTLE_ENDIAN__
18954 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18955  uint8x8_t __ret;
18956  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18957  return __ret;
18958 }
18959 #else
18960 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
18961  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18962  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18963  uint8x8_t __ret;
18964  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18965  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18966  return __ret;
18967 }
18968 #endif
18969 
18970 #ifdef __LITTLE_ENDIAN__
18971 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18972  uint32x2_t __ret;
18973  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18974  return __ret;
18975 }
18976 #else
18977 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
18978  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18979  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18980  uint32x2_t __ret;
18981  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18983  return __ret;
18984 }
18985 #endif
18986 
18987 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
18988  uint64x1_t __ret;
18989  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18990  return __ret;
18991 }
18992 #ifdef __LITTLE_ENDIAN__
18993 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
18994  uint16x4_t __ret;
18995  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18996  return __ret;
18997 }
18998 #else
18999 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
19000  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19001  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19002  uint16x4_t __ret;
19003  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
19004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19005  return __ret;
19006 }
19007 #endif
19008 
19009 #ifdef __LITTLE_ENDIAN__
19010 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
19011  int8x8_t __ret;
19012  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
19013  return __ret;
19014 }
19015 #else
19016 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
19017  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19018  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19019  int8x8_t __ret;
19020  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
19021  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19022  return __ret;
19023 }
19024 #endif
19025 
19026 #ifdef __LITTLE_ENDIAN__
19027 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
19028  int32x2_t __ret;
19029  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19030  return __ret;
19031 }
19032 #else
19033 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
19034  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19035  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19036  int32x2_t __ret;
19037  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19039  return __ret;
19040 }
19041 __ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
19042  int32x2_t __ret;
19043  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19044  return __ret;
19045 }
19046 #endif
19047 
19048 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
19049  int64x1_t __ret;
19050  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19051  return __ret;
19052 }
19053 #ifdef __LITTLE_ENDIAN__
19054 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
19055  int16x4_t __ret;
19056  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19057  return __ret;
19058 }
19059 #else
19060 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
19061  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19062  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19063  int16x4_t __ret;
19064  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19065  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19066  return __ret;
19067 }
19068 __ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
19069  int16x4_t __ret;
19070  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19071  return __ret;
19072 }
19073 #endif
19074 
19075 #ifdef __LITTLE_ENDIAN__
19076 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19077  int64x2_t __ret;
19078  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19079  return __ret;
19080 }
19081 #else
19082 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19083  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19084  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19085  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
19086  int64x2_t __ret;
19087  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
19088  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19089  return __ret;
19090 }
19091 __ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19092  int64x2_t __ret;
19093  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19094  return __ret;
19095 }
19096 #endif
19097 
19098 #ifdef __LITTLE_ENDIAN__
19099 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19100  int32x4_t __ret;
19101  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19102  return __ret;
19103 }
19104 #else
19105 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19106  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19107  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19108  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
19109  int32x4_t __ret;
19110  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
19111  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19112  return __ret;
19113 }
19114 __ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19115  int32x4_t __ret;
19116  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19117  return __ret;
19118 }
19119 #endif
19120 
19121 #ifdef __LITTLE_ENDIAN__
19122 #define vqdmlal_lane_s32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
19123  int64x2_t __s0_110 = __p0_110; \
19124  int32x2_t __s1_110 = __p1_110; \
19125  int32x2_t __s2_110 = __p2_110; \
19126  int64x2_t __ret_110; \
19127  __ret_110 = vqdmlal_s32(__s0_110, __s1_110, splat_lane_s32(__s2_110, __p3_110)); \
19128  __ret_110; \
19129 })
19130 #else
19131 #define vqdmlal_lane_s32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
19132  int64x2_t __s0_111 = __p0_111; \
19133  int32x2_t __s1_111 = __p1_111; \
19134  int32x2_t __s2_111 = __p2_111; \
19135  int64x2_t __rev0_111; __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
19136  int32x2_t __rev1_111; __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
19137  int32x2_t __rev2_111; __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 1, 0); \
19138  int64x2_t __ret_111; \
19139  __ret_111 = __noswap_vqdmlal_s32(__rev0_111, __rev1_111, __noswap_splat_lane_s32(__rev2_111, __p3_111)); \
19140  __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
19141  __ret_111; \
19142 })
19143 #endif
19144 
19145 #ifdef __LITTLE_ENDIAN__
19146 #define vqdmlal_lane_s16(__p0_112, __p1_112, __p2_112, __p3_112) __extension__ ({ \
19147  int32x4_t __s0_112 = __p0_112; \
19148  int16x4_t __s1_112 = __p1_112; \
19149  int16x4_t __s2_112 = __p2_112; \
19150  int32x4_t __ret_112; \
19151  __ret_112 = vqdmlal_s16(__s0_112, __s1_112, splat_lane_s16(__s2_112, __p3_112)); \
19152  __ret_112; \
19153 })
19154 #else
19155 #define vqdmlal_lane_s16(__p0_113, __p1_113, __p2_113, __p3_113) __extension__ ({ \
19156  int32x4_t __s0_113 = __p0_113; \
19157  int16x4_t __s1_113 = __p1_113; \
19158  int16x4_t __s2_113 = __p2_113; \
19159  int32x4_t __rev0_113; __rev0_113 = __builtin_shufflevector(__s0_113, __s0_113, 3, 2, 1, 0); \
19160  int16x4_t __rev1_113; __rev1_113 = __builtin_shufflevector(__s1_113, __s1_113, 3, 2, 1, 0); \
19161  int16x4_t __rev2_113; __rev2_113 = __builtin_shufflevector(__s2_113, __s2_113, 3, 2, 1, 0); \
19162  int32x4_t __ret_113; \
19163  __ret_113 = __noswap_vqdmlal_s16(__rev0_113, __rev1_113, __noswap_splat_lane_s16(__rev2_113, __p3_113)); \
19164  __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 3, 2, 1, 0); \
19165  __ret_113; \
19166 })
19167 #endif
19168 
19169 #ifdef __LITTLE_ENDIAN__
19170 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19171  int64x2_t __ret;
19172  __ret = vqdmlal_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19173  return __ret;
19174 }
19175 #else
19176 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19177  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19178  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19179  int64x2_t __ret;
19180  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, (int32x2_t) {__p2, __p2});
19181  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19182  return __ret;
19183 }
19184 __ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19185  int64x2_t __ret;
19186  __ret = __noswap_vqdmlal_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19187  return __ret;
19188 }
19189 #endif
19190 
19191 #ifdef __LITTLE_ENDIAN__
19192 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19193  int32x4_t __ret;
19194  __ret = vqdmlal_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19195  return __ret;
19196 }
19197 #else
19198 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19199  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19200  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19201  int32x4_t __ret;
19202  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
19203  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19204  return __ret;
19205 }
19206 __ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19207  int32x4_t __ret;
19208  __ret = __noswap_vqdmlal_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19209  return __ret;
19210 }
19211 #endif
19212 
19213 #ifdef __LITTLE_ENDIAN__
19214 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19215  int64x2_t __ret;
19216  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19217  return __ret;
19218 }
19219 #else
19220 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19221  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19222  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19223  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
19224  int64x2_t __ret;
19225  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
19226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19227  return __ret;
19228 }
19229 __ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
19230  int64x2_t __ret;
19231  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
19232  return __ret;
19233 }
19234 #endif
19235 
19236 #ifdef __LITTLE_ENDIAN__
19237 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19238  int32x4_t __ret;
19239  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19240  return __ret;
19241 }
19242 #else
19243 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19244  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19245  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19246  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
19247  int32x4_t __ret;
19248  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
19249  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19250  return __ret;
19251 }
19252 __ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
19253  int32x4_t __ret;
19254  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
19255  return __ret;
19256 }
19257 #endif
19258 
19259 #ifdef __LITTLE_ENDIAN__
19260 #define vqdmlsl_lane_s32(__p0_114, __p1_114, __p2_114, __p3_114) __extension__ ({ \
19261  int64x2_t __s0_114 = __p0_114; \
19262  int32x2_t __s1_114 = __p1_114; \
19263  int32x2_t __s2_114 = __p2_114; \
19264  int64x2_t __ret_114; \
19265  __ret_114 = vqdmlsl_s32(__s0_114, __s1_114, splat_lane_s32(__s2_114, __p3_114)); \
19266  __ret_114; \
19267 })
19268 #else
19269 #define vqdmlsl_lane_s32(__p0_115, __p1_115, __p2_115, __p3_115) __extension__ ({ \
19270  int64x2_t __s0_115 = __p0_115; \
19271  int32x2_t __s1_115 = __p1_115; \
19272  int32x2_t __s2_115 = __p2_115; \
19273  int64x2_t __rev0_115; __rev0_115 = __builtin_shufflevector(__s0_115, __s0_115, 1, 0); \
19274  int32x2_t __rev1_115; __rev1_115 = __builtin_shufflevector(__s1_115, __s1_115, 1, 0); \
19275  int32x2_t __rev2_115; __rev2_115 = __builtin_shufflevector(__s2_115, __s2_115, 1, 0); \
19276  int64x2_t __ret_115; \
19277  __ret_115 = __noswap_vqdmlsl_s32(__rev0_115, __rev1_115, __noswap_splat_lane_s32(__rev2_115, __p3_115)); \
19278  __ret_115 = __builtin_shufflevector(__ret_115, __ret_115, 1, 0); \
19279  __ret_115; \
19280 })
19281 #endif
19282 
19283 #ifdef __LITTLE_ENDIAN__
19284 #define vqdmlsl_lane_s16(__p0_116, __p1_116, __p2_116, __p3_116) __extension__ ({ \
19285  int32x4_t __s0_116 = __p0_116; \
19286  int16x4_t __s1_116 = __p1_116; \
19287  int16x4_t __s2_116 = __p2_116; \
19288  int32x4_t __ret_116; \
19289  __ret_116 = vqdmlsl_s16(__s0_116, __s1_116, splat_lane_s16(__s2_116, __p3_116)); \
19290  __ret_116; \
19291 })
19292 #else
19293 #define vqdmlsl_lane_s16(__p0_117, __p1_117, __p2_117, __p3_117) __extension__ ({ \
19294  int32x4_t __s0_117 = __p0_117; \
19295  int16x4_t __s1_117 = __p1_117; \
19296  int16x4_t __s2_117 = __p2_117; \
19297  int32x4_t __rev0_117; __rev0_117 = __builtin_shufflevector(__s0_117, __s0_117, 3, 2, 1, 0); \
19298  int16x4_t __rev1_117; __rev1_117 = __builtin_shufflevector(__s1_117, __s1_117, 3, 2, 1, 0); \
19299  int16x4_t __rev2_117; __rev2_117 = __builtin_shufflevector(__s2_117, __s2_117, 3, 2, 1, 0); \
19300  int32x4_t __ret_117; \
19301  __ret_117 = __noswap_vqdmlsl_s16(__rev0_117, __rev1_117, __noswap_splat_lane_s16(__rev2_117, __p3_117)); \
19302  __ret_117 = __builtin_shufflevector(__ret_117, __ret_117, 3, 2, 1, 0); \
19303  __ret_117; \
19304 })
19305 #endif
19306 
19307 #ifdef __LITTLE_ENDIAN__
19308 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19309  int64x2_t __ret;
19310  __ret = vqdmlsl_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19311  return __ret;
19312 }
19313 #else
19314 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19315  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19316  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19317  int64x2_t __ret;
19318  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, (int32x2_t) {__p2, __p2});
19319  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19320  return __ret;
19321 }
19322 __ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
19323  int64x2_t __ret;
19324  __ret = __noswap_vqdmlsl_s32(__p0, __p1, (int32x2_t) {__p2, __p2});
19325  return __ret;
19326 }
19327 #endif
19328 
19329 #ifdef __LITTLE_ENDIAN__
19330 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19331  int32x4_t __ret;
19332  __ret = vqdmlsl_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19333  return __ret;
19334 }
19335 #else
19336 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19337  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19338  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19339  int32x4_t __ret;
19340  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
19341  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19342  return __ret;
19343 }
19344 __ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
19345  int32x4_t __ret;
19346  __ret = __noswap_vqdmlsl_s16(__p0, __p1, (int16x4_t) {__p2, __p2, __p2, __p2});
19347  return __ret;
19348 }
19349 #endif
19350 
19351 #ifdef __LITTLE_ENDIAN__
19352 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19353  int32x4_t __ret;
19354  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19355  return __ret;
19356 }
19357 #else
19358 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19359  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19360  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19361  int32x4_t __ret;
19362  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19363  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19364  return __ret;
19365 }
19366 __ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19367  int32x4_t __ret;
19368  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19369  return __ret;
19370 }
19371 #endif
19372 
19373 #ifdef __LITTLE_ENDIAN__
19374 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19375  int16x8_t __ret;
19376  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19377  return __ret;
19378 }
19379 #else
19380 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19381  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19382  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19383  int16x8_t __ret;
19384  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19385  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19386  return __ret;
19387 }
19388 __ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19389  int16x8_t __ret;
19390  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19391  return __ret;
19392 }
19393 #endif
19394 
19395 #ifdef __LITTLE_ENDIAN__
19396 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19397  int32x2_t __ret;
19398  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19399  return __ret;
19400 }
19401 #else
19402 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19403  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19404  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19405  int32x2_t __ret;
19406  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19408  return __ret;
19409 }
19410 __ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19411  int32x2_t __ret;
19412  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19413  return __ret;
19414 }
19415 #endif
19416 
19417 #ifdef __LITTLE_ENDIAN__
19418 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19419  int16x4_t __ret;
19420  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19421  return __ret;
19422 }
19423 #else
19424 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19425  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19426  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19427  int16x4_t __ret;
19428  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19429  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19430  return __ret;
19431 }
19432 __ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19433  int16x4_t __ret;
19434  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19435  return __ret;
19436 }
19437 #endif
19438 
19439 #ifdef __LITTLE_ENDIAN__
19440 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19441  int32x4_t __ret;
19442  __ret = vqdmulhq_s32(__p0, (int32x4_t) {__p1, __p1, __p1, __p1});
19443  return __ret;
19444 }
19445 #else
19446 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
19447  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19448  int32x4_t __ret;
19449  __ret = __noswap_vqdmulhq_s32(__rev0, (int32x4_t) {__p1, __p1, __p1, __p1});
19450  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19451  return __ret;
19452 }
19453 #endif
19454 
19455 #ifdef __LITTLE_ENDIAN__
19456 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19457  int16x8_t __ret;
19458  __ret = vqdmulhq_s16(__p0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19459  return __ret;
19460 }
19461 #else
19462 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
19463  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19464  int16x8_t __ret;
19465  __ret = __noswap_vqdmulhq_s16(__rev0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
19466  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19467  return __ret;
19468 }
19469 #endif
19470 
19471 #ifdef __LITTLE_ENDIAN__
19472 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19473  int32x2_t __ret;
19474  __ret = vqdmulh_s32(__p0, (int32x2_t) {__p1, __p1});
19475  return __ret;
19476 }
19477 #else
19478 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
19479  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19480  int32x2_t __ret;
19481  __ret = __noswap_vqdmulh_s32(__rev0, (int32x2_t) {__p1, __p1});
19482  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19483  return __ret;
19484 }
19485 #endif
19486 
19487 #ifdef __LITTLE_ENDIAN__
19488 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19489  int16x4_t __ret;
19490  __ret = vqdmulh_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19491  return __ret;
19492 }
19493 #else
19494 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
19495  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19496  int16x4_t __ret;
19497  __ret = __noswap_vqdmulh_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
19498  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19499  return __ret;
19500 }
19501 #endif
19502 
19503 #ifdef __LITTLE_ENDIAN__
19504 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19505  int64x2_t __ret;
19506  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
19507  return __ret;
19508 }
19509 #else
19510 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19511  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19512  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19513  int64x2_t __ret;
19514  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
19515  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19516  return __ret;
19517 }
19518 __ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
19519  int64x2_t __ret;
19520  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
19521  return __ret;
19522 }
19523 #endif
19524 
19525 #ifdef __LITTLE_ENDIAN__
19526 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19527  int32x4_t __ret;
19528  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
19529  return __ret;
19530 }
19531 #else
19532 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19533  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19534  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19535  int32x4_t __ret;
19536  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
19537  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19538  return __ret;
19539 }
19540 __ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
19541  int32x4_t __ret;
19542  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
19543  return __ret;
19544 }
19545 #endif
19546 
19547 #ifdef __LITTLE_ENDIAN__
19548 #define vqdmull_lane_s32(__p0_118, __p1_118, __p2_118) __extension__ ({ \
19549  int32x2_t __s0_118 = __p0_118; \
19550  int32x2_t __s1_118 = __p1_118; \
19551  int64x2_t __ret_118; \
19552  __ret_118 = vqdmull_s32(__s0_118, splat_lane_s32(__s1_118, __p2_118)); \
19553  __ret_118; \
19554 })
19555 #else
19556 #define vqdmull_lane_s32(__p0_119, __p1_119, __p2_119) __extension__ ({ \
19557  int32x2_t __s0_119 = __p0_119; \
19558  int32x2_t __s1_119 = __p1_119; \
19559  int32x2_t __rev0_119; __rev0_119 = __builtin_shufflevector(__s0_119, __s0_119, 1, 0); \
19560  int32x2_t __rev1_119; __rev1_119 = __builtin_shufflevector(__s1_119, __s1_119, 1, 0); \
19561  int64x2_t __ret_119; \
19562  __ret_119 = __noswap_vqdmull_s32(__rev0_119, __noswap_splat_lane_s32(__rev1_119, __p2_119)); \
19563  __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 1, 0); \
19564  __ret_119; \
19565 })
19566 #endif
19567 
19568 #ifdef __LITTLE_ENDIAN__
19569 #define vqdmull_lane_s16(__p0_120, __p1_120, __p2_120) __extension__ ({ \
19570  int16x4_t __s0_120 = __p0_120; \
19571  int16x4_t __s1_120 = __p1_120; \
19572  int32x4_t __ret_120; \
19573  __ret_120 = vqdmull_s16(__s0_120, splat_lane_s16(__s1_120, __p2_120)); \
19574  __ret_120; \
19575 })
19576 #else
19577 #define vqdmull_lane_s16(__p0_121, __p1_121, __p2_121) __extension__ ({ \
19578  int16x4_t __s0_121 = __p0_121; \
19579  int16x4_t __s1_121 = __p1_121; \
19580  int16x4_t __rev0_121; __rev0_121 = __builtin_shufflevector(__s0_121, __s0_121, 3, 2, 1, 0); \
19581  int16x4_t __rev1_121; __rev1_121 = __builtin_shufflevector(__s1_121, __s1_121, 3, 2, 1, 0); \
19582  int32x4_t __ret_121; \
19583  __ret_121 = __noswap_vqdmull_s16(__rev0_121, __noswap_splat_lane_s16(__rev1_121, __p2_121)); \
19584  __ret_121 = __builtin_shufflevector(__ret_121, __ret_121, 3, 2, 1, 0); \
19585  __ret_121; \
19586 })
19587 #endif
19588 
19589 #ifdef __LITTLE_ENDIAN__
19590 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19591  int64x2_t __ret;
19592  __ret = vqdmull_s32(__p0, (int32x2_t) {__p1, __p1});
19593  return __ret;
19594 }
19595 #else
19596 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19597  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19598  int64x2_t __ret;
19599  __ret = __noswap_vqdmull_s32(__rev0, (int32x2_t) {__p1, __p1});
19600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19601  return __ret;
19602 }
19603 __ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
19604  int64x2_t __ret;
19605  __ret = __noswap_vqdmull_s32(__p0, (int32x2_t) {__p1, __p1});
19606  return __ret;
19607 }
19608 #endif
19609 
19610 #ifdef __LITTLE_ENDIAN__
19611 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19612  int32x4_t __ret;
19613  __ret = vqdmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19614  return __ret;
19615 }
19616 #else
19617 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19618  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19619  int32x4_t __ret;
19620  __ret = __noswap_vqdmull_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
19621  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19622  return __ret;
19623 }
19624 __ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
19625  int32x4_t __ret;
19626  __ret = __noswap_vqdmull_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
19627  return __ret;
19628 }
19629 #endif
19630 
19631 #ifdef __LITTLE_ENDIAN__
19632 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
19633  uint16x4_t __ret;
19634  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
19635  return __ret;
19636 }
19637 #else
19638 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
19639  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19640  uint16x4_t __ret;
19641  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
19642  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19643  return __ret;
19644 }
19645 __ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
19646  uint16x4_t __ret;
19647  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
19648  return __ret;
19649 }
19650 #endif
19651 
19652 #ifdef __LITTLE_ENDIAN__
19653 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
19654  uint32x2_t __ret;
19655  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
19656  return __ret;
19657 }
19658 #else
19659 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
19660  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19661  uint32x2_t __ret;
19662  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
19663  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19664  return __ret;
19665 }
19666 __ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
19667  uint32x2_t __ret;
19668  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
19669  return __ret;
19670 }
19671 #endif
19672 
19673 #ifdef __LITTLE_ENDIAN__
19674 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
19675  uint8x8_t __ret;
19676  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
19677  return __ret;
19678 }
19679 #else
19680 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
19681  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19682  uint8x8_t __ret;
19683  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
19684  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19685  return __ret;
19686 }
19687 __ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
19688  uint8x8_t __ret;
19689  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
19690  return __ret;
19691 }
19692 #endif
19693 
19694 #ifdef __LITTLE_ENDIAN__
19695 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
19696  int16x4_t __ret;
19697  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
19698  return __ret;
19699 }
19700 #else
19701 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
19702  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19703  int16x4_t __ret;
19704  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
19705  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19706  return __ret;
19707 }
19708 __ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
19709  int16x4_t __ret;
19710  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
19711  return __ret;
19712 }
19713 #endif
19714 
19715 #ifdef __LITTLE_ENDIAN__
19716 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
19717  int32x2_t __ret;
19718  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
19719  return __ret;
19720 }
19721 #else
19722 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
19723  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19724  int32x2_t __ret;
19725  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
19726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19727  return __ret;
19728 }
19729 __ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
19730  int32x2_t __ret;
19731  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
19732  return __ret;
19733 }
19734 #endif
19735 
19736 #ifdef __LITTLE_ENDIAN__
19737 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
19738  int8x8_t __ret;
19739  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
19740  return __ret;
19741 }
19742 #else
19743 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
19744  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19745  int8x8_t __ret;
19746  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
19747  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19748  return __ret;
19749 }
19750 __ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
19751  int8x8_t __ret;
19752  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
19753  return __ret;
19754 }
19755 #endif
19756 
19757 #ifdef __LITTLE_ENDIAN__
19758 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
19759  uint16x4_t __ret;
19760  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
19761  return __ret;
19762 }
19763 #else
19764 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
19765  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19766  uint16x4_t __ret;
19767  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
19768  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19769  return __ret;
19770 }
19771 __ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
19772  uint16x4_t __ret;
19773  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
19774  return __ret;
19775 }
19776 #endif
19777 
19778 #ifdef __LITTLE_ENDIAN__
19779 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
19780  uint32x2_t __ret;
19781  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
19782  return __ret;
19783 }
19784 #else
19785 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
19786  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19787  uint32x2_t __ret;
19788  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
19789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19790  return __ret;
19791 }
19792 __ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
19793  uint32x2_t __ret;
19794  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
19795  return __ret;
19796 }
19797 #endif
19798 
19799 #ifdef __LITTLE_ENDIAN__
19800 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
19801  uint8x8_t __ret;
19802  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
19803  return __ret;
19804 }
19805 #else
19806 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
19807  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19808  uint8x8_t __ret;
19809  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
19810  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19811  return __ret;
19812 }
19813 __ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
19814  uint8x8_t __ret;
19815  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
19816  return __ret;
19817 }
19818 #endif
19819 
19820 #ifdef __LITTLE_ENDIAN__
19821 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
19822  int8x16_t __ret;
19823  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
19824  return __ret;
19825 }
19826 #else
19827 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
19828  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19829  int8x16_t __ret;
19830  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
19831  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19832  return __ret;
19833 }
19834 #endif
19835 
19836 #ifdef __LITTLE_ENDIAN__
19837 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
19838  int32x4_t __ret;
19839  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
19840  return __ret;
19841 }
19842 #else
19843 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
19844  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19845  int32x4_t __ret;
19846  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
19847  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19848  return __ret;
19849 }
19850 #endif
19851 
19852 #ifdef __LITTLE_ENDIAN__
19853 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
19854  int16x8_t __ret;
19855  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
19856  return __ret;
19857 }
19858 #else
19859 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
19860  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19861  int16x8_t __ret;
19862  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
19863  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19864  return __ret;
19865 }
19866 #endif
19867 
19868 #ifdef __LITTLE_ENDIAN__
19869 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
19870  int8x8_t __ret;
19871  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
19872  return __ret;
19873 }
19874 #else
19875 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
19876  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19877  int8x8_t __ret;
19878  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
19879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19880  return __ret;
19881 }
19882 #endif
19883 
19884 #ifdef __LITTLE_ENDIAN__
19885 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
19886  int32x2_t __ret;
19887  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
19888  return __ret;
19889 }
19890 #else
19891 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
19892  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19893  int32x2_t __ret;
19894  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
19895  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19896  return __ret;
19897 }
19898 #endif
19899 
19900 #ifdef __LITTLE_ENDIAN__
19901 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
19902  int16x4_t __ret;
19903  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
19904  return __ret;
19905 }
19906 #else
19907 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
19908  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19909  int16x4_t __ret;
19910  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
19911  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19912  return __ret;
19913 }
19914 #endif
19915 
19916 #ifdef __LITTLE_ENDIAN__
19917 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19918  int32x4_t __ret;
19919  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19920  return __ret;
19921 }
19922 #else
19923 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19924  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19925  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19926  int32x4_t __ret;
19927  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19928  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19929  return __ret;
19930 }
19931 __ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
19932  int32x4_t __ret;
19933  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19934  return __ret;
19935 }
19936 #endif
19937 
19938 #ifdef __LITTLE_ENDIAN__
19939 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19940  int16x8_t __ret;
19941  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19942  return __ret;
19943 }
19944 #else
19945 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19946  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19947  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19948  int16x8_t __ret;
19949  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19950  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19951  return __ret;
19952 }
19953 __ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
19954  int16x8_t __ret;
19955  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19956  return __ret;
19957 }
19958 #endif
19959 
19960 #ifdef __LITTLE_ENDIAN__
19961 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19962  int32x2_t __ret;
19963  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19964  return __ret;
19965 }
19966 #else
19967 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19968  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19969  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19970  int32x2_t __ret;
19971  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19972  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19973  return __ret;
19974 }
19975 __ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
19976  int32x2_t __ret;
19977  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19978  return __ret;
19979 }
19980 #endif
19981 
19982 #ifdef __LITTLE_ENDIAN__
19983 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19984  int16x4_t __ret;
19985  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19986  return __ret;
19987 }
19988 #else
19989 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19990  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19991  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19992  int16x4_t __ret;
19993  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19995  return __ret;
19996 }
19997 __ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
19998  int16x4_t __ret;
19999  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20000  return __ret;
20001 }
20002 #endif
20003 
20004 #ifdef __LITTLE_ENDIAN__
20005 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
20006  int32x4_t __ret;
20007  __ret = vqrdmulhq_s32(__p0, (int32x4_t) {__p1, __p1, __p1, __p1});
20008  return __ret;
20009 }
20010 #else
20011 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
20012  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20013  int32x4_t __ret;
20014  __ret = __noswap_vqrdmulhq_s32(__rev0, (int32x4_t) {__p1, __p1, __p1, __p1});
20015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20016  return __ret;
20017 }
20018 #endif
20019 
20020 #ifdef __LITTLE_ENDIAN__
20021 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
20022  int16x8_t __ret;
20023  __ret = vqrdmulhq_s16(__p0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
20024  return __ret;
20025 }
20026 #else
20027 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
20028  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20029  int16x8_t __ret;
20030  __ret = __noswap_vqrdmulhq_s16(__rev0, (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1});
20031  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20032  return __ret;
20033 }
20034 #endif
20035 
20036 #ifdef __LITTLE_ENDIAN__
20037 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
20038  int32x2_t __ret;
20039  __ret = vqrdmulh_s32(__p0, (int32x2_t) {__p1, __p1});
20040  return __ret;
20041 }
20042 #else
20043 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
20044  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20045  int32x2_t __ret;
20046  __ret = __noswap_vqrdmulh_s32(__rev0, (int32x2_t) {__p1, __p1});
20047  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20048  return __ret;
20049 }
20050 #endif
20051 
20052 #ifdef __LITTLE_ENDIAN__
20053 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
20054  int16x4_t __ret;
20055  __ret = vqrdmulh_s16(__p0, (int16x4_t) {__p1, __p1, __p1, __p1});
20056  return __ret;
20057 }
20058 #else
20059 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
20060  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20061  int16x4_t __ret;
20062  __ret = __noswap_vqrdmulh_s16(__rev0, (int16x4_t) {__p1, __p1, __p1, __p1});
20063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20064  return __ret;
20065 }
20066 #endif
20067 
20068 #ifdef __LITTLE_ENDIAN__
20069 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20070  uint8x16_t __ret;
20071  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20072  return __ret;
20073 }
20074 #else
20075 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20076  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20077  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20078  uint8x16_t __ret;
20079  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20080  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20081  return __ret;
20082 }
20083 #endif
20084 
20085 #ifdef __LITTLE_ENDIAN__
20086 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20087  uint32x4_t __ret;
20088  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20089  return __ret;
20090 }
20091 #else
20092 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20093  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20094  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20095  uint32x4_t __ret;
20096  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20097  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20098  return __ret;
20099 }
20100 #endif
20101 
20102 #ifdef __LITTLE_ENDIAN__
20103 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20104  uint64x2_t __ret;
20105  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20106  return __ret;
20107 }
20108 #else
20109 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20110  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20111  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20112  uint64x2_t __ret;
20113  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20114  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20115  return __ret;
20116 }
20117 #endif
20118 
20119 #ifdef __LITTLE_ENDIAN__
20120 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20121  uint16x8_t __ret;
20122  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20123  return __ret;
20124 }
20125 #else
20126 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20127  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20128  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20129  uint16x8_t __ret;
20130  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20131  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20132  return __ret;
20133 }
20134 #endif
20135 
20136 #ifdef __LITTLE_ENDIAN__
20137 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20138  int8x16_t __ret;
20139  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20140  return __ret;
20141 }
20142 #else
20143 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20144  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20145  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20146  int8x16_t __ret;
20147  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20148  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20149  return __ret;
20150 }
20151 #endif
20152 
20153 #ifdef __LITTLE_ENDIAN__
20154 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20155  int32x4_t __ret;
20156  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20157  return __ret;
20158 }
20159 #else
20160 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20161  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20162  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20163  int32x4_t __ret;
20164  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20165  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20166  return __ret;
20167 }
20168 #endif
20169 
20170 #ifdef __LITTLE_ENDIAN__
20171 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20172  int64x2_t __ret;
20173  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20174  return __ret;
20175 }
20176 #else
20177 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20178  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20179  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20180  int64x2_t __ret;
20181  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
20182  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20183  return __ret;
20184 }
20185 #endif
20186 
20187 #ifdef __LITTLE_ENDIAN__
20188 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20189  int16x8_t __ret;
20190  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20191  return __ret;
20192 }
20193 #else
20194 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20195  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20196  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20197  int16x8_t __ret;
20198  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20199  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20200  return __ret;
20201 }
20202 #endif
20203 
20204 #ifdef __LITTLE_ENDIAN__
20205 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20206  uint8x8_t __ret;
20207  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20208  return __ret;
20209 }
20210 #else
20211 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20212  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20213  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20214  uint8x8_t __ret;
20215  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20217  return __ret;
20218 }
20219 #endif
20220 
20221 #ifdef __LITTLE_ENDIAN__
20222 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20223  uint32x2_t __ret;
20224  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20225  return __ret;
20226 }
20227 #else
20228 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20229  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20230  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20231  uint32x2_t __ret;
20232  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20233  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20234  return __ret;
20235 }
20236 #endif
20237 
20238 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
20239  uint64x1_t __ret;
20240  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
20241  return __ret;
20242 }
20243 #ifdef __LITTLE_ENDIAN__
20244 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20245  uint16x4_t __ret;
20246  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20247  return __ret;
20248 }
20249 #else
20250 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20251  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20252  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20253  uint16x4_t __ret;
20254  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20255  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20256  return __ret;
20257 }
20258 #endif
20259 
20260 #ifdef __LITTLE_ENDIAN__
20261 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
20262  int8x8_t __ret;
20263  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20264  return __ret;
20265 }
20266 #else
20267 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
20268  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20269  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20270  int8x8_t __ret;
20271  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20272  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20273  return __ret;
20274 }
20275 #endif
20276 
20277 #ifdef __LITTLE_ENDIAN__
20278 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
20279  int32x2_t __ret;
20280  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20281  return __ret;
20282 }
20283 #else
20284 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
20285  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20286  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20287  int32x2_t __ret;
20288  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20289  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20290  return __ret;
20291 }
20292 #endif
20293 
20294 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
20295  int64x1_t __ret;
20296  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
20297  return __ret;
20298 }
20299 #ifdef __LITTLE_ENDIAN__
20300 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
20301  int16x4_t __ret;
20302  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20303  return __ret;
20304 }
20305 #else
20306 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
20307  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20308  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20309  int16x4_t __ret;
20310  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20311  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20312  return __ret;
20313 }
20314 #endif
20315 
20316 #ifdef __LITTLE_ENDIAN__
20317 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20318  uint32x4_t __s0 = __p0; \
20319  uint16x4_t __ret; \
20320  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
20321  __ret; \
20322 })
20323 #else
20324 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20325  uint32x4_t __s0 = __p0; \
20326  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20327  uint16x4_t __ret; \
20328  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
20329  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20330  __ret; \
20331 })
20332 #define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
20333  uint32x4_t __s0 = __p0; \
20334  uint16x4_t __ret; \
20335  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
20336  __ret; \
20337 })
20338 #endif
20339 
20340 #ifdef __LITTLE_ENDIAN__
20341 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20342  uint64x2_t __s0 = __p0; \
20343  uint32x2_t __ret; \
20344  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
20345  __ret; \
20346 })
20347 #else
20348 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20349  uint64x2_t __s0 = __p0; \
20350  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20351  uint32x2_t __ret; \
20352  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
20353  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20354  __ret; \
20355 })
20356 #define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
20357  uint64x2_t __s0 = __p0; \
20358  uint32x2_t __ret; \
20359  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
20360  __ret; \
20361 })
20362 #endif
20363 
20364 #ifdef __LITTLE_ENDIAN__
20365 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20366  uint16x8_t __s0 = __p0; \
20367  uint8x8_t __ret; \
20368  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
20369  __ret; \
20370 })
20371 #else
20372 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20373  uint16x8_t __s0 = __p0; \
20374  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20375  uint8x8_t __ret; \
20376  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
20377  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20378  __ret; \
20379 })
20380 #define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
20381  uint16x8_t __s0 = __p0; \
20382  uint8x8_t __ret; \
20383  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
20384  __ret; \
20385 })
20386 #endif
20387 
20388 #ifdef __LITTLE_ENDIAN__
20389 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20390  int32x4_t __s0 = __p0; \
20391  int16x4_t __ret; \
20392  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
20393  __ret; \
20394 })
20395 #else
20396 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20397  int32x4_t __s0 = __p0; \
20398  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20399  int16x4_t __ret; \
20400  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
20401  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20402  __ret; \
20403 })
20404 #define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
20405  int32x4_t __s0 = __p0; \
20406  int16x4_t __ret; \
20407  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
20408  __ret; \
20409 })
20410 #endif
20411 
20412 #ifdef __LITTLE_ENDIAN__
20413 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20414  int64x2_t __s0 = __p0; \
20415  int32x2_t __ret; \
20416  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
20417  __ret; \
20418 })
20419 #else
20420 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20421  int64x2_t __s0 = __p0; \
20422  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20423  int32x2_t __ret; \
20424  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
20425  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20426  __ret; \
20427 })
20428 #define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
20429  int64x2_t __s0 = __p0; \
20430  int32x2_t __ret; \
20431  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
20432  __ret; \
20433 })
20434 #endif
20435 
20436 #ifdef __LITTLE_ENDIAN__
20437 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20438  int16x8_t __s0 = __p0; \
20439  int8x8_t __ret; \
20440  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
20441  __ret; \
20442 })
20443 #else
20444 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20445  int16x8_t __s0 = __p0; \
20446  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20447  int8x8_t __ret; \
20448  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
20449  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20450  __ret; \
20451 })
20452 #define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
20453  int16x8_t __s0 = __p0; \
20454  int8x8_t __ret; \
20455  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
20456  __ret; \
20457 })
20458 #endif
20459 
20460 #ifdef __LITTLE_ENDIAN__
20461 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20462  int32x4_t __s0 = __p0; \
20463  uint16x4_t __ret; \
20464  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
20465  __ret; \
20466 })
20467 #else
20468 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20469  int32x4_t __s0 = __p0; \
20470  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20471  uint16x4_t __ret; \
20472  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
20473  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20474  __ret; \
20475 })
20476 #define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
20477  int32x4_t __s0 = __p0; \
20478  uint16x4_t __ret; \
20479  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
20480  __ret; \
20481 })
20482 #endif
20483 
20484 #ifdef __LITTLE_ENDIAN__
20485 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20486  int64x2_t __s0 = __p0; \
20487  uint32x2_t __ret; \
20488  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
20489  __ret; \
20490 })
20491 #else
20492 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20493  int64x2_t __s0 = __p0; \
20494  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20495  uint32x2_t __ret; \
20496  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
20497  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20498  __ret; \
20499 })
20500 #define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
20501  int64x2_t __s0 = __p0; \
20502  uint32x2_t __ret; \
20503  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
20504  __ret; \
20505 })
20506 #endif
20507 
20508 #ifdef __LITTLE_ENDIAN__
20509 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20510  int16x8_t __s0 = __p0; \
20511  uint8x8_t __ret; \
20512  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
20513  __ret; \
20514 })
20515 #else
20516 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20517  int16x8_t __s0 = __p0; \
20518  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20519  uint8x8_t __ret; \
20520  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
20521  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20522  __ret; \
20523 })
20524 #define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
20525  int16x8_t __s0 = __p0; \
20526  uint8x8_t __ret; \
20527  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
20528  __ret; \
20529 })
20530 #endif
20531 
20532 #ifdef __LITTLE_ENDIAN__
20533 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20534  uint8x16_t __ret;
20535  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20536  return __ret;
20537 }
20538 #else
20539 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20540  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20541  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20542  uint8x16_t __ret;
20543  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20544  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20545  return __ret;
20546 }
20547 #endif
20548 
20549 #ifdef __LITTLE_ENDIAN__
20550 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20551  uint32x4_t __ret;
20552  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20553  return __ret;
20554 }
20555 #else
20556 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20557  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20558  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20559  uint32x4_t __ret;
20560  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20561  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20562  return __ret;
20563 }
20564 #endif
20565 
20566 #ifdef __LITTLE_ENDIAN__
20567 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20568  uint64x2_t __ret;
20569  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20570  return __ret;
20571 }
20572 #else
20573 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20574  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20575  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20576  uint64x2_t __ret;
20577  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20578  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20579  return __ret;
20580 }
20581 #endif
20582 
20583 #ifdef __LITTLE_ENDIAN__
20584 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20585  uint16x8_t __ret;
20586  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20587  return __ret;
20588 }
20589 #else
20590 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20591  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20592  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20593  uint16x8_t __ret;
20594  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20595  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20596  return __ret;
20597 }
20598 #endif
20599 
20600 #ifdef __LITTLE_ENDIAN__
20601 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20602  int8x16_t __ret;
20603  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20604  return __ret;
20605 }
20606 #else
20607 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20608  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20609  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20610  int8x16_t __ret;
20611  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20612  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20613  return __ret;
20614 }
20615 #endif
20616 
20617 #ifdef __LITTLE_ENDIAN__
20618 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20619  int32x4_t __ret;
20620  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20621  return __ret;
20622 }
20623 #else
20624 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20625  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20626  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20627  int32x4_t __ret;
20628  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20630  return __ret;
20631 }
20632 #endif
20633 
20634 #ifdef __LITTLE_ENDIAN__
20635 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20636  int64x2_t __ret;
20637  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20638  return __ret;
20639 }
20640 #else
20641 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20642  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20643  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20644  int64x2_t __ret;
20645  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
20646  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20647  return __ret;
20648 }
20649 #endif
20650 
20651 #ifdef __LITTLE_ENDIAN__
20652 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20653  int16x8_t __ret;
20654  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20655  return __ret;
20656 }
20657 #else
20658 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
20659  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20660  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20661  int16x8_t __ret;
20662  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20663  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20664  return __ret;
20665 }
20666 #endif
20667 
20668 #ifdef __LITTLE_ENDIAN__
20669 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20670  uint8x8_t __ret;
20671  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20672  return __ret;
20673 }
20674 #else
20675 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
20676  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20677  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20678  uint8x8_t __ret;
20679  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20680  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20681  return __ret;
20682 }
20683 #endif
20684 
20685 #ifdef __LITTLE_ENDIAN__
20686 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20687  uint32x2_t __ret;
20688  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20689  return __ret;
20690 }
20691 #else
20692 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
20693  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20694  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20695  uint32x2_t __ret;
20696  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20697  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20698  return __ret;
20699 }
20700 #endif
20701 
20702 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
20703  uint64x1_t __ret;
20704  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
20705  return __ret;
20706 }
20707 #ifdef __LITTLE_ENDIAN__
20708 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20709  uint16x4_t __ret;
20710  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20711  return __ret;
20712 }
20713 #else
20714 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
20715  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20716  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20717  uint16x4_t __ret;
20718  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20719  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20720  return __ret;
20721 }
20722 #endif
20723 
20724 #ifdef __LITTLE_ENDIAN__
20725 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
20726  int8x8_t __ret;
20727  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20728  return __ret;
20729 }
20730 #else
20731 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
20732  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20733  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20734  int8x8_t __ret;
20735  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20736  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20737  return __ret;
20738 }
20739 #endif
20740 
20741 #ifdef __LITTLE_ENDIAN__
20742 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
20743  int32x2_t __ret;
20744  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20745  return __ret;
20746 }
20747 #else
20748 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
20749  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20750  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20751  int32x2_t __ret;
20752  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20753  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20754  return __ret;
20755 }
20756 #endif
20757 
20758 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
20759  int64x1_t __ret;
20760  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
20761  return __ret;
20762 }
20763 #ifdef __LITTLE_ENDIAN__
20764 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
20765  int16x4_t __ret;
20766  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20767  return __ret;
20768 }
20769 #else
20770 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
20771  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20772  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20773  int16x4_t __ret;
20774  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20775  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20776  return __ret;
20777 }
20778 #endif
20779 
20780 #ifdef __LITTLE_ENDIAN__
20781 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
20782  uint8x16_t __s0 = __p0; \
20783  uint8x16_t __ret; \
20784  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
20785  __ret; \
20786 })
20787 #else
20788 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
20789  uint8x16_t __s0 = __p0; \
20790  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20791  uint8x16_t __ret; \
20792  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
20793  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20794  __ret; \
20795 })
20796 #endif
20797 
20798 #ifdef __LITTLE_ENDIAN__
20799 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
20800  uint32x4_t __s0 = __p0; \
20801  uint32x4_t __ret; \
20802  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
20803  __ret; \
20804 })
20805 #else
20806 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
20807  uint32x4_t __s0 = __p0; \
20808  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20809  uint32x4_t __ret; \
20810  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
20811  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20812  __ret; \
20813 })
20814 #endif
20815 
20816 #ifdef __LITTLE_ENDIAN__
20817 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
20818  uint64x2_t __s0 = __p0; \
20819  uint64x2_t __ret; \
20820  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
20821  __ret; \
20822 })
20823 #else
20824 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
20825  uint64x2_t __s0 = __p0; \
20826  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20827  uint64x2_t __ret; \
20828  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
20829  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20830  __ret; \
20831 })
20832 #endif
20833 
20834 #ifdef __LITTLE_ENDIAN__
20835 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
20836  uint16x8_t __s0 = __p0; \
20837  uint16x8_t __ret; \
20838  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
20839  __ret; \
20840 })
20841 #else
20842 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
20843  uint16x8_t __s0 = __p0; \
20844  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20845  uint16x8_t __ret; \
20846  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
20847  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20848  __ret; \
20849 })
20850 #endif
20851 
20852 #ifdef __LITTLE_ENDIAN__
20853 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
20854  int8x16_t __s0 = __p0; \
20855  int8x16_t __ret; \
20856  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
20857  __ret; \
20858 })
20859 #else
20860 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
20861  int8x16_t __s0 = __p0; \
20862  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20863  int8x16_t __ret; \
20864  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
20865  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
20866  __ret; \
20867 })
20868 #endif
20869 
20870 #ifdef __LITTLE_ENDIAN__
20871 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
20872  int32x4_t __s0 = __p0; \
20873  int32x4_t __ret; \
20874  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
20875  __ret; \
20876 })
20877 #else
20878 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
20879  int32x4_t __s0 = __p0; \
20880  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20881  int32x4_t __ret; \
20882  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
20883  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20884  __ret; \
20885 })
20886 #endif
20887 
20888 #ifdef __LITTLE_ENDIAN__
20889 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
20890  int64x2_t __s0 = __p0; \
20891  int64x2_t __ret; \
20892  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
20893  __ret; \
20894 })
20895 #else
20896 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
20897  int64x2_t __s0 = __p0; \
20898  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20899  int64x2_t __ret; \
20900  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
20901  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20902  __ret; \
20903 })
20904 #endif
20905 
20906 #ifdef __LITTLE_ENDIAN__
20907 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
20908  int16x8_t __s0 = __p0; \
20909  int16x8_t __ret; \
20910  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
20911  __ret; \
20912 })
20913 #else
20914 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
20915  int16x8_t __s0 = __p0; \
20916  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20917  int16x8_t __ret; \
20918  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
20919  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20920  __ret; \
20921 })
20922 #endif
20923 
20924 #ifdef __LITTLE_ENDIAN__
20925 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
20926  uint8x8_t __s0 = __p0; \
20927  uint8x8_t __ret; \
20928  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
20929  __ret; \
20930 })
20931 #else
20932 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
20933  uint8x8_t __s0 = __p0; \
20934  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20935  uint8x8_t __ret; \
20936  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
20937  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20938  __ret; \
20939 })
20940 #endif
20941 
20942 #ifdef __LITTLE_ENDIAN__
20943 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
20944  uint32x2_t __s0 = __p0; \
20945  uint32x2_t __ret; \
20946  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
20947  __ret; \
20948 })
20949 #else
20950 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
20951  uint32x2_t __s0 = __p0; \
20952  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
20953  uint32x2_t __ret; \
20954  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
20955  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
20956  __ret; \
20957 })
20958 #endif
20959 
20960 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
20961  uint64x1_t __s0 = __p0; \
20962  uint64x1_t __ret; \
20963  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
20964  __ret; \
20965 })
20966 #ifdef __LITTLE_ENDIAN__
20967 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
20968  uint16x4_t __s0 = __p0; \
20969  uint16x4_t __ret; \
20970  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
20971  __ret; \
20972 })
20973 #else
20974 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
20975  uint16x4_t __s0 = __p0; \
20976  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
20977  uint16x4_t __ret; \
20978  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
20979  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
20980  __ret; \
20981 })
20982 #endif
20983 
20984 #ifdef __LITTLE_ENDIAN__
20985 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
20986  int8x8_t __s0 = __p0; \
20987  int8x8_t __ret; \
20988  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
20989  __ret; \
20990 })
20991 #else
20992 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
20993  int8x8_t __s0 = __p0; \
20994  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
20995  int8x8_t __ret; \
20996  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
20997  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
20998  __ret; \
20999 })
21000 #endif
21001 
21002 #ifdef __LITTLE_ENDIAN__
21003 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
21004  int32x2_t __s0 = __p0; \
21005  int32x2_t __ret; \
21006  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
21007  __ret; \
21008 })
21009 #else
21010 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
21011  int32x2_t __s0 = __p0; \
21012  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21013  int32x2_t __ret; \
21014  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
21015  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21016  __ret; \
21017 })
21018 #endif
21019 
21020 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
21021  int64x1_t __s0 = __p0; \
21022  int64x1_t __ret; \
21023  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
21024  __ret; \
21025 })
21026 #ifdef __LITTLE_ENDIAN__
21027 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
21028  int16x4_t __s0 = __p0; \
21029  int16x4_t __ret; \
21030  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
21031  __ret; \
21032 })
21033 #else
21034 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
21035  int16x4_t __s0 = __p0; \
21036  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21037  int16x4_t __ret; \
21038  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
21039  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21040  __ret; \
21041 })
21042 #endif
21043 
21044 #ifdef __LITTLE_ENDIAN__
21045 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
21046  int8x16_t __s0 = __p0; \
21047  uint8x16_t __ret; \
21048  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
21049  __ret; \
21050 })
21051 #else
21052 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
21053  int8x16_t __s0 = __p0; \
21054  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21055  uint8x16_t __ret; \
21056  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
21057  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21058  __ret; \
21059 })
21060 #endif
21061 
21062 #ifdef __LITTLE_ENDIAN__
21063 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
21064  int32x4_t __s0 = __p0; \
21065  uint32x4_t __ret; \
21066  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
21067  __ret; \
21068 })
21069 #else
21070 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
21071  int32x4_t __s0 = __p0; \
21072  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21073  uint32x4_t __ret; \
21074  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
21075  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21076  __ret; \
21077 })
21078 #endif
21079 
21080 #ifdef __LITTLE_ENDIAN__
21081 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
21082  int64x2_t __s0 = __p0; \
21083  uint64x2_t __ret; \
21084  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
21085  __ret; \
21086 })
21087 #else
21088 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
21089  int64x2_t __s0 = __p0; \
21090  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21091  uint64x2_t __ret; \
21092  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
21093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21094  __ret; \
21095 })
21096 #endif
21097 
21098 #ifdef __LITTLE_ENDIAN__
21099 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
21100  int16x8_t __s0 = __p0; \
21101  uint16x8_t __ret; \
21102  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
21103  __ret; \
21104 })
21105 #else
21106 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
21107  int16x8_t __s0 = __p0; \
21108  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21109  uint16x8_t __ret; \
21110  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
21111  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21112  __ret; \
21113 })
21114 #endif
21115 
21116 #ifdef __LITTLE_ENDIAN__
21117 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
21118  int8x8_t __s0 = __p0; \
21119  uint8x8_t __ret; \
21120  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
21121  __ret; \
21122 })
21123 #else
21124 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
21125  int8x8_t __s0 = __p0; \
21126  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21127  uint8x8_t __ret; \
21128  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
21129  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21130  __ret; \
21131 })
21132 #endif
21133 
21134 #ifdef __LITTLE_ENDIAN__
21135 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
21136  int32x2_t __s0 = __p0; \
21137  uint32x2_t __ret; \
21138  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
21139  __ret; \
21140 })
21141 #else
21142 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
21143  int32x2_t __s0 = __p0; \
21144  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21145  uint32x2_t __ret; \
21146  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
21147  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21148  __ret; \
21149 })
21150 #endif
21151 
21152 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
21153  int64x1_t __s0 = __p0; \
21154  uint64x1_t __ret; \
21155  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
21156  __ret; \
21157 })
21158 #ifdef __LITTLE_ENDIAN__
21159 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
21160  int16x4_t __s0 = __p0; \
21161  uint16x4_t __ret; \
21162  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
21163  __ret; \
21164 })
21165 #else
21166 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
21167  int16x4_t __s0 = __p0; \
21168  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21169  uint16x4_t __ret; \
21170  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
21171  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21172  __ret; \
21173 })
21174 #endif
21175 
21176 #ifdef __LITTLE_ENDIAN__
21177 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21178  uint32x4_t __s0 = __p0; \
21179  uint16x4_t __ret; \
21180  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
21181  __ret; \
21182 })
21183 #else
21184 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21185  uint32x4_t __s0 = __p0; \
21186  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21187  uint16x4_t __ret; \
21188  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21189  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21190  __ret; \
21191 })
21192 #define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
21193  uint32x4_t __s0 = __p0; \
21194  uint16x4_t __ret; \
21195  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
21196  __ret; \
21197 })
21198 #endif
21199 
21200 #ifdef __LITTLE_ENDIAN__
21201 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21202  uint64x2_t __s0 = __p0; \
21203  uint32x2_t __ret; \
21204  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
21205  __ret; \
21206 })
21207 #else
21208 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21209  uint64x2_t __s0 = __p0; \
21210  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21211  uint32x2_t __ret; \
21212  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21213  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21214  __ret; \
21215 })
21216 #define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
21217  uint64x2_t __s0 = __p0; \
21218  uint32x2_t __ret; \
21219  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
21220  __ret; \
21221 })
21222 #endif
21223 
21224 #ifdef __LITTLE_ENDIAN__
21225 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21226  uint16x8_t __s0 = __p0; \
21227  uint8x8_t __ret; \
21228  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
21229  __ret; \
21230 })
21231 #else
21232 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21233  uint16x8_t __s0 = __p0; \
21234  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21235  uint8x8_t __ret; \
21236  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21237  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21238  __ret; \
21239 })
21240 #define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
21241  uint16x8_t __s0 = __p0; \
21242  uint8x8_t __ret; \
21243  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
21244  __ret; \
21245 })
21246 #endif
21247 
21248 #ifdef __LITTLE_ENDIAN__
21249 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21250  int32x4_t __s0 = __p0; \
21251  int16x4_t __ret; \
21252  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
21253  __ret; \
21254 })
21255 #else
21256 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21257  int32x4_t __s0 = __p0; \
21258  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21259  int16x4_t __ret; \
21260  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21261  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21262  __ret; \
21263 })
21264 #define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
21265  int32x4_t __s0 = __p0; \
21266  int16x4_t __ret; \
21267  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
21268  __ret; \
21269 })
21270 #endif
21271 
21272 #ifdef __LITTLE_ENDIAN__
21273 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21274  int64x2_t __s0 = __p0; \
21275  int32x2_t __ret; \
21276  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
21277  __ret; \
21278 })
21279 #else
21280 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21281  int64x2_t __s0 = __p0; \
21282  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21283  int32x2_t __ret; \
21284  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21285  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21286  __ret; \
21287 })
21288 #define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
21289  int64x2_t __s0 = __p0; \
21290  int32x2_t __ret; \
21291  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
21292  __ret; \
21293 })
21294 #endif
21295 
21296 #ifdef __LITTLE_ENDIAN__
21297 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21298  int16x8_t __s0 = __p0; \
21299  int8x8_t __ret; \
21300  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
21301  __ret; \
21302 })
21303 #else
21304 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21305  int16x8_t __s0 = __p0; \
21306  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21307  int8x8_t __ret; \
21308  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21309  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21310  __ret; \
21311 })
21312 #define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
21313  int16x8_t __s0 = __p0; \
21314  int8x8_t __ret; \
21315  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
21316  __ret; \
21317 })
21318 #endif
21319 
21320 #ifdef __LITTLE_ENDIAN__
21321 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21322  int32x4_t __s0 = __p0; \
21323  uint16x4_t __ret; \
21324  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
21325  __ret; \
21326 })
21327 #else
21328 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21329  int32x4_t __s0 = __p0; \
21330  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21331  uint16x4_t __ret; \
21332  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
21333  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21334  __ret; \
21335 })
21336 #define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
21337  int32x4_t __s0 = __p0; \
21338  uint16x4_t __ret; \
21339  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
21340  __ret; \
21341 })
21342 #endif
21343 
21344 #ifdef __LITTLE_ENDIAN__
21345 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21346  int64x2_t __s0 = __p0; \
21347  uint32x2_t __ret; \
21348  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
21349  __ret; \
21350 })
21351 #else
21352 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21353  int64x2_t __s0 = __p0; \
21354  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21355  uint32x2_t __ret; \
21356  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
21357  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21358  __ret; \
21359 })
21360 #define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
21361  int64x2_t __s0 = __p0; \
21362  uint32x2_t __ret; \
21363  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
21364  __ret; \
21365 })
21366 #endif
21367 
21368 #ifdef __LITTLE_ENDIAN__
21369 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21370  int16x8_t __s0 = __p0; \
21371  uint8x8_t __ret; \
21372  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
21373  __ret; \
21374 })
21375 #else
21376 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21377  int16x8_t __s0 = __p0; \
21378  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21379  uint8x8_t __ret; \
21380  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
21381  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21382  __ret; \
21383 })
21384 #define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
21385  int16x8_t __s0 = __p0; \
21386  uint8x8_t __ret; \
21387  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
21388  __ret; \
21389 })
21390 #endif
21391 
21392 #ifdef __LITTLE_ENDIAN__
21393 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
21394  uint8x16_t __ret;
21395  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
21396  return __ret;
21397 }
21398 #else
21399 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
21400  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21401  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21402  uint8x16_t __ret;
21403  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
21404  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21405  return __ret;
21406 }
21407 #endif
21408 
21409 #ifdef __LITTLE_ENDIAN__
21410 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
21411  uint32x4_t __ret;
21412  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
21413  return __ret;
21414 }
21415 #else
21416 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
21417  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21418  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21419  uint32x4_t __ret;
21420  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
21421  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21422  return __ret;
21423 }
21424 #endif
21425 
21426 #ifdef __LITTLE_ENDIAN__
21427 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
21428  uint64x2_t __ret;
21429  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
21430  return __ret;
21431 }
21432 #else
21433 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
21434  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21435  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21436  uint64x2_t __ret;
21437  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
21438  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21439  return __ret;
21440 }
21441 #endif
21442 
21443 #ifdef __LITTLE_ENDIAN__
21444 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
21445  uint16x8_t __ret;
21446  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
21447  return __ret;
21448 }
21449 #else
21450 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
21451  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21452  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21453  uint16x8_t __ret;
21454  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
21455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21456  return __ret;
21457 }
21458 #endif
21459 
21460 #ifdef __LITTLE_ENDIAN__
21461 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
21462  int8x16_t __ret;
21463  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
21464  return __ret;
21465 }
21466 #else
21467 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
21468  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21469  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21470  int8x16_t __ret;
21471  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
21472  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21473  return __ret;
21474 }
21475 #endif
21476 
21477 #ifdef __LITTLE_ENDIAN__
21478 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
21479  int32x4_t __ret;
21480  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
21481  return __ret;
21482 }
21483 #else
21484 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
21485  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21486  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21487  int32x4_t __ret;
21488  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
21489  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21490  return __ret;
21491 }
21492 __ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
21493  int32x4_t __ret;
21494  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
21495  return __ret;
21496 }
21497 #endif
21498 
21499 #ifdef __LITTLE_ENDIAN__
21500 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
21501  int64x2_t __ret;
21502  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
21503  return __ret;
21504 }
21505 #else
21506 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
21507  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21508  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21509  int64x2_t __ret;
21510  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
21511  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21512  return __ret;
21513 }
21514 #endif
21515 
21516 #ifdef __LITTLE_ENDIAN__
21517 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
21518  int16x8_t __ret;
21519  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21520  return __ret;
21521 }
21522 #else
21523 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
21524  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21525  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21526  int16x8_t __ret;
21527  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
21528  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21529  return __ret;
21530 }
21531 __ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
21532  int16x8_t __ret;
21533  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21534  return __ret;
21535 }
21536 #endif
21537 
21538 #ifdef __LITTLE_ENDIAN__
21539 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
21540  uint8x8_t __ret;
21541  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
21542  return __ret;
21543 }
21544 #else
21545 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
21546  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21547  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21548  uint8x8_t __ret;
21549  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
21550  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21551  return __ret;
21552 }
21553 #endif
21554 
21555 #ifdef __LITTLE_ENDIAN__
21556 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
21557  uint32x2_t __ret;
21558  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
21559  return __ret;
21560 }
21561 #else
21562 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
21563  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21564  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21565  uint32x2_t __ret;
21566  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21567  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21568  return __ret;
21569 }
21570 #endif
21571 
21572 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
21573  uint64x1_t __ret;
21574  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21575  return __ret;
21576 }
21577 #ifdef __LITTLE_ENDIAN__
21578 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
21579  uint16x4_t __ret;
21580  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21581  return __ret;
21582 }
21583 #else
21584 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
21585  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21586  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21587  uint16x4_t __ret;
21588  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21589  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21590  return __ret;
21591 }
21592 #endif
21593 
21594 #ifdef __LITTLE_ENDIAN__
21595 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
21596  int8x8_t __ret;
21597  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21598  return __ret;
21599 }
21600 #else
21601 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
21602  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21603  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21604  int8x8_t __ret;
21605  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21606  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21607  return __ret;
21608 }
21609 #endif
21610 
21611 #ifdef __LITTLE_ENDIAN__
21612 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
21613  int32x2_t __ret;
21614  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21615  return __ret;
21616 }
21617 #else
21618 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
21619  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21620  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21621  int32x2_t __ret;
21622  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21623  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21624  return __ret;
21625 }
21626 __ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
21627  int32x2_t __ret;
21628  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21629  return __ret;
21630 }
21631 #endif
21632 
21633 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
21634  int64x1_t __ret;
21635  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21636  return __ret;
21637 }
21638 #ifdef __LITTLE_ENDIAN__
21639 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
21640  int16x4_t __ret;
21641  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21642  return __ret;
21643 }
21644 #else
21645 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
21646  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21647  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21648  int16x4_t __ret;
21649  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21650  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21651  return __ret;
21652 }
21653 __ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
21654  int16x4_t __ret;
21655  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21656  return __ret;
21657 }
21658 #endif
21659 
21660 #ifdef __LITTLE_ENDIAN__
21661 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21662  uint16x4_t __ret;
21663  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21664  return __ret;
21665 }
21666 #else
21667 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21668  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21669  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21670  uint16x4_t __ret;
21671  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
21672  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21673  return __ret;
21674 }
21675 __ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
21676  uint16x4_t __ret;
21677  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
21678  return __ret;
21679 }
21680 #endif
21681 
21682 #ifdef __LITTLE_ENDIAN__
21683 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21684  uint32x2_t __ret;
21685  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
21686  return __ret;
21687 }
21688 #else
21689 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21690  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21691  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21692  uint32x2_t __ret;
21693  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
21694  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21695  return __ret;
21696 }
21697 __ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
21698  uint32x2_t __ret;
21699  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
21700  return __ret;
21701 }
21702 #endif
21703 
21704 #ifdef __LITTLE_ENDIAN__
21705 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21706  uint8x8_t __ret;
21707  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
21708  return __ret;
21709 }
21710 #else
21711 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21712  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21713  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21714  uint8x8_t __ret;
21715  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
21716  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21717  return __ret;
21718 }
21719 __ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
21720  uint8x8_t __ret;
21721  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
21722  return __ret;
21723 }
21724 #endif
21725 
21726 #ifdef __LITTLE_ENDIAN__
21727 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21728  int16x4_t __ret;
21729  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
21730  return __ret;
21731 }
21732 #else
21733 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21734  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21735  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21736  int16x4_t __ret;
21737  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
21738  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21739  return __ret;
21740 }
21741 __ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
21742  int16x4_t __ret;
21743  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
21744  return __ret;
21745 }
21746 #endif
21747 
21748 #ifdef __LITTLE_ENDIAN__
21749 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21750  int32x2_t __ret;
21751  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
21752  return __ret;
21753 }
21754 #else
21755 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21756  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21757  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21758  int32x2_t __ret;
21759  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
21760  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21761  return __ret;
21762 }
21763 __ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
21764  int32x2_t __ret;
21765  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
21766  return __ret;
21767 }
21768 #endif
21769 
21770 #ifdef __LITTLE_ENDIAN__
21771 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21772  int8x8_t __ret;
21773  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
21774  return __ret;
21775 }
21776 #else
21777 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21778  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21779  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21780  int8x8_t __ret;
21781  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
21782  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21783  return __ret;
21784 }
21785 __ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
21786  int8x8_t __ret;
21787  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
21788  return __ret;
21789 }
21790 #endif
21791 
21792 #ifdef __LITTLE_ENDIAN__
21793 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
21794  uint32x4_t __ret;
21795  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
21796  return __ret;
21797 }
21798 #else
21799 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
21800  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21801  uint32x4_t __ret;
21802  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
21803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21804  return __ret;
21805 }
21806 #endif
21807 
21808 #ifdef __LITTLE_ENDIAN__
21809 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
21810  float32x4_t __ret;
21811  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
21812  return __ret;
21813 }
21814 #else
21815 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
21816  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21817  float32x4_t __ret;
21818  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
21819  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21820  return __ret;
21821 }
21822 #endif
21823 
21824 #ifdef __LITTLE_ENDIAN__
21825 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
21826  uint32x2_t __ret;
21827  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
21828  return __ret;
21829 }
21830 #else
21831 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
21832  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21833  uint32x2_t __ret;
21834  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
21835  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21836  return __ret;
21837 }
21838 #endif
21839 
21840 #ifdef __LITTLE_ENDIAN__
21841 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
21842  float32x2_t __ret;
21843  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
21844  return __ret;
21845 }
21846 #else
21847 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
21848  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21849  float32x2_t __ret;
21850  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
21851  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21852  return __ret;
21853 }
21854 #endif
21855 
21856 #ifdef __LITTLE_ENDIAN__
21857 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
21858  float32x4_t __ret;
21859  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21860  return __ret;
21861 }
21862 #else
21863 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
21864  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21865  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21866  float32x4_t __ret;
21867  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21869  return __ret;
21870 }
21871 #endif
21872 
21873 #ifdef __LITTLE_ENDIAN__
21874 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
21875  float32x2_t __ret;
21876  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21877  return __ret;
21878 }
21879 #else
21880 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
21881  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21882  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21883  float32x2_t __ret;
21884  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21886  return __ret;
21887 }
21888 #endif
21889 
21890 #ifdef __LITTLE_ENDIAN__
21891 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
21892  poly8x8_t __ret;
21893  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21894  return __ret;
21895 }
21896 #else
21897 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
21898  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21899  poly8x8_t __ret;
21900  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21902  return __ret;
21903 }
21904 #endif
21905 
21906 #ifdef __LITTLE_ENDIAN__
21907 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
21908  poly8x16_t __ret;
21909  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21910  return __ret;
21911 }
21912 #else
21913 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
21914  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21915  poly8x16_t __ret;
21916  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21917  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21918  return __ret;
21919 }
21920 #endif
21921 
21922 #ifdef __LITTLE_ENDIAN__
21923 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
21924  uint8x16_t __ret;
21925  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21926  return __ret;
21927 }
21928 #else
21929 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
21930  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21931  uint8x16_t __ret;
21932  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21933  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21934  return __ret;
21935 }
21936 #endif
21937 
21938 #ifdef __LITTLE_ENDIAN__
21939 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
21940  int8x16_t __ret;
21941  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21942  return __ret;
21943 }
21944 #else
21945 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
21946  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21947  int8x16_t __ret;
21948  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
21949  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21950  return __ret;
21951 }
21952 #endif
21953 
21954 #ifdef __LITTLE_ENDIAN__
21955 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
21956  uint8x8_t __ret;
21957  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21958  return __ret;
21959 }
21960 #else
21961 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
21962  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21963  uint8x8_t __ret;
21964  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21965  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21966  return __ret;
21967 }
21968 #endif
21969 
21970 #ifdef __LITTLE_ENDIAN__
21971 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
21972  int8x8_t __ret;
21973  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
21974  return __ret;
21975 }
21976 #else
21977 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
21978  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21979  int8x8_t __ret;
21980  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
21981  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21982  return __ret;
21983 }
21984 #endif
21985 
21986 #ifdef __LITTLE_ENDIAN__
21987 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
21988  poly8x8_t __ret;
21989  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
21990  return __ret;
21991 }
21992 #else
21993 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
21994  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21995  poly8x8_t __ret;
21996  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
21997  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21998  return __ret;
21999 }
22000 #endif
22001 
22002 #ifdef __LITTLE_ENDIAN__
22003 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
22004  poly16x4_t __ret;
22005  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22006  return __ret;
22007 }
22008 #else
22009 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
22010  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22011  poly16x4_t __ret;
22012  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22013  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22014  return __ret;
22015 }
22016 #endif
22017 
22018 #ifdef __LITTLE_ENDIAN__
22019 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
22020  poly8x16_t __ret;
22021  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22022  return __ret;
22023 }
22024 #else
22025 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
22026  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22027  poly8x16_t __ret;
22028  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22029  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22030  return __ret;
22031 }
22032 #endif
22033 
22034 #ifdef __LITTLE_ENDIAN__
22035 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
22036  poly16x8_t __ret;
22037  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
22038  return __ret;
22039 }
22040 #else
22041 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
22042  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22043  poly16x8_t __ret;
22044  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
22045  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22046  return __ret;
22047 }
22048 #endif
22049 
22050 #ifdef __LITTLE_ENDIAN__
22051 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
22052  uint8x16_t __ret;
22053  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22054  return __ret;
22055 }
22056 #else
22057 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
22058  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22059  uint8x16_t __ret;
22060  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22061  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22062  return __ret;
22063 }
22064 #endif
22065 
22066 #ifdef __LITTLE_ENDIAN__
22067 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
22068  uint16x8_t __ret;
22069  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
22070  return __ret;
22071 }
22072 #else
22073 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
22074  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22075  uint16x8_t __ret;
22076  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
22077  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22078  return __ret;
22079 }
22080 #endif
22081 
22082 #ifdef __LITTLE_ENDIAN__
22083 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
22084  int8x16_t __ret;
22085  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22086  return __ret;
22087 }
22088 #else
22089 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
22090  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22091  int8x16_t __ret;
22092  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
22093  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22094  return __ret;
22095 }
22096 #endif
22097 
22098 #ifdef __LITTLE_ENDIAN__
22099 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
22100  int16x8_t __ret;
22101  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
22102  return __ret;
22103 }
22104 #else
22105 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
22106  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22107  int16x8_t __ret;
22108  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
22109  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22110  return __ret;
22111 }
22112 #endif
22113 
22114 #ifdef __LITTLE_ENDIAN__
22115 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
22116  uint8x8_t __ret;
22117  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22118  return __ret;
22119 }
22120 #else
22121 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
22122  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22123  uint8x8_t __ret;
22124  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22126  return __ret;
22127 }
22128 #endif
22129 
22130 #ifdef __LITTLE_ENDIAN__
22131 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
22132  uint16x4_t __ret;
22133  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22134  return __ret;
22135 }
22136 #else
22137 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
22138  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22139  uint16x4_t __ret;
22140  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22141  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22142  return __ret;
22143 }
22144 #endif
22145 
22146 #ifdef __LITTLE_ENDIAN__
22147 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
22148  int8x8_t __ret;
22149  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22150  return __ret;
22151 }
22152 #else
22153 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
22154  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22155  int8x8_t __ret;
22156  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22157  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22158  return __ret;
22159 }
22160 #endif
22161 
22162 #ifdef __LITTLE_ENDIAN__
22163 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
22164  int16x4_t __ret;
22165  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22166  return __ret;
22167 }
22168 #else
22169 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
22170  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22171  int16x4_t __ret;
22172  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22173  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22174  return __ret;
22175 }
22176 #endif
22177 
22178 #ifdef __LITTLE_ENDIAN__
22179 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
22180  poly8x8_t __ret;
22181  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22182  return __ret;
22183 }
22184 #else
22185 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
22186  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22187  poly8x8_t __ret;
22188  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22189  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22190  return __ret;
22191 }
22192 #endif
22193 
22194 #ifdef __LITTLE_ENDIAN__
22195 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
22196  poly16x4_t __ret;
22197  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22198  return __ret;
22199 }
22200 #else
22201 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
22202  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22203  poly16x4_t __ret;
22204  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22206  return __ret;
22207 }
22208 #endif
22209 
22210 #ifdef __LITTLE_ENDIAN__
22211 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
22212  poly8x16_t __ret;
22213  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22214  return __ret;
22215 }
22216 #else
22217 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
22218  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22219  poly8x16_t __ret;
22220  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22221  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22222  return __ret;
22223 }
22224 #endif
22225 
22226 #ifdef __LITTLE_ENDIAN__
22227 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
22228  poly16x8_t __ret;
22229  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22230  return __ret;
22231 }
22232 #else
22233 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
22234  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22235  poly16x8_t __ret;
22236  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22237  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22238  return __ret;
22239 }
22240 #endif
22241 
22242 #ifdef __LITTLE_ENDIAN__
22243 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
22244  uint8x16_t __ret;
22245  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22246  return __ret;
22247 }
22248 #else
22249 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
22250  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22251  uint8x16_t __ret;
22252  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22253  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22254  return __ret;
22255 }
22256 #endif
22257 
22258 #ifdef __LITTLE_ENDIAN__
22259 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
22260  uint32x4_t __ret;
22261  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22262  return __ret;
22263 }
22264 #else
22265 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
22266  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22267  uint32x4_t __ret;
22268  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22269  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22270  return __ret;
22271 }
22272 #endif
22273 
22274 #ifdef __LITTLE_ENDIAN__
22275 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
22276  uint16x8_t __ret;
22277  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22278  return __ret;
22279 }
22280 #else
22281 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
22282  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22283  uint16x8_t __ret;
22284  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22285  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22286  return __ret;
22287 }
22288 #endif
22289 
22290 #ifdef __LITTLE_ENDIAN__
22291 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
22292  int8x16_t __ret;
22293  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22294  return __ret;
22295 }
22296 #else
22297 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
22298  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22299  int8x16_t __ret;
22300  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
22301  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22302  return __ret;
22303 }
22304 #endif
22305 
22306 #ifdef __LITTLE_ENDIAN__
22307 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
22308  float32x4_t __ret;
22309  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22310  return __ret;
22311 }
22312 #else
22313 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
22314  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22315  float32x4_t __ret;
22316  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22317  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22318  return __ret;
22319 }
22320 #endif
22321 
22322 #ifdef __LITTLE_ENDIAN__
22323 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
22324  int32x4_t __ret;
22325  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
22326  return __ret;
22327 }
22328 #else
22329 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
22330  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22331  int32x4_t __ret;
22332  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
22333  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22334  return __ret;
22335 }
22336 #endif
22337 
22338 #ifdef __LITTLE_ENDIAN__
22339 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
22340  int16x8_t __ret;
22341  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
22342  return __ret;
22343 }
22344 #else
22345 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
22346  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22347  int16x8_t __ret;
22348  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
22349  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22350  return __ret;
22351 }
22352 #endif
22353 
22354 #ifdef __LITTLE_ENDIAN__
22355 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
22356  uint8x8_t __ret;
22357  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22358  return __ret;
22359 }
22360 #else
22361 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
22362  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22363  uint8x8_t __ret;
22364  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22365  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22366  return __ret;
22367 }
22368 #endif
22369 
22370 #ifdef __LITTLE_ENDIAN__
22371 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
22372  uint32x2_t __ret;
22373  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22374  return __ret;
22375 }
22376 #else
22377 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
22378  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22379  uint32x2_t __ret;
22380  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22381  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22382  return __ret;
22383 }
22384 #endif
22385 
22386 #ifdef __LITTLE_ENDIAN__
22387 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
22388  uint16x4_t __ret;
22389  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22390  return __ret;
22391 }
22392 #else
22393 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
22394  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22395  uint16x4_t __ret;
22396  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22397  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22398  return __ret;
22399 }
22400 #endif
22401 
22402 #ifdef __LITTLE_ENDIAN__
22403 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
22404  int8x8_t __ret;
22405  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22406  return __ret;
22407 }
22408 #else
22409 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
22410  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22411  int8x8_t __ret;
22412  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
22413  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22414  return __ret;
22415 }
22416 #endif
22417 
22418 #ifdef __LITTLE_ENDIAN__
22419 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
22420  float32x2_t __ret;
22421  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22422  return __ret;
22423 }
22424 #else
22425 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
22426  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22427  float32x2_t __ret;
22428  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22429  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22430  return __ret;
22431 }
22432 #endif
22433 
22434 #ifdef __LITTLE_ENDIAN__
22435 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
22436  int32x2_t __ret;
22437  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
22438  return __ret;
22439 }
22440 #else
22441 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
22442  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22443  int32x2_t __ret;
22444  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
22445  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22446  return __ret;
22447 }
22448 #endif
22449 
22450 #ifdef __LITTLE_ENDIAN__
22451 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
22452  int16x4_t __ret;
22453  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22454  return __ret;
22455 }
22456 #else
22457 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
22458  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22459  int16x4_t __ret;
22460  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
22461  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22462  return __ret;
22463 }
22464 #endif
22465 
22466 #ifdef __LITTLE_ENDIAN__
22467 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
22468  uint8x16_t __ret;
22469  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22470  return __ret;
22471 }
22472 #else
22473 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
22474  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22475  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22476  uint8x16_t __ret;
22477  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22478  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22479  return __ret;
22480 }
22481 #endif
22482 
22483 #ifdef __LITTLE_ENDIAN__
22484 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
22485  uint32x4_t __ret;
22486  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22487  return __ret;
22488 }
22489 #else
22490 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
22491  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22492  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22493  uint32x4_t __ret;
22494  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22495  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22496  return __ret;
22497 }
22498 #endif
22499 
22500 #ifdef __LITTLE_ENDIAN__
22501 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
22502  uint16x8_t __ret;
22503  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22504  return __ret;
22505 }
22506 #else
22507 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
22508  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22509  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22510  uint16x8_t __ret;
22511  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22512  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22513  return __ret;
22514 }
22515 #endif
22516 
22517 #ifdef __LITTLE_ENDIAN__
22518 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
22519  int8x16_t __ret;
22520  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22521  return __ret;
22522 }
22523 #else
22524 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
22525  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22526  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22527  int8x16_t __ret;
22528  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22529  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22530  return __ret;
22531 }
22532 #endif
22533 
22534 #ifdef __LITTLE_ENDIAN__
22535 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
22536  int32x4_t __ret;
22537  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22538  return __ret;
22539 }
22540 #else
22541 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
22542  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22543  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22544  int32x4_t __ret;
22545  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22546  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22547  return __ret;
22548 }
22549 #endif
22550 
22551 #ifdef __LITTLE_ENDIAN__
22552 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
22553  int16x8_t __ret;
22554  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22555  return __ret;
22556 }
22557 #else
22558 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
22559  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22560  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22561  int16x8_t __ret;
22562  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22563  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22564  return __ret;
22565 }
22566 #endif
22567 
22568 #ifdef __LITTLE_ENDIAN__
22569 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
22570  uint8x8_t __ret;
22571  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22572  return __ret;
22573 }
22574 #else
22575 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
22576  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22577  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22578  uint8x8_t __ret;
22579  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22581  return __ret;
22582 }
22583 #endif
22584 
22585 #ifdef __LITTLE_ENDIAN__
22586 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
22587  uint32x2_t __ret;
22588  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22589  return __ret;
22590 }
22591 #else
22592 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
22593  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22594  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22595  uint32x2_t __ret;
22596  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22597  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22598  return __ret;
22599 }
22600 #endif
22601 
22602 #ifdef __LITTLE_ENDIAN__
22603 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
22604  uint16x4_t __ret;
22605  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22606  return __ret;
22607 }
22608 #else
22609 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
22610  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22611  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22612  uint16x4_t __ret;
22613  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22614  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22615  return __ret;
22616 }
22617 #endif
22618 
22619 #ifdef __LITTLE_ENDIAN__
22620 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
22621  int8x8_t __ret;
22622  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22623  return __ret;
22624 }
22625 #else
22626 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
22627  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22628  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22629  int8x8_t __ret;
22630  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22631  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22632  return __ret;
22633 }
22634 #endif
22635 
22636 #ifdef __LITTLE_ENDIAN__
22637 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
22638  int32x2_t __ret;
22639  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22640  return __ret;
22641 }
22642 #else
22643 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
22644  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22645  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22646  int32x2_t __ret;
22647  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22648  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22649  return __ret;
22650 }
22651 #endif
22652 
22653 #ifdef __LITTLE_ENDIAN__
22654 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
22655  int16x4_t __ret;
22656  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22657  return __ret;
22658 }
22659 #else
22660 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
22661  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22662  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22663  int16x4_t __ret;
22664  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22665  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22666  return __ret;
22667 }
22668 #endif
22669 
22670 #ifdef __LITTLE_ENDIAN__
22671 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22672  uint8x16_t __ret;
22673  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22674  return __ret;
22675 }
22676 #else
22677 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22678  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22679  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22680  uint8x16_t __ret;
22681  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22682  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22683  return __ret;
22684 }
22685 #endif
22686 
22687 #ifdef __LITTLE_ENDIAN__
22688 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22689  uint32x4_t __ret;
22690  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22691  return __ret;
22692 }
22693 #else
22694 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22695  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22696  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22697  uint32x4_t __ret;
22698  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22699  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22700  return __ret;
22701 }
22702 #endif
22703 
22704 #ifdef __LITTLE_ENDIAN__
22705 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22706  uint64x2_t __ret;
22707  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22708  return __ret;
22709 }
22710 #else
22711 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22712  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22713  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22714  uint64x2_t __ret;
22715  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22716  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22717  return __ret;
22718 }
22719 #endif
22720 
22721 #ifdef __LITTLE_ENDIAN__
22722 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22723  uint16x8_t __ret;
22724  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22725  return __ret;
22726 }
22727 #else
22728 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22729  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22730  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22731  uint16x8_t __ret;
22732  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22733  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22734  return __ret;
22735 }
22736 #endif
22737 
22738 #ifdef __LITTLE_ENDIAN__
22739 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22740  int8x16_t __ret;
22741  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22742  return __ret;
22743 }
22744 #else
22745 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22746  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22747  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22748  int8x16_t __ret;
22749  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22750  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22751  return __ret;
22752 }
22753 #endif
22754 
22755 #ifdef __LITTLE_ENDIAN__
22756 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22757  int32x4_t __ret;
22758  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22759  return __ret;
22760 }
22761 #else
22762 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22763  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22764  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22765  int32x4_t __ret;
22766  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22767  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22768  return __ret;
22769 }
22770 #endif
22771 
22772 #ifdef __LITTLE_ENDIAN__
22773 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22774  int64x2_t __ret;
22775  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22776  return __ret;
22777 }
22778 #else
22779 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22780  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22781  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22782  int64x2_t __ret;
22783  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22784  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22785  return __ret;
22786 }
22787 #endif
22788 
22789 #ifdef __LITTLE_ENDIAN__
22790 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22791  int16x8_t __ret;
22792  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22793  return __ret;
22794 }
22795 #else
22796 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22797  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22798  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22799  int16x8_t __ret;
22800  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22801  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22802  return __ret;
22803 }
22804 #endif
22805 
22806 #ifdef __LITTLE_ENDIAN__
22807 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22808  uint8x8_t __ret;
22809  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22810  return __ret;
22811 }
22812 #else
22813 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22814  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22815  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22816  uint8x8_t __ret;
22817  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22818  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22819  return __ret;
22820 }
22821 #endif
22822 
22823 #ifdef __LITTLE_ENDIAN__
22824 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22825  uint32x2_t __ret;
22826  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22827  return __ret;
22828 }
22829 #else
22830 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22831  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22832  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22833  uint32x2_t __ret;
22834  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22835  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22836  return __ret;
22837 }
22838 #endif
22839 
22840 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22841  uint64x1_t __ret;
22842  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22843  return __ret;
22844 }
22845 #ifdef __LITTLE_ENDIAN__
22846 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22847  uint16x4_t __ret;
22848  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22849  return __ret;
22850 }
22851 #else
22852 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22853  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22854  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22855  uint16x4_t __ret;
22856  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22858  return __ret;
22859 }
22860 #endif
22861 
22862 #ifdef __LITTLE_ENDIAN__
22863 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
22864  int8x8_t __ret;
22865  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22866  return __ret;
22867 }
22868 #else
22869 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
22870  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22871  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22872  int8x8_t __ret;
22873  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22874  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22875  return __ret;
22876 }
22877 #endif
22878 
22879 #ifdef __LITTLE_ENDIAN__
22880 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
22881  int32x2_t __ret;
22882  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22883  return __ret;
22884 }
22885 #else
22886 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
22887  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22888  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22889  int32x2_t __ret;
22890  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22891  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22892  return __ret;
22893 }
22894 #endif
22895 
22896 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
22897  int64x1_t __ret;
22898  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22899  return __ret;
22900 }
22901 #ifdef __LITTLE_ENDIAN__
22902 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
22903  int16x4_t __ret;
22904  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22905  return __ret;
22906 }
22907 #else
22908 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
22909  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22910  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22911  int16x4_t __ret;
22912  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22913  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22914  return __ret;
22915 }
22916 #endif
22917 
22918 #ifdef __LITTLE_ENDIAN__
22919 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
22920  uint8x16_t __s0 = __p0; \
22921  uint8x16_t __ret; \
22922  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
22923  __ret; \
22924 })
22925 #else
22926 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
22927  uint8x16_t __s0 = __p0; \
22928  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22929  uint8x16_t __ret; \
22930  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
22931  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22932  __ret; \
22933 })
22934 #endif
22935 
22936 #ifdef __LITTLE_ENDIAN__
22937 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
22938  uint32x4_t __s0 = __p0; \
22939  uint32x4_t __ret; \
22940  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
22941  __ret; \
22942 })
22943 #else
22944 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
22945  uint32x4_t __s0 = __p0; \
22946  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22947  uint32x4_t __ret; \
22948  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
22949  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22950  __ret; \
22951 })
22952 #endif
22953 
22954 #ifdef __LITTLE_ENDIAN__
22955 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
22956  uint64x2_t __s0 = __p0; \
22957  uint64x2_t __ret; \
22958  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
22959  __ret; \
22960 })
22961 #else
22962 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
22963  uint64x2_t __s0 = __p0; \
22964  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22965  uint64x2_t __ret; \
22966  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
22967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22968  __ret; \
22969 })
22970 #endif
22971 
22972 #ifdef __LITTLE_ENDIAN__
22973 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
22974  uint16x8_t __s0 = __p0; \
22975  uint16x8_t __ret; \
22976  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
22977  __ret; \
22978 })
22979 #else
22980 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
22981  uint16x8_t __s0 = __p0; \
22982  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
22983  uint16x8_t __ret; \
22984  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
22985  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22986  __ret; \
22987 })
22988 #endif
22989 
22990 #ifdef __LITTLE_ENDIAN__
22991 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
22992  int8x16_t __s0 = __p0; \
22993  int8x16_t __ret; \
22994  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
22995  __ret; \
22996 })
22997 #else
22998 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
22999  int8x16_t __s0 = __p0; \
23000  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23001  int8x16_t __ret; \
23002  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
23003  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23004  __ret; \
23005 })
23006 #endif
23007 
23008 #ifdef __LITTLE_ENDIAN__
23009 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
23010  int32x4_t __s0 = __p0; \
23011  int32x4_t __ret; \
23012  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
23013  __ret; \
23014 })
23015 #else
23016 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
23017  int32x4_t __s0 = __p0; \
23018  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23019  int32x4_t __ret; \
23020  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
23021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23022  __ret; \
23023 })
23024 #endif
23025 
23026 #ifdef __LITTLE_ENDIAN__
23027 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
23028  int64x2_t __s0 = __p0; \
23029  int64x2_t __ret; \
23030  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
23031  __ret; \
23032 })
23033 #else
23034 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
23035  int64x2_t __s0 = __p0; \
23036  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23037  int64x2_t __ret; \
23038  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
23039  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23040  __ret; \
23041 })
23042 #endif
23043 
23044 #ifdef __LITTLE_ENDIAN__
23045 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
23046  int16x8_t __s0 = __p0; \
23047  int16x8_t __ret; \
23048  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
23049  __ret; \
23050 })
23051 #else
23052 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
23053  int16x8_t __s0 = __p0; \
23054  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23055  int16x8_t __ret; \
23056  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
23057  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23058  __ret; \
23059 })
23060 #endif
23061 
23062 #ifdef __LITTLE_ENDIAN__
23063 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
23064  uint8x8_t __s0 = __p0; \
23065  uint8x8_t __ret; \
23066  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
23067  __ret; \
23068 })
23069 #else
23070 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
23071  uint8x8_t __s0 = __p0; \
23072  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23073  uint8x8_t __ret; \
23074  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
23075  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23076  __ret; \
23077 })
23078 #endif
23079 
23080 #ifdef __LITTLE_ENDIAN__
23081 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
23082  uint32x2_t __s0 = __p0; \
23083  uint32x2_t __ret; \
23084  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
23085  __ret; \
23086 })
23087 #else
23088 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
23089  uint32x2_t __s0 = __p0; \
23090  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23091  uint32x2_t __ret; \
23092  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
23093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23094  __ret; \
23095 })
23096 #endif
23097 
23098 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
23099  uint64x1_t __s0 = __p0; \
23100  uint64x1_t __ret; \
23101  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
23102  __ret; \
23103 })
23104 #ifdef __LITTLE_ENDIAN__
23105 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
23106  uint16x4_t __s0 = __p0; \
23107  uint16x4_t __ret; \
23108  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
23109  __ret; \
23110 })
23111 #else
23112 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
23113  uint16x4_t __s0 = __p0; \
23114  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23115  uint16x4_t __ret; \
23116  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
23117  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23118  __ret; \
23119 })
23120 #endif
23121 
23122 #ifdef __LITTLE_ENDIAN__
23123 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
23124  int8x8_t __s0 = __p0; \
23125  int8x8_t __ret; \
23126  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
23127  __ret; \
23128 })
23129 #else
23130 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
23131  int8x8_t __s0 = __p0; \
23132  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23133  int8x8_t __ret; \
23134  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
23135  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23136  __ret; \
23137 })
23138 #endif
23139 
23140 #ifdef __LITTLE_ENDIAN__
23141 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
23142  int32x2_t __s0 = __p0; \
23143  int32x2_t __ret; \
23144  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
23145  __ret; \
23146 })
23147 #else
23148 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
23149  int32x2_t __s0 = __p0; \
23150  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23151  int32x2_t __ret; \
23152  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
23153  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23154  __ret; \
23155 })
23156 #endif
23157 
23158 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
23159  int64x1_t __s0 = __p0; \
23160  int64x1_t __ret; \
23161  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
23162  __ret; \
23163 })
23164 #ifdef __LITTLE_ENDIAN__
23165 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
23166  int16x4_t __s0 = __p0; \
23167  int16x4_t __ret; \
23168  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
23169  __ret; \
23170 })
23171 #else
23172 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
23173  int16x4_t __s0 = __p0; \
23174  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23175  int16x4_t __ret; \
23176  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
23177  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23178  __ret; \
23179 })
23180 #endif
23181 
23182 #ifdef __LITTLE_ENDIAN__
23183 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23184  uint32x4_t __s0 = __p0; \
23185  uint16x4_t __ret; \
23186  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
23187  __ret; \
23188 })
23189 #else
23190 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23191  uint32x4_t __s0 = __p0; \
23192  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23193  uint16x4_t __ret; \
23194  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23195  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23196  __ret; \
23197 })
23198 #define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
23199  uint32x4_t __s0 = __p0; \
23200  uint16x4_t __ret; \
23201  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
23202  __ret; \
23203 })
23204 #endif
23205 
23206 #ifdef __LITTLE_ENDIAN__
23207 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23208  uint64x2_t __s0 = __p0; \
23209  uint32x2_t __ret; \
23210  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
23211  __ret; \
23212 })
23213 #else
23214 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23215  uint64x2_t __s0 = __p0; \
23216  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23217  uint32x2_t __ret; \
23218  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23219  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23220  __ret; \
23221 })
23222 #define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
23223  uint64x2_t __s0 = __p0; \
23224  uint32x2_t __ret; \
23225  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
23226  __ret; \
23227 })
23228 #endif
23229 
23230 #ifdef __LITTLE_ENDIAN__
23231 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23232  uint16x8_t __s0 = __p0; \
23233  uint8x8_t __ret; \
23234  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
23235  __ret; \
23236 })
23237 #else
23238 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23239  uint16x8_t __s0 = __p0; \
23240  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23241  uint8x8_t __ret; \
23242  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23243  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23244  __ret; \
23245 })
23246 #define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
23247  uint16x8_t __s0 = __p0; \
23248  uint8x8_t __ret; \
23249  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
23250  __ret; \
23251 })
23252 #endif
23253 
23254 #ifdef __LITTLE_ENDIAN__
23255 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23256  int32x4_t __s0 = __p0; \
23257  int16x4_t __ret; \
23258  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
23259  __ret; \
23260 })
23261 #else
23262 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23263  int32x4_t __s0 = __p0; \
23264  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23265  int16x4_t __ret; \
23266  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23267  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23268  __ret; \
23269 })
23270 #define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
23271  int32x4_t __s0 = __p0; \
23272  int16x4_t __ret; \
23273  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
23274  __ret; \
23275 })
23276 #endif
23277 
23278 #ifdef __LITTLE_ENDIAN__
23279 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23280  int64x2_t __s0 = __p0; \
23281  int32x2_t __ret; \
23282  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
23283  __ret; \
23284 })
23285 #else
23286 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23287  int64x2_t __s0 = __p0; \
23288  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23289  int32x2_t __ret; \
23290  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23291  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23292  __ret; \
23293 })
23294 #define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
23295  int64x2_t __s0 = __p0; \
23296  int32x2_t __ret; \
23297  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
23298  __ret; \
23299 })
23300 #endif
23301 
23302 #ifdef __LITTLE_ENDIAN__
23303 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23304  int16x8_t __s0 = __p0; \
23305  int8x8_t __ret; \
23306  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
23307  __ret; \
23308 })
23309 #else
23310 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23311  int16x8_t __s0 = __p0; \
23312  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23313  int8x8_t __ret; \
23314  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23315  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23316  __ret; \
23317 })
23318 #define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
23319  int16x8_t __s0 = __p0; \
23320  int8x8_t __ret; \
23321  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
23322  __ret; \
23323 })
23324 #endif
23325 
23326 #ifdef __LITTLE_ENDIAN__
23327 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
23328  uint32x4_t __ret;
23329  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
23330  return __ret;
23331 }
23332 #else
23333 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
23334  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23335  uint32x4_t __ret;
23336  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
23337  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23338  return __ret;
23339 }
23340 #endif
23341 
23342 #ifdef __LITTLE_ENDIAN__
23343 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
23344  float32x4_t __ret;
23345  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
23346  return __ret;
23347 }
23348 #else
23349 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
23350  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23351  float32x4_t __ret;
23352  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
23353  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23354  return __ret;
23355 }
23356 #endif
23357 
23358 #ifdef __LITTLE_ENDIAN__
23359 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
23360  uint32x2_t __ret;
23361  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
23362  return __ret;
23363 }
23364 #else
23365 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
23366  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23367  uint32x2_t __ret;
23368  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
23369  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23370  return __ret;
23371 }
23372 #endif
23373 
23374 #ifdef __LITTLE_ENDIAN__
23375 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
23376  float32x2_t __ret;
23377  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
23378  return __ret;
23379 }
23380 #else
23381 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
23382  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23383  float32x2_t __ret;
23384  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
23385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23386  return __ret;
23387 }
23388 #endif
23389 
23390 #ifdef __LITTLE_ENDIAN__
23391 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
23392  float32x4_t __ret;
23393  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
23394  return __ret;
23395 }
23396 #else
23397 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
23398  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23399  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23400  float32x4_t __ret;
23401  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
23402  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23403  return __ret;
23404 }
23405 #endif
23406 
23407 #ifdef __LITTLE_ENDIAN__
23408 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
23409  float32x2_t __ret;
23410  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
23411  return __ret;
23412 }
23413 #else
23414 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
23415  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23416  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23417  float32x2_t __ret;
23418  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
23419  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23420  return __ret;
23421 }
23422 #endif
23423 
23424 #ifdef __LITTLE_ENDIAN__
23425 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23426  uint8x16_t __s0 = __p0; \
23427  uint8x16_t __s1 = __p1; \
23428  uint8x16_t __ret; \
23429  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23430  __ret; \
23431 })
23432 #else
23433 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23434  uint8x16_t __s0 = __p0; \
23435  uint8x16_t __s1 = __p1; \
23436  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23437  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23438  uint8x16_t __ret; \
23439  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23441  __ret; \
23442 })
23443 #endif
23444 
23445 #ifdef __LITTLE_ENDIAN__
23446 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23447  uint32x4_t __s0 = __p0; \
23448  uint32x4_t __s1 = __p1; \
23449  uint32x4_t __ret; \
23450  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
23451  __ret; \
23452 })
23453 #else
23454 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23455  uint32x4_t __s0 = __p0; \
23456  uint32x4_t __s1 = __p1; \
23457  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23458  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23459  uint32x4_t __ret; \
23460  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
23461  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23462  __ret; \
23463 })
23464 #endif
23465 
23466 #ifdef __LITTLE_ENDIAN__
23467 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23468  uint64x2_t __s0 = __p0; \
23469  uint64x2_t __s1 = __p1; \
23470  uint64x2_t __ret; \
23471  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
23472  __ret; \
23473 })
23474 #else
23475 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23476  uint64x2_t __s0 = __p0; \
23477  uint64x2_t __s1 = __p1; \
23478  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23479  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23480  uint64x2_t __ret; \
23481  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
23482  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23483  __ret; \
23484 })
23485 #endif
23486 
23487 #ifdef __LITTLE_ENDIAN__
23488 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23489  uint16x8_t __s0 = __p0; \
23490  uint16x8_t __s1 = __p1; \
23491  uint16x8_t __ret; \
23492  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
23493  __ret; \
23494 })
23495 #else
23496 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
23497  uint16x8_t __s0 = __p0; \
23498  uint16x8_t __s1 = __p1; \
23499  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23500  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23501  uint16x8_t __ret; \
23502  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
23503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23504  __ret; \
23505 })
23506 #endif
23507 
23508 #ifdef __LITTLE_ENDIAN__
23509 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23510  int8x16_t __s0 = __p0; \
23511  int8x16_t __s1 = __p1; \
23512  int8x16_t __ret; \
23513  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
23514  __ret; \
23515 })
23516 #else
23517 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
23518  int8x16_t __s0 = __p0; \
23519  int8x16_t __s1 = __p1; \
23520  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23521  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23522  int8x16_t __ret; \
23523  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
23524  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23525  __ret; \
23526 })
23527 #endif
23528 
23529 #ifdef __LITTLE_ENDIAN__
23530 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
23531  int32x4_t __s0 = __p0; \
23532  int32x4_t __s1 = __p1; \
23533  int32x4_t __ret; \
23534  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
23535  __ret; \
23536 })
23537 #else
23538 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
23539  int32x4_t __s0 = __p0; \
23540  int32x4_t __s1 = __p1; \
23541  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23542  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23543  int32x4_t __ret; \
23544  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
23545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23546  __ret; \
23547 })
23548 #endif
23549 
23550 #ifdef __LITTLE_ENDIAN__
23551 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
23552  int64x2_t __s0 = __p0; \
23553  int64x2_t __s1 = __p1; \
23554  int64x2_t __ret; \
23555  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
23556  __ret; \
23557 })
23558 #else
23559 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
23560  int64x2_t __s0 = __p0; \
23561  int64x2_t __s1 = __p1; \
23562  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23563  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23564  int64x2_t __ret; \
23565  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
23566  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23567  __ret; \
23568 })
23569 #endif
23570 
23571 #ifdef __LITTLE_ENDIAN__
23572 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
23573  int16x8_t __s0 = __p0; \
23574  int16x8_t __s1 = __p1; \
23575  int16x8_t __ret; \
23576  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
23577  __ret; \
23578 })
23579 #else
23580 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
23581  int16x8_t __s0 = __p0; \
23582  int16x8_t __s1 = __p1; \
23583  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23584  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23585  int16x8_t __ret; \
23586  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
23587  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23588  __ret; \
23589 })
23590 #endif
23591 
23592 #ifdef __LITTLE_ENDIAN__
23593 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
23594  uint8x8_t __s0 = __p0; \
23595  uint8x8_t __s1 = __p1; \
23596  uint8x8_t __ret; \
23597  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
23598  __ret; \
23599 })
23600 #else
23601 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
23602  uint8x8_t __s0 = __p0; \
23603  uint8x8_t __s1 = __p1; \
23604  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23605  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23606  uint8x8_t __ret; \
23607  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
23608  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23609  __ret; \
23610 })
23611 #endif
23612 
23613 #ifdef __LITTLE_ENDIAN__
23614 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
23615  uint32x2_t __s0 = __p0; \
23616  uint32x2_t __s1 = __p1; \
23617  uint32x2_t __ret; \
23618  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
23619  __ret; \
23620 })
23621 #else
23622 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
23623  uint32x2_t __s0 = __p0; \
23624  uint32x2_t __s1 = __p1; \
23625  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23626  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23627  uint32x2_t __ret; \
23628  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
23629  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23630  __ret; \
23631 })
23632 #endif
23633 
23634 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
23635  uint64x1_t __s0 = __p0; \
23636  uint64x1_t __s1 = __p1; \
23637  uint64x1_t __ret; \
23638  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
23639  __ret; \
23640 })
23641 #ifdef __LITTLE_ENDIAN__
23642 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
23643  uint16x4_t __s0 = __p0; \
23644  uint16x4_t __s1 = __p1; \
23645  uint16x4_t __ret; \
23646  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
23647  __ret; \
23648 })
23649 #else
23650 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
23651  uint16x4_t __s0 = __p0; \
23652  uint16x4_t __s1 = __p1; \
23653  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23654  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23655  uint16x4_t __ret; \
23656  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
23657  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23658  __ret; \
23659 })
23660 #endif
23661 
23662 #ifdef __LITTLE_ENDIAN__
23663 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
23664  int8x8_t __s0 = __p0; \
23665  int8x8_t __s1 = __p1; \
23666  int8x8_t __ret; \
23667  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
23668  __ret; \
23669 })
23670 #else
23671 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
23672  int8x8_t __s0 = __p0; \
23673  int8x8_t __s1 = __p1; \
23674  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23675  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23676  int8x8_t __ret; \
23677  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
23678  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23679  __ret; \
23680 })
23681 #endif
23682 
23683 #ifdef __LITTLE_ENDIAN__
23684 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
23685  int32x2_t __s0 = __p0; \
23686  int32x2_t __s1 = __p1; \
23687  int32x2_t __ret; \
23688  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
23689  __ret; \
23690 })
23691 #else
23692 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
23693  int32x2_t __s0 = __p0; \
23694  int32x2_t __s1 = __p1; \
23695  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23696  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23697  int32x2_t __ret; \
23698  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
23699  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23700  __ret; \
23701 })
23702 #endif
23703 
23704 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
23705  int64x1_t __s0 = __p0; \
23706  int64x1_t __s1 = __p1; \
23707  int64x1_t __ret; \
23708  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
23709  __ret; \
23710 })
23711 #ifdef __LITTLE_ENDIAN__
23712 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
23713  int16x4_t __s0 = __p0; \
23714  int16x4_t __s1 = __p1; \
23715  int16x4_t __ret; \
23716  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
23717  __ret; \
23718 })
23719 #else
23720 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
23721  int16x4_t __s0 = __p0; \
23722  int16x4_t __s1 = __p1; \
23723  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23724  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23725  int16x4_t __ret; \
23726  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
23727  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23728  __ret; \
23729 })
23730 #endif
23731 
23732 #ifdef __LITTLE_ENDIAN__
23733 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23734  uint16x4_t __ret;
23735  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
23736  return __ret;
23737 }
23738 #else
23739 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23740  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23741  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23742  uint16x4_t __ret;
23743  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
23744  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23745  return __ret;
23746 }
23747 __ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
23748  uint16x4_t __ret;
23749  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
23750  return __ret;
23751 }
23752 #endif
23753 
23754 #ifdef __LITTLE_ENDIAN__
23755 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23756  uint32x2_t __ret;
23757  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
23758  return __ret;
23759 }
23760 #else
23761 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23762  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23763  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23764  uint32x2_t __ret;
23765  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
23766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23767  return __ret;
23768 }
23769 __ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
23770  uint32x2_t __ret;
23771  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
23772  return __ret;
23773 }
23774 #endif
23775 
23776 #ifdef __LITTLE_ENDIAN__
23777 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23778  uint8x8_t __ret;
23779  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
23780  return __ret;
23781 }
23782 #else
23783 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23784  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
23785  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
23786  uint8x8_t __ret;
23787  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
23788  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
23789  return __ret;
23790 }
23791 __ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
23792  uint8x8_t __ret;
23793  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
23794  return __ret;
23795 }
23796 #endif
23797 
23798 #ifdef __LITTLE_ENDIAN__
23799 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23800  int16x4_t __ret;
23801  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
23802  return __ret;
23803 }
23804 #else
23805 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23806  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23807  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23808  int16x4_t __ret;
23809  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
23810  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23811  return __ret;
23812 }
23813 __ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
23814  int16x4_t __ret;
23815  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
23816  return __ret;
23817 }
23818 #endif
23819 
23820 #ifdef __LITTLE_ENDIAN__
23821 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23822  int32x2_t __ret;
23823  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
23824  return __ret;
23825 }
23826 #else
23827 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23828  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
23829  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
23830  int32x2_t __ret;
23831  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
23832  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23833  return __ret;
23834 }
23835 __ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
23836  int32x2_t __ret;
23837  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
23838  return __ret;
23839 }
23840 #endif
23841 
23842 #ifdef __LITTLE_ENDIAN__
23843 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23844  int8x8_t __ret;
23845  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
23846  return __ret;
23847 }
23848 #else
23849 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23850  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
23851  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
23852  int8x8_t __ret;
23853  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
23854  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
23855  return __ret;
23856 }
23857 __ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
23858  int8x8_t __ret;
23859  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
23860  return __ret;
23861 }
23862 #endif
23863 
23864 #ifdef __LITTLE_ENDIAN__
23865 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23866  poly8_t __s0 = __p0; \
23867  poly8x8_t __s1 = __p1; \
23868  poly8x8_t __ret; \
23869  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__s1, __p2); \
23870  __ret; \
23871 })
23872 #else
23873 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23874  poly8_t __s0 = __p0; \
23875  poly8x8_t __s1 = __p1; \
23876  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23877  poly8x8_t __ret; \
23878  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__rev1, __p2); \
23879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23880  __ret; \
23881 })
23882 #define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23883  poly8_t __s0 = __p0; \
23884  poly8x8_t __s1 = __p1; \
23885  poly8x8_t __ret; \
23886  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (poly8x8_t)__s1, __p2); \
23887  __ret; \
23888 })
23889 #endif
23890 
23891 #ifdef __LITTLE_ENDIAN__
23892 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23893  poly16_t __s0 = __p0; \
23894  poly16x4_t __s1 = __p1; \
23895  poly16x4_t __ret; \
23896  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__s1, __p2); \
23897  __ret; \
23898 })
23899 #else
23900 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23901  poly16_t __s0 = __p0; \
23902  poly16x4_t __s1 = __p1; \
23903  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23904  poly16x4_t __ret; \
23905  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__rev1, __p2); \
23906  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23907  __ret; \
23908 })
23909 #define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23910  poly16_t __s0 = __p0; \
23911  poly16x4_t __s1 = __p1; \
23912  poly16x4_t __ret; \
23913  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (poly16x4_t)__s1, __p2); \
23914  __ret; \
23915 })
23916 #endif
23917 
23918 #ifdef __LITTLE_ENDIAN__
23919 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23920  poly8_t __s0 = __p0; \
23921  poly8x16_t __s1 = __p1; \
23922  poly8x16_t __ret; \
23923  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__s1, __p2); \
23924  __ret; \
23925 })
23926 #else
23927 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23928  poly8_t __s0 = __p0; \
23929  poly8x16_t __s1 = __p1; \
23930  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23931  poly8x16_t __ret; \
23932  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__rev1, __p2); \
23933  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23934  __ret; \
23935 })
23936 #define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
23937  poly8_t __s0 = __p0; \
23938  poly8x16_t __s1 = __p1; \
23939  poly8x16_t __ret; \
23940  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (poly8x16_t)__s1, __p2); \
23941  __ret; \
23942 })
23943 #endif
23944 
23945 #ifdef __LITTLE_ENDIAN__
23946 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23947  poly16_t __s0 = __p0; \
23948  poly16x8_t __s1 = __p1; \
23949  poly16x8_t __ret; \
23950  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__s1, __p2); \
23951  __ret; \
23952 })
23953 #else
23954 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23955  poly16_t __s0 = __p0; \
23956  poly16x8_t __s1 = __p1; \
23957  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23958  poly16x8_t __ret; \
23959  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__rev1, __p2); \
23960  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23961  __ret; \
23962 })
23963 #define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
23964  poly16_t __s0 = __p0; \
23965  poly16x8_t __s1 = __p1; \
23966  poly16x8_t __ret; \
23967  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (poly16x8_t)__s1, __p2); \
23968  __ret; \
23969 })
23970 #endif
23971 
23972 #ifdef __LITTLE_ENDIAN__
23973 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23974  uint8_t __s0 = __p0; \
23975  uint8x16_t __s1 = __p1; \
23976  uint8x16_t __ret; \
23977  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23978  __ret; \
23979 })
23980 #else
23981 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23982  uint8_t __s0 = __p0; \
23983  uint8x16_t __s1 = __p1; \
23984  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23985  uint8x16_t __ret; \
23986  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
23987  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23988  __ret; \
23989 })
23990 #define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
23991  uint8_t __s0 = __p0; \
23992  uint8x16_t __s1 = __p1; \
23993  uint8x16_t __ret; \
23994  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
23995  __ret; \
23996 })
23997 #endif
23998 
23999 #ifdef __LITTLE_ENDIAN__
24000 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24001  uint32_t __s0 = __p0; \
24002  uint32x4_t __s1 = __p1; \
24003  uint32x4_t __ret; \
24004  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
24005  __ret; \
24006 })
24007 #else
24008 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24009  uint32_t __s0 = __p0; \
24010  uint32x4_t __s1 = __p1; \
24011  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24012  uint32x4_t __ret; \
24013  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__rev1, __p2); \
24014  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24015  __ret; \
24016 })
24017 #define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24018  uint32_t __s0 = __p0; \
24019  uint32x4_t __s1 = __p1; \
24020  uint32x4_t __ret; \
24021  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
24022  __ret; \
24023 })
24024 #endif
24025 
24026 #ifdef __LITTLE_ENDIAN__
24027 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
24028  uint64_t __s0 = __p0; \
24029  uint64x2_t __s1 = __p1; \
24030  uint64x2_t __ret; \
24031  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24032  __ret; \
24033 })
24034 #else
24035 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
24036  uint64_t __s0 = __p0; \
24037  uint64x2_t __s1 = __p1; \
24038  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24039  uint64x2_t __ret; \
24040  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__rev1, __p2); \
24041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24042  __ret; \
24043 })
24044 #define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
24045  uint64_t __s0 = __p0; \
24046  uint64x2_t __s1 = __p1; \
24047  uint64x2_t __ret; \
24048  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24049  __ret; \
24050 })
24051 #endif
24052 
24053 #ifdef __LITTLE_ENDIAN__
24054 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24055  uint16_t __s0 = __p0; \
24056  uint16x8_t __s1 = __p1; \
24057  uint16x8_t __ret; \
24058  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24059  __ret; \
24060 })
24061 #else
24062 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24063  uint16_t __s0 = __p0; \
24064  uint16x8_t __s1 = __p1; \
24065  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24066  uint16x8_t __ret; \
24067  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__rev1, __p2); \
24068  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24069  __ret; \
24070 })
24071 #define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24072  uint16_t __s0 = __p0; \
24073  uint16x8_t __s1 = __p1; \
24074  uint16x8_t __ret; \
24075  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24076  __ret; \
24077 })
24078 #endif
24079 
24080 #ifdef __LITTLE_ENDIAN__
24081 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24082  int8_t __s0 = __p0; \
24083  int8x16_t __s1 = __p1; \
24084  int8x16_t __ret; \
24085  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
24086  __ret; \
24087 })
24088 #else
24089 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24090  int8_t __s0 = __p0; \
24091  int8x16_t __s1 = __p1; \
24092  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24093  int8x16_t __ret; \
24094  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
24095  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24096  __ret; \
24097 })
24098 #define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24099  int8_t __s0 = __p0; \
24100  int8x16_t __s1 = __p1; \
24101  int8x16_t __ret; \
24102  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
24103  __ret; \
24104 })
24105 #endif
24106 
24107 #ifdef __LITTLE_ENDIAN__
24108 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24109  float32_t __s0 = __p0; \
24110  float32x4_t __s1 = __p1; \
24111  float32x4_t __ret; \
24112  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__s1, __p2); \
24113  __ret; \
24114 })
24115 #else
24116 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24117  float32_t __s0 = __p0; \
24118  float32x4_t __s1 = __p1; \
24119  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24120  float32x4_t __ret; \
24121  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__rev1, __p2); \
24122  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24123  __ret; \
24124 })
24125 #define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24126  float32_t __s0 = __p0; \
24127  float32x4_t __s1 = __p1; \
24128  float32x4_t __ret; \
24129  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (float32x4_t)__s1, __p2); \
24130  __ret; \
24131 })
24132 #endif
24133 
24134 #ifdef __LITTLE_ENDIAN__
24135 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24136  int32_t __s0 = __p0; \
24137  int32x4_t __s1 = __p1; \
24138  int32x4_t __ret; \
24139  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
24140  __ret; \
24141 })
24142 #else
24143 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24144  int32_t __s0 = __p0; \
24145  int32x4_t __s1 = __p1; \
24146  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24147  int32x4_t __ret; \
24148  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__rev1, __p2); \
24149  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24150  __ret; \
24151 })
24152 #define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24153  int32_t __s0 = __p0; \
24154  int32x4_t __s1 = __p1; \
24155  int32x4_t __ret; \
24156  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int32x4_t)__s1, __p2); \
24157  __ret; \
24158 })
24159 #endif
24160 
24161 #ifdef __LITTLE_ENDIAN__
24162 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24163  int64_t __s0 = __p0; \
24164  int64x2_t __s1 = __p1; \
24165  int64x2_t __ret; \
24166  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24167  __ret; \
24168 })
24169 #else
24170 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24171  int64_t __s0 = __p0; \
24172  int64x2_t __s1 = __p1; \
24173  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24174  int64x2_t __ret; \
24175  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__rev1, __p2); \
24176  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24177  __ret; \
24178 })
24179 #define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24180  int64_t __s0 = __p0; \
24181  int64x2_t __s1 = __p1; \
24182  int64x2_t __ret; \
24183  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int64x2_t)__s1, __p2); \
24184  __ret; \
24185 })
24186 #endif
24187 
24188 #ifdef __LITTLE_ENDIAN__
24189 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24190  int16_t __s0 = __p0; \
24191  int16x8_t __s1 = __p1; \
24192  int16x8_t __ret; \
24193  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24194  __ret; \
24195 })
24196 #else
24197 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24198  int16_t __s0 = __p0; \
24199  int16x8_t __s1 = __p1; \
24200  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24201  int16x8_t __ret; \
24202  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__rev1, __p2); \
24203  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24204  __ret; \
24205 })
24206 #define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24207  int16_t __s0 = __p0; \
24208  int16x8_t __s1 = __p1; \
24209  int16x8_t __ret; \
24210  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int16x8_t)__s1, __p2); \
24211  __ret; \
24212 })
24213 #endif
24214 
24215 #ifdef __LITTLE_ENDIAN__
24216 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24217  uint8_t __s0 = __p0; \
24218  uint8x8_t __s1 = __p1; \
24219  uint8x8_t __ret; \
24220  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24221  __ret; \
24222 })
24223 #else
24224 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24225  uint8_t __s0 = __p0; \
24226  uint8x8_t __s1 = __p1; \
24227  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24228  uint8x8_t __ret; \
24229  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
24230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24231  __ret; \
24232 })
24233 #define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
24234  uint8_t __s0 = __p0; \
24235  uint8x8_t __s1 = __p1; \
24236  uint8x8_t __ret; \
24237  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24238  __ret; \
24239 })
24240 #endif
24241 
24242 #ifdef __LITTLE_ENDIAN__
24243 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24244  uint32_t __s0 = __p0; \
24245  uint32x2_t __s1 = __p1; \
24246  uint32x2_t __ret; \
24247  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24248  __ret; \
24249 })
24250 #else
24251 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24252  uint32_t __s0 = __p0; \
24253  uint32x2_t __s1 = __p1; \
24254  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24255  uint32x2_t __ret; \
24256  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__rev1, __p2); \
24257  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24258  __ret; \
24259 })
24260 #define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
24261  uint32_t __s0 = __p0; \
24262  uint32x2_t __s1 = __p1; \
24263  uint32x2_t __ret; \
24264  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24265  __ret; \
24266 })
24267 #endif
24268 
24269 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
24270  uint64_t __s0 = __p0; \
24271  uint64x1_t __s1 = __p1; \
24272  uint64x1_t __ret; \
24273  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int64x1_t)__s1, __p2); \
24274  __ret; \
24275 })
24276 #ifdef __LITTLE_ENDIAN__
24277 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24278  uint16_t __s0 = __p0; \
24279  uint16x4_t __s1 = __p1; \
24280  uint16x4_t __ret; \
24281  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24282  __ret; \
24283 })
24284 #else
24285 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24286  uint16_t __s0 = __p0; \
24287  uint16x4_t __s1 = __p1; \
24288  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24289  uint16x4_t __ret; \
24290  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__rev1, __p2); \
24291  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24292  __ret; \
24293 })
24294 #define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
24295  uint16_t __s0 = __p0; \
24296  uint16x4_t __s1 = __p1; \
24297  uint16x4_t __ret; \
24298  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24299  __ret; \
24300 })
24301 #endif
24302 
24303 #ifdef __LITTLE_ENDIAN__
24304 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24305  int8_t __s0 = __p0; \
24306  int8x8_t __s1 = __p1; \
24307  int8x8_t __ret; \
24308  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24309  __ret; \
24310 })
24311 #else
24312 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24313  int8_t __s0 = __p0; \
24314  int8x8_t __s1 = __p1; \
24315  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24316  int8x8_t __ret; \
24317  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
24318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24319  __ret; \
24320 })
24321 #define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
24322  int8_t __s0 = __p0; \
24323  int8x8_t __s1 = __p1; \
24324  int8x8_t __ret; \
24325  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
24326  __ret; \
24327 })
24328 #endif
24329 
24330 #ifdef __LITTLE_ENDIAN__
24331 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24332  float32_t __s0 = __p0; \
24333  float32x2_t __s1 = __p1; \
24334  float32x2_t __ret; \
24335  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__s1, __p2); \
24336  __ret; \
24337 })
24338 #else
24339 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24340  float32_t __s0 = __p0; \
24341  float32x2_t __s1 = __p1; \
24342  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24343  float32x2_t __ret; \
24344  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__rev1, __p2); \
24345  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24346  __ret; \
24347 })
24348 #define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
24349  float32_t __s0 = __p0; \
24350  float32x2_t __s1 = __p1; \
24351  float32x2_t __ret; \
24352  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (float32x2_t)__s1, __p2); \
24353  __ret; \
24354 })
24355 #endif
24356 
24357 #ifdef __LITTLE_ENDIAN__
24358 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24359  int32_t __s0 = __p0; \
24360  int32x2_t __s1 = __p1; \
24361  int32x2_t __ret; \
24362  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24363  __ret; \
24364 })
24365 #else
24366 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24367  int32_t __s0 = __p0; \
24368  int32x2_t __s1 = __p1; \
24369  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24370  int32x2_t __ret; \
24371  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__rev1, __p2); \
24372  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24373  __ret; \
24374 })
24375 #define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
24376  int32_t __s0 = __p0; \
24377  int32x2_t __s1 = __p1; \
24378  int32x2_t __ret; \
24379  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int32x2_t)__s1, __p2); \
24380  __ret; \
24381 })
24382 #endif
24383 
24384 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
24385  int64_t __s0 = __p0; \
24386  int64x1_t __s1 = __p1; \
24387  int64x1_t __ret; \
24388  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int64x1_t)__s1, __p2); \
24389  __ret; \
24390 })
24391 #ifdef __LITTLE_ENDIAN__
24392 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24393  int16_t __s0 = __p0; \
24394  int16x4_t __s1 = __p1; \
24395  int16x4_t __ret; \
24396  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24397  __ret; \
24398 })
24399 #else
24400 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24401  int16_t __s0 = __p0; \
24402  int16x4_t __s1 = __p1; \
24403  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24404  int16x4_t __ret; \
24405  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__rev1, __p2); \
24406  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24407  __ret; \
24408 })
24409 #define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
24410  int16_t __s0 = __p0; \
24411  int16x4_t __s1 = __p1; \
24412  int16x4_t __ret; \
24413  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int16x4_t)__s1, __p2); \
24414  __ret; \
24415 })
24416 #endif
24417 
24418 #ifdef __LITTLE_ENDIAN__
24419 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
24420  uint8x16_t __ret;
24421  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
24422  return __ret;
24423 }
24424 #else
24425 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
24426  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24427  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24428  uint8x16_t __ret;
24429  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
24430  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24431  return __ret;
24432 }
24433 #endif
24434 
24435 #ifdef __LITTLE_ENDIAN__
24436 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
24437  uint32x4_t __ret;
24438  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
24439  return __ret;
24440 }
24441 #else
24442 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
24443  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24444  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24445  uint32x4_t __ret;
24446  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
24447  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24448  return __ret;
24449 }
24450 #endif
24451 
24452 #ifdef __LITTLE_ENDIAN__
24453 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
24454  uint64x2_t __ret;
24455  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
24456  return __ret;
24457 }
24458 #else
24459 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
24460  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24461  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24462  uint64x2_t __ret;
24463  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
24464  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24465  return __ret;
24466 }
24467 #endif
24468 
24469 #ifdef __LITTLE_ENDIAN__
24470 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
24471  uint16x8_t __ret;
24472  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
24473  return __ret;
24474 }
24475 #else
24476 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
24477  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24478  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24479  uint16x8_t __ret;
24480  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
24481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24482  return __ret;
24483 }
24484 #endif
24485 
24486 #ifdef __LITTLE_ENDIAN__
24487 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
24488  int8x16_t __ret;
24489  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
24490  return __ret;
24491 }
24492 #else
24493 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
24494  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24495  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24496  int8x16_t __ret;
24497  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
24498  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
24499  return __ret;
24500 }
24501 #endif
24502 
24503 #ifdef __LITTLE_ENDIAN__
24504 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
24505  int32x4_t __ret;
24506  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
24507  return __ret;
24508 }
24509 #else
24510 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
24511  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24512  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24513  int32x4_t __ret;
24514  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
24515  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24516  return __ret;
24517 }
24518 #endif
24519 
24520 #ifdef __LITTLE_ENDIAN__
24521 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
24522  int64x2_t __ret;
24523  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
24524  return __ret;
24525 }
24526 #else
24527 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
24528  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24529  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24530  int64x2_t __ret;
24531  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
24532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24533  return __ret;
24534 }
24535 #endif
24536 
24537 #ifdef __LITTLE_ENDIAN__
24538 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
24539  int16x8_t __ret;
24540  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
24541  return __ret;
24542 }
24543 #else
24544 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
24545  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24546  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24547  int16x8_t __ret;
24548  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
24549  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24550  return __ret;
24551 }
24552 #endif
24553 
24554 #ifdef __LITTLE_ENDIAN__
24555 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
24556  uint8x8_t __ret;
24557  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
24558  return __ret;
24559 }
24560 #else
24561 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
24562  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24563  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24564  uint8x8_t __ret;
24565  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
24566  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24567  return __ret;
24568 }
24569 #endif
24570 
24571 #ifdef __LITTLE_ENDIAN__
24572 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
24573  uint32x2_t __ret;
24574  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
24575  return __ret;
24576 }
24577 #else
24578 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
24579  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24580  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24581  uint32x2_t __ret;
24582  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
24583  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24584  return __ret;
24585 }
24586 #endif
24587 
24588 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
24589  uint64x1_t __ret;
24590  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
24591  return __ret;
24592 }
24593 #ifdef __LITTLE_ENDIAN__
24594 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
24595  uint16x4_t __ret;
24596  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
24597  return __ret;
24598 }
24599 #else
24600 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
24601  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24602  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24603  uint16x4_t __ret;
24604  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
24605  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24606  return __ret;
24607 }
24608 #endif
24609 
24610 #ifdef __LITTLE_ENDIAN__
24611 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
24612  int8x8_t __ret;
24613  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
24614  return __ret;
24615 }
24616 #else
24617 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
24618  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
24619  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
24620  int8x8_t __ret;
24621  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
24622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
24623  return __ret;
24624 }
24625 #endif
24626 
24627 #ifdef __LITTLE_ENDIAN__
24628 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
24629  int32x2_t __ret;
24630  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
24631  return __ret;
24632 }
24633 #else
24634 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
24635  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
24636  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
24637  int32x2_t __ret;
24638  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
24639  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
24640  return __ret;
24641 }
24642 #endif
24643 
24644 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
24645  int64x1_t __ret;
24646  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
24647  return __ret;
24648 }
24649 #ifdef __LITTLE_ENDIAN__
24650 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
24651  int16x4_t __ret;
24652  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
24653  return __ret;
24654 }
24655 #else
24656 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
24657  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
24658  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
24659  int16x4_t __ret;
24660  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
24661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
24662  return __ret;
24663 }
24664 #endif
24665 
24666 #ifdef __LITTLE_ENDIAN__
24667 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
24668  uint8x16_t __s0 = __p0; \
24669  uint8x16_t __ret; \
24670  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
24671  __ret; \
24672 })
24673 #else
24674 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
24675  uint8x16_t __s0 = __p0; \
24676  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24677  uint8x16_t __ret; \
24678  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
24679  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24680  __ret; \
24681 })
24682 #endif
24683 
24684 #ifdef __LITTLE_ENDIAN__
24685 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
24686  uint32x4_t __s0 = __p0; \
24687  uint32x4_t __ret; \
24688  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
24689  __ret; \
24690 })
24691 #else
24692 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
24693  uint32x4_t __s0 = __p0; \
24694  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24695  uint32x4_t __ret; \
24696  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
24697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24698  __ret; \
24699 })
24700 #endif
24701 
24702 #ifdef __LITTLE_ENDIAN__
24703 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
24704  uint64x2_t __s0 = __p0; \
24705  uint64x2_t __ret; \
24706  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
24707  __ret; \
24708 })
24709 #else
24710 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
24711  uint64x2_t __s0 = __p0; \
24712  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24713  uint64x2_t __ret; \
24714  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
24715  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24716  __ret; \
24717 })
24718 #endif
24719 
24720 #ifdef __LITTLE_ENDIAN__
24721 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
24722  uint16x8_t __s0 = __p0; \
24723  uint16x8_t __ret; \
24724  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
24725  __ret; \
24726 })
24727 #else
24728 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
24729  uint16x8_t __s0 = __p0; \
24730  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24731  uint16x8_t __ret; \
24732  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
24733  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24734  __ret; \
24735 })
24736 #endif
24737 
24738 #ifdef __LITTLE_ENDIAN__
24739 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
24740  int8x16_t __s0 = __p0; \
24741  int8x16_t __ret; \
24742  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
24743  __ret; \
24744 })
24745 #else
24746 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
24747  int8x16_t __s0 = __p0; \
24748  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24749  int8x16_t __ret; \
24750  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
24751  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24752  __ret; \
24753 })
24754 #endif
24755 
24756 #ifdef __LITTLE_ENDIAN__
24757 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
24758  int32x4_t __s0 = __p0; \
24759  int32x4_t __ret; \
24760  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
24761  __ret; \
24762 })
24763 #else
24764 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
24765  int32x4_t __s0 = __p0; \
24766  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24767  int32x4_t __ret; \
24768  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
24769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24770  __ret; \
24771 })
24772 #endif
24773 
24774 #ifdef __LITTLE_ENDIAN__
24775 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
24776  int64x2_t __s0 = __p0; \
24777  int64x2_t __ret; \
24778  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
24779  __ret; \
24780 })
24781 #else
24782 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
24783  int64x2_t __s0 = __p0; \
24784  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24785  int64x2_t __ret; \
24786  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
24787  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24788  __ret; \
24789 })
24790 #endif
24791 
24792 #ifdef __LITTLE_ENDIAN__
24793 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
24794  int16x8_t __s0 = __p0; \
24795  int16x8_t __ret; \
24796  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
24797  __ret; \
24798 })
24799 #else
24800 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
24801  int16x8_t __s0 = __p0; \
24802  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24803  int16x8_t __ret; \
24804  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
24805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24806  __ret; \
24807 })
24808 #endif
24809 
24810 #ifdef __LITTLE_ENDIAN__
24811 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
24812  uint8x8_t __s0 = __p0; \
24813  uint8x8_t __ret; \
24814  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
24815  __ret; \
24816 })
24817 #else
24818 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
24819  uint8x8_t __s0 = __p0; \
24820  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24821  uint8x8_t __ret; \
24822  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
24823  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24824  __ret; \
24825 })
24826 #endif
24827 
24828 #ifdef __LITTLE_ENDIAN__
24829 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
24830  uint32x2_t __s0 = __p0; \
24831  uint32x2_t __ret; \
24832  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
24833  __ret; \
24834 })
24835 #else
24836 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
24837  uint32x2_t __s0 = __p0; \
24838  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24839  uint32x2_t __ret; \
24840  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
24841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24842  __ret; \
24843 })
24844 #endif
24845 
24846 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
24847  uint64x1_t __s0 = __p0; \
24848  uint64x1_t __ret; \
24849  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
24850  __ret; \
24851 })
24852 #ifdef __LITTLE_ENDIAN__
24853 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
24854  uint16x4_t __s0 = __p0; \
24855  uint16x4_t __ret; \
24856  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
24857  __ret; \
24858 })
24859 #else
24860 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
24861  uint16x4_t __s0 = __p0; \
24862  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24863  uint16x4_t __ret; \
24864  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
24865  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24866  __ret; \
24867 })
24868 #endif
24869 
24870 #ifdef __LITTLE_ENDIAN__
24871 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
24872  int8x8_t __s0 = __p0; \
24873  int8x8_t __ret; \
24874  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
24875  __ret; \
24876 })
24877 #else
24878 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
24879  int8x8_t __s0 = __p0; \
24880  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24881  int8x8_t __ret; \
24882  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
24883  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24884  __ret; \
24885 })
24886 #endif
24887 
24888 #ifdef __LITTLE_ENDIAN__
24889 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
24890  int32x2_t __s0 = __p0; \
24891  int32x2_t __ret; \
24892  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
24893  __ret; \
24894 })
24895 #else
24896 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
24897  int32x2_t __s0 = __p0; \
24898  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24899  int32x2_t __ret; \
24900  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
24901  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24902  __ret; \
24903 })
24904 #endif
24905 
24906 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
24907  int64x1_t __s0 = __p0; \
24908  int64x1_t __ret; \
24909  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
24910  __ret; \
24911 })
24912 #ifdef __LITTLE_ENDIAN__
24913 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
24914  int16x4_t __s0 = __p0; \
24915  int16x4_t __ret; \
24916  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
24917  __ret; \
24918 })
24919 #else
24920 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
24921  int16x4_t __s0 = __p0; \
24922  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24923  int16x4_t __ret; \
24924  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
24925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24926  __ret; \
24927 })
24928 #endif
24929 
24930 #ifdef __LITTLE_ENDIAN__
24931 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
24932  uint8x8_t __s0 = __p0; \
24933  uint16x8_t __ret; \
24934  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
24935  __ret; \
24936 })
24937 #else
24938 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
24939  uint8x8_t __s0 = __p0; \
24940  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24941  uint16x8_t __ret; \
24942  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
24943  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24944  __ret; \
24945 })
24946 #define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
24947  uint8x8_t __s0 = __p0; \
24948  uint16x8_t __ret; \
24949  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
24950  __ret; \
24951 })
24952 #endif
24953 
24954 #ifdef __LITTLE_ENDIAN__
24955 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
24956  uint32x2_t __s0 = __p0; \
24957  uint64x2_t __ret; \
24958  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
24959  __ret; \
24960 })
24961 #else
24962 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
24963  uint32x2_t __s0 = __p0; \
24964  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24965  uint64x2_t __ret; \
24966  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
24967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24968  __ret; \
24969 })
24970 #define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
24971  uint32x2_t __s0 = __p0; \
24972  uint64x2_t __ret; \
24973  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
24974  __ret; \
24975 })
24976 #endif
24977 
24978 #ifdef __LITTLE_ENDIAN__
24979 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
24980  uint16x4_t __s0 = __p0; \
24981  uint32x4_t __ret; \
24982  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
24983  __ret; \
24984 })
24985 #else
24986 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
24987  uint16x4_t __s0 = __p0; \
24988  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24989  uint32x4_t __ret; \
24990  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
24991  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24992  __ret; \
24993 })
24994 #define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
24995  uint16x4_t __s0 = __p0; \
24996  uint32x4_t __ret; \
24997  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
24998  __ret; \
24999 })
25000 #endif
25001 
25002 #ifdef __LITTLE_ENDIAN__
25003 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
25004  int8x8_t __s0 = __p0; \
25005  int16x8_t __ret; \
25006  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
25007  __ret; \
25008 })
25009 #else
25010 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
25011  int8x8_t __s0 = __p0; \
25012  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25013  int16x8_t __ret; \
25014  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
25015  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25016  __ret; \
25017 })
25018 #define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
25019  int8x8_t __s0 = __p0; \
25020  int16x8_t __ret; \
25021  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
25022  __ret; \
25023 })
25024 #endif
25025 
25026 #ifdef __LITTLE_ENDIAN__
25027 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
25028  int32x2_t __s0 = __p0; \
25029  int64x2_t __ret; \
25030  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
25031  __ret; \
25032 })
25033 #else
25034 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
25035  int32x2_t __s0 = __p0; \
25036  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25037  int64x2_t __ret; \
25038  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
25039  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25040  __ret; \
25041 })
25042 #define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
25043  int32x2_t __s0 = __p0; \
25044  int64x2_t __ret; \
25045  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
25046  __ret; \
25047 })
25048 #endif
25049 
25050 #ifdef __LITTLE_ENDIAN__
25051 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
25052  int16x4_t __s0 = __p0; \
25053  int32x4_t __ret; \
25054  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
25055  __ret; \
25056 })
25057 #else
25058 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
25059  int16x4_t __s0 = __p0; \
25060  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25061  int32x4_t __ret; \
25062  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
25063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25064  __ret; \
25065 })
25066 #define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
25067  int16x4_t __s0 = __p0; \
25068  int32x4_t __ret; \
25069  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
25070  __ret; \
25071 })
25072 #endif
25073 
25074 #ifdef __LITTLE_ENDIAN__
25075 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
25076  uint8x16_t __s0 = __p0; \
25077  uint8x16_t __ret; \
25078  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
25079  __ret; \
25080 })
25081 #else
25082 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
25083  uint8x16_t __s0 = __p0; \
25084  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25085  uint8x16_t __ret; \
25086  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
25087  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25088  __ret; \
25089 })
25090 #endif
25091 
25092 #ifdef __LITTLE_ENDIAN__
25093 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
25094  uint32x4_t __s0 = __p0; \
25095  uint32x4_t __ret; \
25096  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
25097  __ret; \
25098 })
25099 #else
25100 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
25101  uint32x4_t __s0 = __p0; \
25102  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25103  uint32x4_t __ret; \
25104  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
25105  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25106  __ret; \
25107 })
25108 #endif
25109 
25110 #ifdef __LITTLE_ENDIAN__
25111 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
25112  uint64x2_t __s0 = __p0; \
25113  uint64x2_t __ret; \
25114  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
25115  __ret; \
25116 })
25117 #else
25118 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
25119  uint64x2_t __s0 = __p0; \
25120  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25121  uint64x2_t __ret; \
25122  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
25123  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25124  __ret; \
25125 })
25126 #endif
25127 
25128 #ifdef __LITTLE_ENDIAN__
25129 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
25130  uint16x8_t __s0 = __p0; \
25131  uint16x8_t __ret; \
25132  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
25133  __ret; \
25134 })
25135 #else
25136 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
25137  uint16x8_t __s0 = __p0; \
25138  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25139  uint16x8_t __ret; \
25140  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
25141  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25142  __ret; \
25143 })
25144 #endif
25145 
25146 #ifdef __LITTLE_ENDIAN__
25147 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
25148  int8x16_t __s0 = __p0; \
25149  int8x16_t __ret; \
25150  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
25151  __ret; \
25152 })
25153 #else
25154 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
25155  int8x16_t __s0 = __p0; \
25156  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25157  int8x16_t __ret; \
25158  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
25159  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25160  __ret; \
25161 })
25162 #endif
25163 
25164 #ifdef __LITTLE_ENDIAN__
25165 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
25166  int32x4_t __s0 = __p0; \
25167  int32x4_t __ret; \
25168  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
25169  __ret; \
25170 })
25171 #else
25172 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
25173  int32x4_t __s0 = __p0; \
25174  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25175  int32x4_t __ret; \
25176  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
25177  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25178  __ret; \
25179 })
25180 #endif
25181 
25182 #ifdef __LITTLE_ENDIAN__
25183 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
25184  int64x2_t __s0 = __p0; \
25185  int64x2_t __ret; \
25186  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
25187  __ret; \
25188 })
25189 #else
25190 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
25191  int64x2_t __s0 = __p0; \
25192  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25193  int64x2_t __ret; \
25194  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
25195  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25196  __ret; \
25197 })
25198 #endif
25199 
25200 #ifdef __LITTLE_ENDIAN__
25201 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
25202  int16x8_t __s0 = __p0; \
25203  int16x8_t __ret; \
25204  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
25205  __ret; \
25206 })
25207 #else
25208 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
25209  int16x8_t __s0 = __p0; \
25210  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25211  int16x8_t __ret; \
25212  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
25213  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25214  __ret; \
25215 })
25216 #endif
25217 
25218 #ifdef __LITTLE_ENDIAN__
25219 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
25220  uint8x8_t __s0 = __p0; \
25221  uint8x8_t __ret; \
25222  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
25223  __ret; \
25224 })
25225 #else
25226 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
25227  uint8x8_t __s0 = __p0; \
25228  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25229  uint8x8_t __ret; \
25230  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
25231  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25232  __ret; \
25233 })
25234 #endif
25235 
25236 #ifdef __LITTLE_ENDIAN__
25237 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
25238  uint32x2_t __s0 = __p0; \
25239  uint32x2_t __ret; \
25240  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
25241  __ret; \
25242 })
25243 #else
25244 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
25245  uint32x2_t __s0 = __p0; \
25246  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25247  uint32x2_t __ret; \
25248  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
25249  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25250  __ret; \
25251 })
25252 #endif
25253 
25254 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
25255  uint64x1_t __s0 = __p0; \
25256  uint64x1_t __ret; \
25257  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
25258  __ret; \
25259 })
25260 #ifdef __LITTLE_ENDIAN__
25261 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
25262  uint16x4_t __s0 = __p0; \
25263  uint16x4_t __ret; \
25264  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
25265  __ret; \
25266 })
25267 #else
25268 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
25269  uint16x4_t __s0 = __p0; \
25270  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25271  uint16x4_t __ret; \
25272  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
25273  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25274  __ret; \
25275 })
25276 #endif
25277 
25278 #ifdef __LITTLE_ENDIAN__
25279 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
25280  int8x8_t __s0 = __p0; \
25281  int8x8_t __ret; \
25282  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
25283  __ret; \
25284 })
25285 #else
25286 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
25287  int8x8_t __s0 = __p0; \
25288  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25289  int8x8_t __ret; \
25290  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
25291  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25292  __ret; \
25293 })
25294 #endif
25295 
25296 #ifdef __LITTLE_ENDIAN__
25297 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
25298  int32x2_t __s0 = __p0; \
25299  int32x2_t __ret; \
25300  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
25301  __ret; \
25302 })
25303 #else
25304 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
25305  int32x2_t __s0 = __p0; \
25306  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25307  int32x2_t __ret; \
25308  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
25309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25310  __ret; \
25311 })
25312 #endif
25313 
25314 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
25315  int64x1_t __s0 = __p0; \
25316  int64x1_t __ret; \
25317  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
25318  __ret; \
25319 })
25320 #ifdef __LITTLE_ENDIAN__
25321 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
25322  int16x4_t __s0 = __p0; \
25323  int16x4_t __ret; \
25324  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
25325  __ret; \
25326 })
25327 #else
25328 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
25329  int16x4_t __s0 = __p0; \
25330  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25331  int16x4_t __ret; \
25332  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
25333  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25334  __ret; \
25335 })
25336 #endif
25337 
25338 #ifdef __LITTLE_ENDIAN__
25339 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
25340  uint32x4_t __s0 = __p0; \
25341  uint16x4_t __ret; \
25342  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
25343  __ret; \
25344 })
25345 #else
25346 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
25347  uint32x4_t __s0 = __p0; \
25348  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25349  uint16x4_t __ret; \
25350  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
25351  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25352  __ret; \
25353 })
25354 #define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
25355  uint32x4_t __s0 = __p0; \
25356  uint16x4_t __ret; \
25357  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
25358  __ret; \
25359 })
25360 #endif
25361 
25362 #ifdef __LITTLE_ENDIAN__
25363 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
25364  uint64x2_t __s0 = __p0; \
25365  uint32x2_t __ret; \
25366  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
25367  __ret; \
25368 })
25369 #else
25370 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
25371  uint64x2_t __s0 = __p0; \
25372  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25373  uint32x2_t __ret; \
25374  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
25375  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25376  __ret; \
25377 })
25378 #define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
25379  uint64x2_t __s0 = __p0; \
25380  uint32x2_t __ret; \
25381  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
25382  __ret; \
25383 })
25384 #endif
25385 
25386 #ifdef __LITTLE_ENDIAN__
25387 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
25388  uint16x8_t __s0 = __p0; \
25389  uint8x8_t __ret; \
25390  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
25391  __ret; \
25392 })
25393 #else
25394 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
25395  uint16x8_t __s0 = __p0; \
25396  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25397  uint8x8_t __ret; \
25398  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
25399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25400  __ret; \
25401 })
25402 #define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
25403  uint16x8_t __s0 = __p0; \
25404  uint8x8_t __ret; \
25405  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
25406  __ret; \
25407 })
25408 #endif
25409 
25410 #ifdef __LITTLE_ENDIAN__
25411 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
25412  int32x4_t __s0 = __p0; \
25413  int16x4_t __ret; \
25414  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
25415  __ret; \
25416 })
25417 #else
25418 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
25419  int32x4_t __s0 = __p0; \
25420  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25421  int16x4_t __ret; \
25422  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
25423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25424  __ret; \
25425 })
25426 #define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
25427  int32x4_t __s0 = __p0; \
25428  int16x4_t __ret; \
25429  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
25430  __ret; \
25431 })
25432 #endif
25433 
25434 #ifdef __LITTLE_ENDIAN__
25435 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
25436  int64x2_t __s0 = __p0; \
25437  int32x2_t __ret; \
25438  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
25439  __ret; \
25440 })
25441 #else
25442 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
25443  int64x2_t __s0 = __p0; \
25444  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25445  int32x2_t __ret; \
25446  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
25447  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25448  __ret; \
25449 })
25450 #define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
25451  int64x2_t __s0 = __p0; \
25452  int32x2_t __ret; \
25453  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
25454  __ret; \
25455 })
25456 #endif
25457 
25458 #ifdef __LITTLE_ENDIAN__
25459 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
25460  int16x8_t __s0 = __p0; \
25461  int8x8_t __ret; \
25462  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
25463  __ret; \
25464 })
25465 #else
25466 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
25467  int16x8_t __s0 = __p0; \
25468  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25469  int8x8_t __ret; \
25470  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
25471  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25472  __ret; \
25473 })
25474 #define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
25475  int16x8_t __s0 = __p0; \
25476  int8x8_t __ret; \
25477  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
25478  __ret; \
25479 })
25480 #endif
25481 
25482 #ifdef __LITTLE_ENDIAN__
25483 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
25484  poly8x8_t __s0 = __p0; \
25485  poly8x8_t __s1 = __p1; \
25486  poly8x8_t __ret; \
25487  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
25488  __ret; \
25489 })
25490 #else
25491 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
25492  poly8x8_t __s0 = __p0; \
25493  poly8x8_t __s1 = __p1; \
25494  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25495  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25496  poly8x8_t __ret; \
25497  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
25498  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25499  __ret; \
25500 })
25501 #endif
25502 
25503 #ifdef __LITTLE_ENDIAN__
25504 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
25505  poly16x4_t __s0 = __p0; \
25506  poly16x4_t __s1 = __p1; \
25507  poly16x4_t __ret; \
25508  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
25509  __ret; \
25510 })
25511 #else
25512 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
25513  poly16x4_t __s0 = __p0; \
25514  poly16x4_t __s1 = __p1; \
25515  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25516  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25517  poly16x4_t __ret; \
25518  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
25519  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25520  __ret; \
25521 })
25522 #endif
25523 
25524 #ifdef __LITTLE_ENDIAN__
25525 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
25526  poly8x16_t __s0 = __p0; \
25527  poly8x16_t __s1 = __p1; \
25528  poly8x16_t __ret; \
25529  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
25530  __ret; \
25531 })
25532 #else
25533 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
25534  poly8x16_t __s0 = __p0; \
25535  poly8x16_t __s1 = __p1; \
25536  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25537  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25538  poly8x16_t __ret; \
25539  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
25540  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25541  __ret; \
25542 })
25543 #endif
25544 
25545 #ifdef __LITTLE_ENDIAN__
25546 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
25547  poly16x8_t __s0 = __p0; \
25548  poly16x8_t __s1 = __p1; \
25549  poly16x8_t __ret; \
25550  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
25551  __ret; \
25552 })
25553 #else
25554 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
25555  poly16x8_t __s0 = __p0; \
25556  poly16x8_t __s1 = __p1; \
25557  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25558  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25559  poly16x8_t __ret; \
25560  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
25561  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25562  __ret; \
25563 })
25564 #endif
25565 
25566 #ifdef __LITTLE_ENDIAN__
25567 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25568  uint8x16_t __s0 = __p0; \
25569  uint8x16_t __s1 = __p1; \
25570  uint8x16_t __ret; \
25571  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
25572  __ret; \
25573 })
25574 #else
25575 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25576  uint8x16_t __s0 = __p0; \
25577  uint8x16_t __s1 = __p1; \
25578  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25579  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25580  uint8x16_t __ret; \
25581  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
25582  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25583  __ret; \
25584 })
25585 #endif
25586 
25587 #ifdef __LITTLE_ENDIAN__
25588 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25589  uint32x4_t __s0 = __p0; \
25590  uint32x4_t __s1 = __p1; \
25591  uint32x4_t __ret; \
25592  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
25593  __ret; \
25594 })
25595 #else
25596 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25597  uint32x4_t __s0 = __p0; \
25598  uint32x4_t __s1 = __p1; \
25599  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25600  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25601  uint32x4_t __ret; \
25602  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
25603  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25604  __ret; \
25605 })
25606 #endif
25607 
25608 #ifdef __LITTLE_ENDIAN__
25609 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25610  uint64x2_t __s0 = __p0; \
25611  uint64x2_t __s1 = __p1; \
25612  uint64x2_t __ret; \
25613  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
25614  __ret; \
25615 })
25616 #else
25617 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25618  uint64x2_t __s0 = __p0; \
25619  uint64x2_t __s1 = __p1; \
25620  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25621  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25622  uint64x2_t __ret; \
25623  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
25624  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25625  __ret; \
25626 })
25627 #endif
25628 
25629 #ifdef __LITTLE_ENDIAN__
25630 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25631  uint16x8_t __s0 = __p0; \
25632  uint16x8_t __s1 = __p1; \
25633  uint16x8_t __ret; \
25634  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
25635  __ret; \
25636 })
25637 #else
25638 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25639  uint16x8_t __s0 = __p0; \
25640  uint16x8_t __s1 = __p1; \
25641  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25642  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25643  uint16x8_t __ret; \
25644  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
25645  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25646  __ret; \
25647 })
25648 #endif
25649 
25650 #ifdef __LITTLE_ENDIAN__
25651 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25652  int8x16_t __s0 = __p0; \
25653  int8x16_t __s1 = __p1; \
25654  int8x16_t __ret; \
25655  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
25656  __ret; \
25657 })
25658 #else
25659 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25660  int8x16_t __s0 = __p0; \
25661  int8x16_t __s1 = __p1; \
25662  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25663  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25664  int8x16_t __ret; \
25665  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
25666  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25667  __ret; \
25668 })
25669 #endif
25670 
25671 #ifdef __LITTLE_ENDIAN__
25672 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25673  int32x4_t __s0 = __p0; \
25674  int32x4_t __s1 = __p1; \
25675  int32x4_t __ret; \
25676  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
25677  __ret; \
25678 })
25679 #else
25680 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25681  int32x4_t __s0 = __p0; \
25682  int32x4_t __s1 = __p1; \
25683  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25684  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25685  int32x4_t __ret; \
25686  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
25687  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25688  __ret; \
25689 })
25690 #endif
25691 
25692 #ifdef __LITTLE_ENDIAN__
25693 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25694  int64x2_t __s0 = __p0; \
25695  int64x2_t __s1 = __p1; \
25696  int64x2_t __ret; \
25697  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
25698  __ret; \
25699 })
25700 #else
25701 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
25702  int64x2_t __s0 = __p0; \
25703  int64x2_t __s1 = __p1; \
25704  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25705  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25706  int64x2_t __ret; \
25707  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
25708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25709  __ret; \
25710 })
25711 #endif
25712 
25713 #ifdef __LITTLE_ENDIAN__
25714 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25715  int16x8_t __s0 = __p0; \
25716  int16x8_t __s1 = __p1; \
25717  int16x8_t __ret; \
25718  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
25719  __ret; \
25720 })
25721 #else
25722 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
25723  int16x8_t __s0 = __p0; \
25724  int16x8_t __s1 = __p1; \
25725  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25726  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25727  int16x8_t __ret; \
25728  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
25729  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25730  __ret; \
25731 })
25732 #endif
25733 
25734 #ifdef __LITTLE_ENDIAN__
25735 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
25736  uint8x8_t __s0 = __p0; \
25737  uint8x8_t __s1 = __p1; \
25738  uint8x8_t __ret; \
25739  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
25740  __ret; \
25741 })
25742 #else
25743 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
25744  uint8x8_t __s0 = __p0; \
25745  uint8x8_t __s1 = __p1; \
25746  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25747  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25748  uint8x8_t __ret; \
25749  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
25750  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25751  __ret; \
25752 })
25753 #endif
25754 
25755 #ifdef __LITTLE_ENDIAN__
25756 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
25757  uint32x2_t __s0 = __p0; \
25758  uint32x2_t __s1 = __p1; \
25759  uint32x2_t __ret; \
25760  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
25761  __ret; \
25762 })
25763 #else
25764 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
25765  uint32x2_t __s0 = __p0; \
25766  uint32x2_t __s1 = __p1; \
25767  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25768  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25769  uint32x2_t __ret; \
25770  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
25771  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25772  __ret; \
25773 })
25774 #endif
25775 
25776 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
25777  uint64x1_t __s0 = __p0; \
25778  uint64x1_t __s1 = __p1; \
25779  uint64x1_t __ret; \
25780  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
25781  __ret; \
25782 })
25783 #ifdef __LITTLE_ENDIAN__
25784 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
25785  uint16x4_t __s0 = __p0; \
25786  uint16x4_t __s1 = __p1; \
25787  uint16x4_t __ret; \
25788  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
25789  __ret; \
25790 })
25791 #else
25792 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
25793  uint16x4_t __s0 = __p0; \
25794  uint16x4_t __s1 = __p1; \
25795  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25796  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25797  uint16x4_t __ret; \
25798  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
25799  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25800  __ret; \
25801 })
25802 #endif
25803 
25804 #ifdef __LITTLE_ENDIAN__
25805 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
25806  int8x8_t __s0 = __p0; \
25807  int8x8_t __s1 = __p1; \
25808  int8x8_t __ret; \
25809  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
25810  __ret; \
25811 })
25812 #else
25813 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
25814  int8x8_t __s0 = __p0; \
25815  int8x8_t __s1 = __p1; \
25816  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25817  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25818  int8x8_t __ret; \
25819  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
25820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25821  __ret; \
25822 })
25823 #endif
25824 
25825 #ifdef __LITTLE_ENDIAN__
25826 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
25827  int32x2_t __s0 = __p0; \
25828  int32x2_t __s1 = __p1; \
25829  int32x2_t __ret; \
25830  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
25831  __ret; \
25832 })
25833 #else
25834 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
25835  int32x2_t __s0 = __p0; \
25836  int32x2_t __s1 = __p1; \
25837  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25838  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25839  int32x2_t __ret; \
25840  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
25841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25842  __ret; \
25843 })
25844 #endif
25845 
25846 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
25847  int64x1_t __s0 = __p0; \
25848  int64x1_t __s1 = __p1; \
25849  int64x1_t __ret; \
25850  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25851  __ret; \
25852 })
25853 #ifdef __LITTLE_ENDIAN__
25854 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
25855  int16x4_t __s0 = __p0; \
25856  int16x4_t __s1 = __p1; \
25857  int16x4_t __ret; \
25858  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
25859  __ret; \
25860 })
25861 #else
25862 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
25863  int16x4_t __s0 = __p0; \
25864  int16x4_t __s1 = __p1; \
25865  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25866  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25867  int16x4_t __ret; \
25868  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
25869  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25870  __ret; \
25871 })
25872 #endif
25873 
25874 #ifdef __LITTLE_ENDIAN__
25875 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25876  uint8x16_t __s0 = __p0; \
25877  uint8x16_t __s1 = __p1; \
25878  uint8x16_t __ret; \
25879  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
25880  __ret; \
25881 })
25882 #else
25883 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
25884  uint8x16_t __s0 = __p0; \
25885  uint8x16_t __s1 = __p1; \
25886  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25887  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25888  uint8x16_t __ret; \
25889  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
25890  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25891  __ret; \
25892 })
25893 #endif
25894 
25895 #ifdef __LITTLE_ENDIAN__
25896 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25897  uint32x4_t __s0 = __p0; \
25898  uint32x4_t __s1 = __p1; \
25899  uint32x4_t __ret; \
25900  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
25901  __ret; \
25902 })
25903 #else
25904 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
25905  uint32x4_t __s0 = __p0; \
25906  uint32x4_t __s1 = __p1; \
25907  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25908  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25909  uint32x4_t __ret; \
25910  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
25911  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25912  __ret; \
25913 })
25914 #endif
25915 
25916 #ifdef __LITTLE_ENDIAN__
25917 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25918  uint64x2_t __s0 = __p0; \
25919  uint64x2_t __s1 = __p1; \
25920  uint64x2_t __ret; \
25921  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
25922  __ret; \
25923 })
25924 #else
25925 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
25926  uint64x2_t __s0 = __p0; \
25927  uint64x2_t __s1 = __p1; \
25928  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25929  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25930  uint64x2_t __ret; \
25931  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
25932  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25933  __ret; \
25934 })
25935 #endif
25936 
25937 #ifdef __LITTLE_ENDIAN__
25938 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25939  uint16x8_t __s0 = __p0; \
25940  uint16x8_t __s1 = __p1; \
25941  uint16x8_t __ret; \
25942  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
25943  __ret; \
25944 })
25945 #else
25946 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
25947  uint16x8_t __s0 = __p0; \
25948  uint16x8_t __s1 = __p1; \
25949  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
25950  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25951  uint16x8_t __ret; \
25952  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
25953  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
25954  __ret; \
25955 })
25956 #endif
25957 
25958 #ifdef __LITTLE_ENDIAN__
25959 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25960  int8x16_t __s0 = __p0; \
25961  int8x16_t __s1 = __p1; \
25962  int8x16_t __ret; \
25963  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
25964  __ret; \
25965 })
25966 #else
25967 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
25968  int8x16_t __s0 = __p0; \
25969  int8x16_t __s1 = __p1; \
25970  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25971  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25972  int8x16_t __ret; \
25973  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
25974  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25975  __ret; \
25976 })
25977 #endif
25978 
25979 #ifdef __LITTLE_ENDIAN__
25980 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25981  int32x4_t __s0 = __p0; \
25982  int32x4_t __s1 = __p1; \
25983  int32x4_t __ret; \
25984  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
25985  __ret; \
25986 })
25987 #else
25988 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
25989  int32x4_t __s0 = __p0; \
25990  int32x4_t __s1 = __p1; \
25991  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25992  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25993  int32x4_t __ret; \
25994  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
25995  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25996  __ret; \
25997 })
25998 #endif
25999 
26000 #ifdef __LITTLE_ENDIAN__
26001 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26002  int64x2_t __s0 = __p0; \
26003  int64x2_t __s1 = __p1; \
26004  int64x2_t __ret; \
26005  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
26006  __ret; \
26007 })
26008 #else
26009 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26010  int64x2_t __s0 = __p0; \
26011  int64x2_t __s1 = __p1; \
26012  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26013  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26014  int64x2_t __ret; \
26015  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
26016  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26017  __ret; \
26018 })
26019 #endif
26020 
26021 #ifdef __LITTLE_ENDIAN__
26022 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26023  int16x8_t __s0 = __p0; \
26024  int16x8_t __s1 = __p1; \
26025  int16x8_t __ret; \
26026  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
26027  __ret; \
26028 })
26029 #else
26030 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26031  int16x8_t __s0 = __p0; \
26032  int16x8_t __s1 = __p1; \
26033  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26034  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26035  int16x8_t __ret; \
26036  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
26037  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26038  __ret; \
26039 })
26040 #endif
26041 
26042 #ifdef __LITTLE_ENDIAN__
26043 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
26044  uint8x8_t __s0 = __p0; \
26045  uint8x8_t __s1 = __p1; \
26046  uint8x8_t __ret; \
26047  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
26048  __ret; \
26049 })
26050 #else
26051 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
26052  uint8x8_t __s0 = __p0; \
26053  uint8x8_t __s1 = __p1; \
26054  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26055  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26056  uint8x8_t __ret; \
26057  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
26058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26059  __ret; \
26060 })
26061 #endif
26062 
26063 #ifdef __LITTLE_ENDIAN__
26064 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
26065  uint32x2_t __s0 = __p0; \
26066  uint32x2_t __s1 = __p1; \
26067  uint32x2_t __ret; \
26068  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
26069  __ret; \
26070 })
26071 #else
26072 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
26073  uint32x2_t __s0 = __p0; \
26074  uint32x2_t __s1 = __p1; \
26075  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26076  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26077  uint32x2_t __ret; \
26078  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
26079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26080  __ret; \
26081 })
26082 #endif
26083 
26084 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
26085  uint64x1_t __s0 = __p0; \
26086  uint64x1_t __s1 = __p1; \
26087  uint64x1_t __ret; \
26088  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
26089  __ret; \
26090 })
26091 #ifdef __LITTLE_ENDIAN__
26092 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
26093  uint16x4_t __s0 = __p0; \
26094  uint16x4_t __s1 = __p1; \
26095  uint16x4_t __ret; \
26096  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
26097  __ret; \
26098 })
26099 #else
26100 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
26101  uint16x4_t __s0 = __p0; \
26102  uint16x4_t __s1 = __p1; \
26103  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26104  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26105  uint16x4_t __ret; \
26106  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
26107  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26108  __ret; \
26109 })
26110 #endif
26111 
26112 #ifdef __LITTLE_ENDIAN__
26113 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
26114  int8x8_t __s0 = __p0; \
26115  int8x8_t __s1 = __p1; \
26116  int8x8_t __ret; \
26117  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
26118  __ret; \
26119 })
26120 #else
26121 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
26122  int8x8_t __s0 = __p0; \
26123  int8x8_t __s1 = __p1; \
26124  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26125  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26126  int8x8_t __ret; \
26127  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
26128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26129  __ret; \
26130 })
26131 #endif
26132 
26133 #ifdef __LITTLE_ENDIAN__
26134 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
26135  int32x2_t __s0 = __p0; \
26136  int32x2_t __s1 = __p1; \
26137  int32x2_t __ret; \
26138  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
26139  __ret; \
26140 })
26141 #else
26142 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
26143  int32x2_t __s0 = __p0; \
26144  int32x2_t __s1 = __p1; \
26145  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26146  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26147  int32x2_t __ret; \
26148  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
26149  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26150  __ret; \
26151 })
26152 #endif
26153 
26154 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
26155  int64x1_t __s0 = __p0; \
26156  int64x1_t __s1 = __p1; \
26157  int64x1_t __ret; \
26158  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
26159  __ret; \
26160 })
26161 #ifdef __LITTLE_ENDIAN__
26162 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
26163  int16x4_t __s0 = __p0; \
26164  int16x4_t __s1 = __p1; \
26165  int16x4_t __ret; \
26166  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
26167  __ret; \
26168 })
26169 #else
26170 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
26171  int16x4_t __s0 = __p0; \
26172  int16x4_t __s1 = __p1; \
26173  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26174  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26175  int16x4_t __ret; \
26176  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
26177  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26178  __ret; \
26179 })
26180 #endif
26181 
26182 #ifdef __LITTLE_ENDIAN__
26183 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
26184  poly8x8_t __s0 = __p0; \
26185  poly8x8_t __s1 = __p1; \
26186  poly8x8_t __ret; \
26187  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
26188  __ret; \
26189 })
26190 #else
26191 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
26192  poly8x8_t __s0 = __p0; \
26193  poly8x8_t __s1 = __p1; \
26194  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26195  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26196  poly8x8_t __ret; \
26197  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
26198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26199  __ret; \
26200 })
26201 #endif
26202 
26203 #ifdef __LITTLE_ENDIAN__
26204 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
26205  poly16x4_t __s0 = __p0; \
26206  poly16x4_t __s1 = __p1; \
26207  poly16x4_t __ret; \
26208  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
26209  __ret; \
26210 })
26211 #else
26212 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
26213  poly16x4_t __s0 = __p0; \
26214  poly16x4_t __s1 = __p1; \
26215  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26216  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26217  poly16x4_t __ret; \
26218  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
26219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26220  __ret; \
26221 })
26222 #endif
26223 
26224 #ifdef __LITTLE_ENDIAN__
26225 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
26226  poly8x16_t __s0 = __p0; \
26227  poly8x16_t __s1 = __p1; \
26228  poly8x16_t __ret; \
26229  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
26230  __ret; \
26231 })
26232 #else
26233 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
26234  poly8x16_t __s0 = __p0; \
26235  poly8x16_t __s1 = __p1; \
26236  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26237  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26238  poly8x16_t __ret; \
26239  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
26240  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26241  __ret; \
26242 })
26243 #endif
26244 
26245 #ifdef __LITTLE_ENDIAN__
26246 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
26247  poly16x8_t __s0 = __p0; \
26248  poly16x8_t __s1 = __p1; \
26249  poly16x8_t __ret; \
26250  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
26251  __ret; \
26252 })
26253 #else
26254 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
26255  poly16x8_t __s0 = __p0; \
26256  poly16x8_t __s1 = __p1; \
26257  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26258  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26259  poly16x8_t __ret; \
26260  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
26261  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26262  __ret; \
26263 })
26264 #endif
26265 
26266 #ifdef __LITTLE_ENDIAN__
26267 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
26268  uint8x16_t __s0 = __p0; \
26269  uint8x16_t __s1 = __p1; \
26270  uint8x16_t __ret; \
26271  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
26272  __ret; \
26273 })
26274 #else
26275 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
26276  uint8x16_t __s0 = __p0; \
26277  uint8x16_t __s1 = __p1; \
26278  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26279  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26280  uint8x16_t __ret; \
26281  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
26282  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26283  __ret; \
26284 })
26285 #endif
26286 
26287 #ifdef __LITTLE_ENDIAN__
26288 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
26289  uint32x4_t __s0 = __p0; \
26290  uint32x4_t __s1 = __p1; \
26291  uint32x4_t __ret; \
26292  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
26293  __ret; \
26294 })
26295 #else
26296 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
26297  uint32x4_t __s0 = __p0; \
26298  uint32x4_t __s1 = __p1; \
26299  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26300  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26301  uint32x4_t __ret; \
26302  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
26303  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26304  __ret; \
26305 })
26306 #endif
26307 
26308 #ifdef __LITTLE_ENDIAN__
26309 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
26310  uint64x2_t __s0 = __p0; \
26311  uint64x2_t __s1 = __p1; \
26312  uint64x2_t __ret; \
26313  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
26314  __ret; \
26315 })
26316 #else
26317 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
26318  uint64x2_t __s0 = __p0; \
26319  uint64x2_t __s1 = __p1; \
26320  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26321  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26322  uint64x2_t __ret; \
26323  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
26324  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26325  __ret; \
26326 })
26327 #endif
26328 
26329 #ifdef __LITTLE_ENDIAN__
26330 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
26331  uint16x8_t __s0 = __p0; \
26332  uint16x8_t __s1 = __p1; \
26333  uint16x8_t __ret; \
26334  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
26335  __ret; \
26336 })
26337 #else
26338 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
26339  uint16x8_t __s0 = __p0; \
26340  uint16x8_t __s1 = __p1; \
26341  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26342  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26343  uint16x8_t __ret; \
26344  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
26345  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26346  __ret; \
26347 })
26348 #endif
26349 
26350 #ifdef __LITTLE_ENDIAN__
26351 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
26352  int8x16_t __s0 = __p0; \
26353  int8x16_t __s1 = __p1; \
26354  int8x16_t __ret; \
26355  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
26356  __ret; \
26357 })
26358 #else
26359 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
26360  int8x16_t __s0 = __p0; \
26361  int8x16_t __s1 = __p1; \
26362  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26363  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26364  int8x16_t __ret; \
26365  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
26366  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26367  __ret; \
26368 })
26369 #endif
26370 
26371 #ifdef __LITTLE_ENDIAN__
26372 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
26373  int32x4_t __s0 = __p0; \
26374  int32x4_t __s1 = __p1; \
26375  int32x4_t __ret; \
26376  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
26377  __ret; \
26378 })
26379 #else
26380 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
26381  int32x4_t __s0 = __p0; \
26382  int32x4_t __s1 = __p1; \
26383  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26384  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26385  int32x4_t __ret; \
26386  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
26387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26388  __ret; \
26389 })
26390 #endif
26391 
26392 #ifdef __LITTLE_ENDIAN__
26393 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26394  int64x2_t __s0 = __p0; \
26395  int64x2_t __s1 = __p1; \
26396  int64x2_t __ret; \
26397  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
26398  __ret; \
26399 })
26400 #else
26401 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
26402  int64x2_t __s0 = __p0; \
26403  int64x2_t __s1 = __p1; \
26404  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26405  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26406  int64x2_t __ret; \
26407  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
26408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26409  __ret; \
26410 })
26411 #endif
26412 
26413 #ifdef __LITTLE_ENDIAN__
26414 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26415  int16x8_t __s0 = __p0; \
26416  int16x8_t __s1 = __p1; \
26417  int16x8_t __ret; \
26418  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
26419  __ret; \
26420 })
26421 #else
26422 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
26423  int16x8_t __s0 = __p0; \
26424  int16x8_t __s1 = __p1; \
26425  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26426  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26427  int16x8_t __ret; \
26428  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
26429  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26430  __ret; \
26431 })
26432 #endif
26433 
26434 #ifdef __LITTLE_ENDIAN__
26435 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
26436  uint8x8_t __s0 = __p0; \
26437  uint8x8_t __s1 = __p1; \
26438  uint8x8_t __ret; \
26439  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
26440  __ret; \
26441 })
26442 #else
26443 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
26444  uint8x8_t __s0 = __p0; \
26445  uint8x8_t __s1 = __p1; \
26446  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26447  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26448  uint8x8_t __ret; \
26449  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
26450  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26451  __ret; \
26452 })
26453 #endif
26454 
26455 #ifdef __LITTLE_ENDIAN__
26456 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
26457  uint32x2_t __s0 = __p0; \
26458  uint32x2_t __s1 = __p1; \
26459  uint32x2_t __ret; \
26460  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
26461  __ret; \
26462 })
26463 #else
26464 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
26465  uint32x2_t __s0 = __p0; \
26466  uint32x2_t __s1 = __p1; \
26467  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26468  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26469  uint32x2_t __ret; \
26470  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
26471  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26472  __ret; \
26473 })
26474 #endif
26475 
26476 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
26477  uint64x1_t __s0 = __p0; \
26478  uint64x1_t __s1 = __p1; \
26479  uint64x1_t __ret; \
26480  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
26481  __ret; \
26482 })
26483 #ifdef __LITTLE_ENDIAN__
26484 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
26485  uint16x4_t __s0 = __p0; \
26486  uint16x4_t __s1 = __p1; \
26487  uint16x4_t __ret; \
26488  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
26489  __ret; \
26490 })
26491 #else
26492 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
26493  uint16x4_t __s0 = __p0; \
26494  uint16x4_t __s1 = __p1; \
26495  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26496  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26497  uint16x4_t __ret; \
26498  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
26499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26500  __ret; \
26501 })
26502 #endif
26503 
26504 #ifdef __LITTLE_ENDIAN__
26505 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
26506  int8x8_t __s0 = __p0; \
26507  int8x8_t __s1 = __p1; \
26508  int8x8_t __ret; \
26509  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
26510  __ret; \
26511 })
26512 #else
26513 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
26514  int8x8_t __s0 = __p0; \
26515  int8x8_t __s1 = __p1; \
26516  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
26517  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26518  int8x8_t __ret; \
26519  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
26520  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
26521  __ret; \
26522 })
26523 #endif
26524 
26525 #ifdef __LITTLE_ENDIAN__
26526 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
26527  int32x2_t __s0 = __p0; \
26528  int32x2_t __s1 = __p1; \
26529  int32x2_t __ret; \
26530  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
26531  __ret; \
26532 })
26533 #else
26534 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
26535  int32x2_t __s0 = __p0; \
26536  int32x2_t __s1 = __p1; \
26537  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
26538  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26539  int32x2_t __ret; \
26540  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
26541  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
26542  __ret; \
26543 })
26544 #endif
26545 
26546 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
26547  int64x1_t __s0 = __p0; \
26548  int64x1_t __s1 = __p1; \
26549  int64x1_t __ret; \
26550  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
26551  __ret; \
26552 })
26553 #ifdef __LITTLE_ENDIAN__
26554 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
26555  int16x4_t __s0 = __p0; \
26556  int16x4_t __s1 = __p1; \
26557  int16x4_t __ret; \
26558  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
26559  __ret; \
26560 })
26561 #else
26562 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
26563  int16x4_t __s0 = __p0; \
26564  int16x4_t __s1 = __p1; \
26565  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
26566  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26567  int16x4_t __ret; \
26568  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
26569  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
26570  __ret; \
26571 })
26572 #endif
26573 
26574 #ifdef __LITTLE_ENDIAN__
26575 #define vst1_p8(__p0, __p1) __extension__ ({ \
26576  poly8x8_t __s1 = __p1; \
26577  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
26578 })
26579 #else
26580 #define vst1_p8(__p0, __p1) __extension__ ({ \
26581  poly8x8_t __s1 = __p1; \
26582  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26583  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
26584 })
26585 #endif
26586 
26587 #ifdef __LITTLE_ENDIAN__
26588 #define vst1_p16(__p0, __p1) __extension__ ({ \
26589  poly16x4_t __s1 = __p1; \
26590  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
26591 })
26592 #else
26593 #define vst1_p16(__p0, __p1) __extension__ ({ \
26594  poly16x4_t __s1 = __p1; \
26595  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26596  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
26597 })
26598 #endif
26599 
26600 #ifdef __LITTLE_ENDIAN__
26601 #define vst1q_p8(__p0, __p1) __extension__ ({ \
26602  poly8x16_t __s1 = __p1; \
26603  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
26604 })
26605 #else
26606 #define vst1q_p8(__p0, __p1) __extension__ ({ \
26607  poly8x16_t __s1 = __p1; \
26608  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26609  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
26610 })
26611 #endif
26612 
26613 #ifdef __LITTLE_ENDIAN__
26614 #define vst1q_p16(__p0, __p1) __extension__ ({ \
26615  poly16x8_t __s1 = __p1; \
26616  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
26617 })
26618 #else
26619 #define vst1q_p16(__p0, __p1) __extension__ ({ \
26620  poly16x8_t __s1 = __p1; \
26621  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26622  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
26623 })
26624 #endif
26625 
26626 #ifdef __LITTLE_ENDIAN__
26627 #define vst1q_u8(__p0, __p1) __extension__ ({ \
26628  uint8x16_t __s1 = __p1; \
26629  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
26630 })
26631 #else
26632 #define vst1q_u8(__p0, __p1) __extension__ ({ \
26633  uint8x16_t __s1 = __p1; \
26634  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26635  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
26636 })
26637 #endif
26638 
26639 #ifdef __LITTLE_ENDIAN__
26640 #define vst1q_u32(__p0, __p1) __extension__ ({ \
26641  uint32x4_t __s1 = __p1; \
26642  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
26643 })
26644 #else
26645 #define vst1q_u32(__p0, __p1) __extension__ ({ \
26646  uint32x4_t __s1 = __p1; \
26647  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26648  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
26649 })
26650 #endif
26651 
26652 #ifdef __LITTLE_ENDIAN__
26653 #define vst1q_u64(__p0, __p1) __extension__ ({ \
26654  uint64x2_t __s1 = __p1; \
26655  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
26656 })
26657 #else
26658 #define vst1q_u64(__p0, __p1) __extension__ ({ \
26659  uint64x2_t __s1 = __p1; \
26660  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26661  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
26662 })
26663 #endif
26664 
26665 #ifdef __LITTLE_ENDIAN__
26666 #define vst1q_u16(__p0, __p1) __extension__ ({ \
26667  uint16x8_t __s1 = __p1; \
26668  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
26669 })
26670 #else
26671 #define vst1q_u16(__p0, __p1) __extension__ ({ \
26672  uint16x8_t __s1 = __p1; \
26673  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26674  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
26675 })
26676 #endif
26677 
26678 #ifdef __LITTLE_ENDIAN__
26679 #define vst1q_s8(__p0, __p1) __extension__ ({ \
26680  int8x16_t __s1 = __p1; \
26681  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
26682 })
26683 #else
26684 #define vst1q_s8(__p0, __p1) __extension__ ({ \
26685  int8x16_t __s1 = __p1; \
26686  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26687  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
26688 })
26689 #endif
26690 
26691 #ifdef __LITTLE_ENDIAN__
26692 #define vst1q_f32(__p0, __p1) __extension__ ({ \
26693  float32x4_t __s1 = __p1; \
26694  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
26695 })
26696 #else
26697 #define vst1q_f32(__p0, __p1) __extension__ ({ \
26698  float32x4_t __s1 = __p1; \
26699  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26700  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
26701 })
26702 #endif
26703 
26704 #ifdef __LITTLE_ENDIAN__
26705 #define vst1q_s32(__p0, __p1) __extension__ ({ \
26706  int32x4_t __s1 = __p1; \
26707  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
26708 })
26709 #else
26710 #define vst1q_s32(__p0, __p1) __extension__ ({ \
26711  int32x4_t __s1 = __p1; \
26712  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26713  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
26714 })
26715 #endif
26716 
26717 #ifdef __LITTLE_ENDIAN__
26718 #define vst1q_s64(__p0, __p1) __extension__ ({ \
26719  int64x2_t __s1 = __p1; \
26720  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
26721 })
26722 #else
26723 #define vst1q_s64(__p0, __p1) __extension__ ({ \
26724  int64x2_t __s1 = __p1; \
26725  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26726  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
26727 })
26728 #endif
26729 
26730 #ifdef __LITTLE_ENDIAN__
26731 #define vst1q_s16(__p0, __p1) __extension__ ({ \
26732  int16x8_t __s1 = __p1; \
26733  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
26734 })
26735 #else
26736 #define vst1q_s16(__p0, __p1) __extension__ ({ \
26737  int16x8_t __s1 = __p1; \
26738  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26739  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
26740 })
26741 #endif
26742 
26743 #ifdef __LITTLE_ENDIAN__
26744 #define vst1_u8(__p0, __p1) __extension__ ({ \
26745  uint8x8_t __s1 = __p1; \
26746  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
26747 })
26748 #else
26749 #define vst1_u8(__p0, __p1) __extension__ ({ \
26750  uint8x8_t __s1 = __p1; \
26751  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26752  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
26753 })
26754 #endif
26755 
26756 #ifdef __LITTLE_ENDIAN__
26757 #define vst1_u32(__p0, __p1) __extension__ ({ \
26758  uint32x2_t __s1 = __p1; \
26759  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
26760 })
26761 #else
26762 #define vst1_u32(__p0, __p1) __extension__ ({ \
26763  uint32x2_t __s1 = __p1; \
26764  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26765  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
26766 })
26767 #endif
26768 
26769 #define vst1_u64(__p0, __p1) __extension__ ({ \
26770  uint64x1_t __s1 = __p1; \
26771  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
26772 })
26773 #ifdef __LITTLE_ENDIAN__
26774 #define vst1_u16(__p0, __p1) __extension__ ({ \
26775  uint16x4_t __s1 = __p1; \
26776  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
26777 })
26778 #else
26779 #define vst1_u16(__p0, __p1) __extension__ ({ \
26780  uint16x4_t __s1 = __p1; \
26781  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26782  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
26783 })
26784 #endif
26785 
26786 #ifdef __LITTLE_ENDIAN__
26787 #define vst1_s8(__p0, __p1) __extension__ ({ \
26788  int8x8_t __s1 = __p1; \
26789  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
26790 })
26791 #else
26792 #define vst1_s8(__p0, __p1) __extension__ ({ \
26793  int8x8_t __s1 = __p1; \
26794  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26795  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
26796 })
26797 #endif
26798 
26799 #ifdef __LITTLE_ENDIAN__
26800 #define vst1_f32(__p0, __p1) __extension__ ({ \
26801  float32x2_t __s1 = __p1; \
26802  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
26803 })
26804 #else
26805 #define vst1_f32(__p0, __p1) __extension__ ({ \
26806  float32x2_t __s1 = __p1; \
26807  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26808  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
26809 })
26810 #endif
26811 
26812 #ifdef __LITTLE_ENDIAN__
26813 #define vst1_s32(__p0, __p1) __extension__ ({ \
26814  int32x2_t __s1 = __p1; \
26815  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
26816 })
26817 #else
26818 #define vst1_s32(__p0, __p1) __extension__ ({ \
26819  int32x2_t __s1 = __p1; \
26820  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26821  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
26822 })
26823 #endif
26824 
26825 #define vst1_s64(__p0, __p1) __extension__ ({ \
26826  int64x1_t __s1 = __p1; \
26827  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
26828 })
26829 #ifdef __LITTLE_ENDIAN__
26830 #define vst1_s16(__p0, __p1) __extension__ ({ \
26831  int16x4_t __s1 = __p1; \
26832  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
26833 })
26834 #else
26835 #define vst1_s16(__p0, __p1) __extension__ ({ \
26836  int16x4_t __s1 = __p1; \
26837  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26838  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
26839 })
26840 #endif
26841 
26842 #ifdef __LITTLE_ENDIAN__
26843 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26844  poly8x8_t __s1 = __p1; \
26845  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
26846 })
26847 #else
26848 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26849  poly8x8_t __s1 = __p1; \
26850  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26851  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
26852 })
26853 #endif
26854 
26855 #ifdef __LITTLE_ENDIAN__
26856 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26857  poly16x4_t __s1 = __p1; \
26858  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
26859 })
26860 #else
26861 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26862  poly16x4_t __s1 = __p1; \
26863  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26864  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
26865 })
26866 #endif
26867 
26868 #ifdef __LITTLE_ENDIAN__
26869 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26870  poly8x16_t __s1 = __p1; \
26871  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
26872 })
26873 #else
26874 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26875  poly8x16_t __s1 = __p1; \
26876  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26877  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
26878 })
26879 #endif
26880 
26881 #ifdef __LITTLE_ENDIAN__
26882 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26883  poly16x8_t __s1 = __p1; \
26884  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
26885 })
26886 #else
26887 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26888  poly16x8_t __s1 = __p1; \
26889  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26890  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
26891 })
26892 #endif
26893 
26894 #ifdef __LITTLE_ENDIAN__
26895 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26896  uint8x16_t __s1 = __p1; \
26897  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
26898 })
26899 #else
26900 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26901  uint8x16_t __s1 = __p1; \
26902  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26903  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
26904 })
26905 #endif
26906 
26907 #ifdef __LITTLE_ENDIAN__
26908 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26909  uint32x4_t __s1 = __p1; \
26910  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
26911 })
26912 #else
26913 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26914  uint32x4_t __s1 = __p1; \
26915  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26916  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
26917 })
26918 #endif
26919 
26920 #ifdef __LITTLE_ENDIAN__
26921 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
26922  uint64x2_t __s1 = __p1; \
26923  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
26924 })
26925 #else
26926 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
26927  uint64x2_t __s1 = __p1; \
26928  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26929  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
26930 })
26931 #endif
26932 
26933 #ifdef __LITTLE_ENDIAN__
26934 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26935  uint16x8_t __s1 = __p1; \
26936  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
26937 })
26938 #else
26939 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26940  uint16x8_t __s1 = __p1; \
26941  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
26942  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
26943 })
26944 #endif
26945 
26946 #ifdef __LITTLE_ENDIAN__
26947 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26948  int8x16_t __s1 = __p1; \
26949  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
26950 })
26951 #else
26952 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26953  int8x16_t __s1 = __p1; \
26954  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26955  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
26956 })
26957 #endif
26958 
26959 #ifdef __LITTLE_ENDIAN__
26960 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26961  float32x4_t __s1 = __p1; \
26962  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
26963 })
26964 #else
26965 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26966  float32x4_t __s1 = __p1; \
26967  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26968  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
26969 })
26970 #endif
26971 
26972 #ifdef __LITTLE_ENDIAN__
26973 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26974  int32x4_t __s1 = __p1; \
26975  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
26976 })
26977 #else
26978 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26979  int32x4_t __s1 = __p1; \
26980  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
26981  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
26982 })
26983 #endif
26984 
26985 #ifdef __LITTLE_ENDIAN__
26986 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
26987  int64x2_t __s1 = __p1; \
26988  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
26989 })
26990 #else
26991 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
26992  int64x2_t __s1 = __p1; \
26993  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
26994  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
26995 })
26996 #endif
26997 
26998 #ifdef __LITTLE_ENDIAN__
26999 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27000  int16x8_t __s1 = __p1; \
27001  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
27002 })
27003 #else
27004 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27005  int16x8_t __s1 = __p1; \
27006  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
27007  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
27008 })
27009 #endif
27010 
27011 #ifdef __LITTLE_ENDIAN__
27012 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27013  uint8x8_t __s1 = __p1; \
27014  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
27015 })
27016 #else
27017 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27018  uint8x8_t __s1 = __p1; \
27019  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
27020  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
27021 })
27022 #endif
27023 
27024 #ifdef __LITTLE_ENDIAN__
27025 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27026  uint32x2_t __s1 = __p1; \
27027  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
27028 })
27029 #else
27030 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27031  uint32x2_t __s1 = __p1; \
27032  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
27033  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
27034 })
27035 #endif
27036 
27037 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
27038  uint64x1_t __s1 = __p1; \
27039  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
27040 })
27041 #ifdef __LITTLE_ENDIAN__
27042 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27043  uint16x4_t __s1 = __p1; \
27044  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
27045 })
27046 #else
27047 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27048  uint16x4_t __s1 = __p1; \
27049  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
27050  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
27051 })
27052 #endif
27053 
27054 #ifdef __LITTLE_ENDIAN__
27055 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27056  int8x8_t __s1 = __p1; \
27057  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
27058 })
27059 #else
27060 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27061  int8x8_t __s1 = __p1; \
27062  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
27063  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
27064 })
27065 #endif
27066 
27067 #ifdef __LITTLE_ENDIAN__
27068 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27069  float32x2_t __s1 = __p1; \
27070  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
27071 })
27072 #else
27073 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27074  float32x2_t __s1 = __p1; \
27075  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
27076  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
27077 })
27078 #endif
27079 
27080 #ifdef __LITTLE_ENDIAN__
27081 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27082  int32x2_t __s1 = __p1; \
27083  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
27084 })
27085 #else
27086 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27087  int32x2_t __s1 = __p1; \
27088  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
27089  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
27090 })
27091 #endif
27092 
27093 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
27094  int64x1_t __s1 = __p1; \
27095  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
27096 })
27097 #ifdef __LITTLE_ENDIAN__
27098 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27099  int16x4_t __s1 = __p1; \
27100  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
27101 })
27102 #else
27103 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27104  int16x4_t __s1 = __p1; \
27105  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
27106  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
27107 })
27108 #endif
27109 
27110 #ifdef __LITTLE_ENDIAN__
27111 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
27112  poly8x8x2_t __s1 = __p1; \
27113  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
27114 })
27115 #else
27116 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
27117  poly8x8x2_t __s1 = __p1; \
27118  poly8x8x2_t __rev1; \
27119  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27120  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27121  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
27122 })
27123 #endif
27124 
27125 #ifdef __LITTLE_ENDIAN__
27126 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
27127  poly16x4x2_t __s1 = __p1; \
27128  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
27129 })
27130 #else
27131 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
27132  poly16x4x2_t __s1 = __p1; \
27133  poly16x4x2_t __rev1; \
27134  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27135  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27136  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
27137 })
27138 #endif
27139 
27140 #ifdef __LITTLE_ENDIAN__
27141 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
27142  poly8x16x2_t __s1 = __p1; \
27143  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
27144 })
27145 #else
27146 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
27147  poly8x16x2_t __s1 = __p1; \
27148  poly8x16x2_t __rev1; \
27149  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27150  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27151  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
27152 })
27153 #endif
27154 
27155 #ifdef __LITTLE_ENDIAN__
27156 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
27157  poly16x8x2_t __s1 = __p1; \
27158  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
27159 })
27160 #else
27161 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
27162  poly16x8x2_t __s1 = __p1; \
27163  poly16x8x2_t __rev1; \
27164  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27165  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27166  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
27167 })
27168 #endif
27169 
27170 #ifdef __LITTLE_ENDIAN__
27171 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
27172  uint8x16x2_t __s1 = __p1; \
27173  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
27174 })
27175 #else
27176 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
27177  uint8x16x2_t __s1 = __p1; \
27178  uint8x16x2_t __rev1; \
27179  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27180  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27181  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
27182 })
27183 #endif
27184 
27185 #ifdef __LITTLE_ENDIAN__
27186 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
27187  uint32x4x2_t __s1 = __p1; \
27188  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
27189 })
27190 #else
27191 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
27192  uint32x4x2_t __s1 = __p1; \
27193  uint32x4x2_t __rev1; \
27194  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27195  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27196  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
27197 })
27198 #endif
27199 
27200 #ifdef __LITTLE_ENDIAN__
27201 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
27202  uint64x2x2_t __s1 = __p1; \
27203  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
27204 })
27205 #else
27206 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
27207  uint64x2x2_t __s1 = __p1; \
27208  uint64x2x2_t __rev1; \
27209  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27210  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27211  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
27212 })
27213 #endif
27214 
27215 #ifdef __LITTLE_ENDIAN__
27216 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
27217  uint16x8x2_t __s1 = __p1; \
27218  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
27219 })
27220 #else
27221 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
27222  uint16x8x2_t __s1 = __p1; \
27223  uint16x8x2_t __rev1; \
27224  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27225  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27226  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
27227 })
27228 #endif
27229 
27230 #ifdef __LITTLE_ENDIAN__
27231 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
27232  int8x16x2_t __s1 = __p1; \
27233  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
27234 })
27235 #else
27236 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
27237  int8x16x2_t __s1 = __p1; \
27238  int8x16x2_t __rev1; \
27239  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27240  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27241  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
27242 })
27243 #endif
27244 
27245 #ifdef __LITTLE_ENDIAN__
27246 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
27247  float32x4x2_t __s1 = __p1; \
27248  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 41); \
27249 })
27250 #else
27251 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
27252  float32x4x2_t __s1 = __p1; \
27253  float32x4x2_t __rev1; \
27254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27256  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 41); \
27257 })
27258 #endif
27259 
27260 #ifdef __LITTLE_ENDIAN__
27261 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
27262  int32x4x2_t __s1 = __p1; \
27263  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 34); \
27264 })
27265 #else
27266 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
27267  int32x4x2_t __s1 = __p1; \
27268  int32x4x2_t __rev1; \
27269  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27270  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27271  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 34); \
27272 })
27273 #endif
27274 
27275 #ifdef __LITTLE_ENDIAN__
27276 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
27277  int64x2x2_t __s1 = __p1; \
27278  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 35); \
27279 })
27280 #else
27281 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
27282  int64x2x2_t __s1 = __p1; \
27283  int64x2x2_t __rev1; \
27284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27286  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 35); \
27287 })
27288 #endif
27289 
27290 #ifdef __LITTLE_ENDIAN__
27291 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
27292  int16x8x2_t __s1 = __p1; \
27293  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 33); \
27294 })
27295 #else
27296 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
27297  int16x8x2_t __s1 = __p1; \
27298  int16x8x2_t __rev1; \
27299  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27300  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27301  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 33); \
27302 })
27303 #endif
27304 
27305 #ifdef __LITTLE_ENDIAN__
27306 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
27307  uint8x8x2_t __s1 = __p1; \
27308  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
27309 })
27310 #else
27311 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
27312  uint8x8x2_t __s1 = __p1; \
27313  uint8x8x2_t __rev1; \
27314  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27315  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27316  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
27317 })
27318 #endif
27319 
27320 #ifdef __LITTLE_ENDIAN__
27321 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
27322  uint32x2x2_t __s1 = __p1; \
27323  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
27324 })
27325 #else
27326 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
27327  uint32x2x2_t __s1 = __p1; \
27328  uint32x2x2_t __rev1; \
27329  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27330  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27331  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
27332 })
27333 #endif
27334 
27335 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
27336  uint64x1x2_t __s1 = __p1; \
27337  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
27338 })
27339 #ifdef __LITTLE_ENDIAN__
27340 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
27341  uint16x4x2_t __s1 = __p1; \
27342  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
27343 })
27344 #else
27345 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
27346  uint16x4x2_t __s1 = __p1; \
27347  uint16x4x2_t __rev1; \
27348  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27349  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27350  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
27351 })
27352 #endif
27353 
27354 #ifdef __LITTLE_ENDIAN__
27355 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
27356  int8x8x2_t __s1 = __p1; \
27357  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
27358 })
27359 #else
27360 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
27361  int8x8x2_t __s1 = __p1; \
27362  int8x8x2_t __rev1; \
27363  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27364  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27365  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
27366 })
27367 #endif
27368 
27369 #ifdef __LITTLE_ENDIAN__
27370 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
27371  float32x2x2_t __s1 = __p1; \
27372  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 9); \
27373 })
27374 #else
27375 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
27376  float32x2x2_t __s1 = __p1; \
27377  float32x2x2_t __rev1; \
27378  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27379  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27380  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 9); \
27381 })
27382 #endif
27383 
27384 #ifdef __LITTLE_ENDIAN__
27385 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
27386  int32x2x2_t __s1 = __p1; \
27387  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 2); \
27388 })
27389 #else
27390 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
27391  int32x2x2_t __s1 = __p1; \
27392  int32x2x2_t __rev1; \
27393  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27394  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27395  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 2); \
27396 })
27397 #endif
27398 
27399 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
27400  int64x1x2_t __s1 = __p1; \
27401  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 3); \
27402 })
27403 #ifdef __LITTLE_ENDIAN__
27404 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
27405  int16x4x2_t __s1 = __p1; \
27406  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 1); \
27407 })
27408 #else
27409 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
27410  int16x4x2_t __s1 = __p1; \
27411  int16x4x2_t __rev1; \
27412  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27413  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27414  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 1); \
27415 })
27416 #endif
27417 
27418 #ifdef __LITTLE_ENDIAN__
27419 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
27420  poly8x8x3_t __s1 = __p1; \
27421  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
27422 })
27423 #else
27424 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
27425  poly8x8x3_t __s1 = __p1; \
27426  poly8x8x3_t __rev1; \
27427  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27428  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27429  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27430  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
27431 })
27432 #endif
27433 
27434 #ifdef __LITTLE_ENDIAN__
27435 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
27436  poly16x4x3_t __s1 = __p1; \
27437  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
27438 })
27439 #else
27440 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
27441  poly16x4x3_t __s1 = __p1; \
27442  poly16x4x3_t __rev1; \
27443  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27444  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27445  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27446  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
27447 })
27448 #endif
27449 
27450 #ifdef __LITTLE_ENDIAN__
27451 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
27452  poly8x16x3_t __s1 = __p1; \
27453  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
27454 })
27455 #else
27456 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
27457  poly8x16x3_t __s1 = __p1; \
27458  poly8x16x3_t __rev1; \
27459  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27460  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27461  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27462  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
27463 })
27464 #endif
27465 
27466 #ifdef __LITTLE_ENDIAN__
27467 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
27468  poly16x8x3_t __s1 = __p1; \
27469  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
27470 })
27471 #else
27472 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
27473  poly16x8x3_t __s1 = __p1; \
27474  poly16x8x3_t __rev1; \
27475  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27476  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27477  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27478  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
27479 })
27480 #endif
27481 
27482 #ifdef __LITTLE_ENDIAN__
27483 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
27484  uint8x16x3_t __s1 = __p1; \
27485  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
27486 })
27487 #else
27488 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
27489  uint8x16x3_t __s1 = __p1; \
27490  uint8x16x3_t __rev1; \
27491  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27492  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27493  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27494  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
27495 })
27496 #endif
27497 
27498 #ifdef __LITTLE_ENDIAN__
27499 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
27500  uint32x4x3_t __s1 = __p1; \
27501  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
27502 })
27503 #else
27504 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
27505  uint32x4x3_t __s1 = __p1; \
27506  uint32x4x3_t __rev1; \
27507  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27508  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27509  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27510  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
27511 })
27512 #endif
27513 
27514 #ifdef __LITTLE_ENDIAN__
27515 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
27516  uint64x2x3_t __s1 = __p1; \
27517  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
27518 })
27519 #else
27520 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
27521  uint64x2x3_t __s1 = __p1; \
27522  uint64x2x3_t __rev1; \
27523  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27524  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27525  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27526  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
27527 })
27528 #endif
27529 
27530 #ifdef __LITTLE_ENDIAN__
27531 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
27532  uint16x8x3_t __s1 = __p1; \
27533  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
27534 })
27535 #else
27536 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
27537  uint16x8x3_t __s1 = __p1; \
27538  uint16x8x3_t __rev1; \
27539  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27540  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27541  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27542  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
27543 })
27544 #endif
27545 
27546 #ifdef __LITTLE_ENDIAN__
27547 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
27548  int8x16x3_t __s1 = __p1; \
27549  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
27550 })
27551 #else
27552 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
27553  int8x16x3_t __s1 = __p1; \
27554  int8x16x3_t __rev1; \
27555  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27556  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27557  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27558  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
27559 })
27560 #endif
27561 
27562 #ifdef __LITTLE_ENDIAN__
27563 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
27564  float32x4x3_t __s1 = __p1; \
27565  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 41); \
27566 })
27567 #else
27568 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
27569  float32x4x3_t __s1 = __p1; \
27570  float32x4x3_t __rev1; \
27571  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27572  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27573  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27574  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 41); \
27575 })
27576 #endif
27577 
27578 #ifdef __LITTLE_ENDIAN__
27579 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
27580  int32x4x3_t __s1 = __p1; \
27581  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 34); \
27582 })
27583 #else
27584 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
27585  int32x4x3_t __s1 = __p1; \
27586  int32x4x3_t __rev1; \
27587  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27588  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27589  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27590  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 34); \
27591 })
27592 #endif
27593 
27594 #ifdef __LITTLE_ENDIAN__
27595 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
27596  int64x2x3_t __s1 = __p1; \
27597  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 35); \
27598 })
27599 #else
27600 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
27601  int64x2x3_t __s1 = __p1; \
27602  int64x2x3_t __rev1; \
27603  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27604  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27605  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27606  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 35); \
27607 })
27608 #endif
27609 
27610 #ifdef __LITTLE_ENDIAN__
27611 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
27612  int16x8x3_t __s1 = __p1; \
27613  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 33); \
27614 })
27615 #else
27616 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
27617  int16x8x3_t __s1 = __p1; \
27618  int16x8x3_t __rev1; \
27619  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27620  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27621  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27622  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 33); \
27623 })
27624 #endif
27625 
27626 #ifdef __LITTLE_ENDIAN__
27627 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
27628  uint8x8x3_t __s1 = __p1; \
27629  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
27630 })
27631 #else
27632 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
27633  uint8x8x3_t __s1 = __p1; \
27634  uint8x8x3_t __rev1; \
27635  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27636  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27637  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27638  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
27639 })
27640 #endif
27641 
27642 #ifdef __LITTLE_ENDIAN__
27643 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
27644  uint32x2x3_t __s1 = __p1; \
27645  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
27646 })
27647 #else
27648 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
27649  uint32x2x3_t __s1 = __p1; \
27650  uint32x2x3_t __rev1; \
27651  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27652  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27653  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27654  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
27655 })
27656 #endif
27657 
27658 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
27659  uint64x1x3_t __s1 = __p1; \
27660  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
27661 })
27662 #ifdef __LITTLE_ENDIAN__
27663 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
27664  uint16x4x3_t __s1 = __p1; \
27665  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
27666 })
27667 #else
27668 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
27669  uint16x4x3_t __s1 = __p1; \
27670  uint16x4x3_t __rev1; \
27671  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27672  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27673  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27674  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
27675 })
27676 #endif
27677 
27678 #ifdef __LITTLE_ENDIAN__
27679 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
27680  int8x8x3_t __s1 = __p1; \
27681  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
27682 })
27683 #else
27684 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
27685  int8x8x3_t __s1 = __p1; \
27686  int8x8x3_t __rev1; \
27687  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27688  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27689  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27690  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
27691 })
27692 #endif
27693 
27694 #ifdef __LITTLE_ENDIAN__
27695 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
27696  float32x2x3_t __s1 = __p1; \
27697  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 9); \
27698 })
27699 #else
27700 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
27701  float32x2x3_t __s1 = __p1; \
27702  float32x2x3_t __rev1; \
27703  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27704  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27705  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27706  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 9); \
27707 })
27708 #endif
27709 
27710 #ifdef __LITTLE_ENDIAN__
27711 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
27712  int32x2x3_t __s1 = __p1; \
27713  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 2); \
27714 })
27715 #else
27716 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
27717  int32x2x3_t __s1 = __p1; \
27718  int32x2x3_t __rev1; \
27719  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27720  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27721  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27722  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 2); \
27723 })
27724 #endif
27725 
27726 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
27727  int64x1x3_t __s1 = __p1; \
27728  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 3); \
27729 })
27730 #ifdef __LITTLE_ENDIAN__
27731 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
27732  int16x4x3_t __s1 = __p1; \
27733  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 1); \
27734 })
27735 #else
27736 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
27737  int16x4x3_t __s1 = __p1; \
27738  int16x4x3_t __rev1; \
27739  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27740  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27741  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27742  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 1); \
27743 })
27744 #endif
27745 
27746 #ifdef __LITTLE_ENDIAN__
27747 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
27748  poly8x8x4_t __s1 = __p1; \
27749  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
27750 })
27751 #else
27752 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
27753  poly8x8x4_t __s1 = __p1; \
27754  poly8x8x4_t __rev1; \
27755  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27756  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27757  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27758  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27759  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
27760 })
27761 #endif
27762 
27763 #ifdef __LITTLE_ENDIAN__
27764 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
27765  poly16x4x4_t __s1 = __p1; \
27766  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
27767 })
27768 #else
27769 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
27770  poly16x4x4_t __s1 = __p1; \
27771  poly16x4x4_t __rev1; \
27772  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27773  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27774  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27775  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27776  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
27777 })
27778 #endif
27779 
27780 #ifdef __LITTLE_ENDIAN__
27781 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
27782  poly8x16x4_t __s1 = __p1; \
27783  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
27784 })
27785 #else
27786 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
27787  poly8x16x4_t __s1 = __p1; \
27788  poly8x16x4_t __rev1; \
27789  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27790  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27791  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27792  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27793  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
27794 })
27795 #endif
27796 
27797 #ifdef __LITTLE_ENDIAN__
27798 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
27799  poly16x8x4_t __s1 = __p1; \
27800  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
27801 })
27802 #else
27803 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
27804  poly16x8x4_t __s1 = __p1; \
27805  poly16x8x4_t __rev1; \
27806  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27807  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27808  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27809  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27810  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
27811 })
27812 #endif
27813 
27814 #ifdef __LITTLE_ENDIAN__
27815 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
27816  uint8x16x4_t __s1 = __p1; \
27817  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
27818 })
27819 #else
27820 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
27821  uint8x16x4_t __s1 = __p1; \
27822  uint8x16x4_t __rev1; \
27823  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27824  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27825  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27826  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27827  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
27828 })
27829 #endif
27830 
27831 #ifdef __LITTLE_ENDIAN__
27832 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
27833  uint32x4x4_t __s1 = __p1; \
27834  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
27835 })
27836 #else
27837 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
27838  uint32x4x4_t __s1 = __p1; \
27839  uint32x4x4_t __rev1; \
27840  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27841  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27842  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27843  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27844  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
27845 })
27846 #endif
27847 
27848 #ifdef __LITTLE_ENDIAN__
27849 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
27850  uint64x2x4_t __s1 = __p1; \
27851  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
27852 })
27853 #else
27854 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
27855  uint64x2x4_t __s1 = __p1; \
27856  uint64x2x4_t __rev1; \
27857  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27858  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27859  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27860  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27861  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
27862 })
27863 #endif
27864 
27865 #ifdef __LITTLE_ENDIAN__
27866 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
27867  uint16x8x4_t __s1 = __p1; \
27868  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
27869 })
27870 #else
27871 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
27872  uint16x8x4_t __s1 = __p1; \
27873  uint16x8x4_t __rev1; \
27874  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27875  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27876  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27877  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27878  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
27879 })
27880 #endif
27881 
27882 #ifdef __LITTLE_ENDIAN__
27883 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
27884  int8x16x4_t __s1 = __p1; \
27885  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
27886 })
27887 #else
27888 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
27889  int8x16x4_t __s1 = __p1; \
27890  int8x16x4_t __rev1; \
27891  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27892  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27893  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27894  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
27895  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
27896 })
27897 #endif
27898 
27899 #ifdef __LITTLE_ENDIAN__
27900 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
27901  float32x4x4_t __s1 = __p1; \
27902  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 41); \
27903 })
27904 #else
27905 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
27906  float32x4x4_t __s1 = __p1; \
27907  float32x4x4_t __rev1; \
27908  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27909  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27910  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27911  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27912  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 41); \
27913 })
27914 #endif
27915 
27916 #ifdef __LITTLE_ENDIAN__
27917 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
27918  int32x4x4_t __s1 = __p1; \
27919  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 34); \
27920 })
27921 #else
27922 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
27923  int32x4x4_t __s1 = __p1; \
27924  int32x4x4_t __rev1; \
27925  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27926  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27927  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27928  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27929  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 34); \
27930 })
27931 #endif
27932 
27933 #ifdef __LITTLE_ENDIAN__
27934 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
27935  int64x2x4_t __s1 = __p1; \
27936  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 35); \
27937 })
27938 #else
27939 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
27940  int64x2x4_t __s1 = __p1; \
27941  int64x2x4_t __rev1; \
27942  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27943  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27944  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27945  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27946  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 35); \
27947 })
27948 #endif
27949 
27950 #ifdef __LITTLE_ENDIAN__
27951 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
27952  int16x8x4_t __s1 = __p1; \
27953  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 33); \
27954 })
27955 #else
27956 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
27957  int16x8x4_t __s1 = __p1; \
27958  int16x8x4_t __rev1; \
27959  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27960  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27961  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27962  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27963  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 33); \
27964 })
27965 #endif
27966 
27967 #ifdef __LITTLE_ENDIAN__
27968 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
27969  uint8x8x4_t __s1 = __p1; \
27970  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
27971 })
27972 #else
27973 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
27974  uint8x8x4_t __s1 = __p1; \
27975  uint8x8x4_t __rev1; \
27976  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27977  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27978  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27979  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27980  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
27981 })
27982 #endif
27983 
27984 #ifdef __LITTLE_ENDIAN__
27985 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
27986  uint32x2x4_t __s1 = __p1; \
27987  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27988 })
27989 #else
27990 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
27991  uint32x2x4_t __s1 = __p1; \
27992  uint32x2x4_t __rev1; \
27993  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27994  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27995  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27996  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27997  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27998 })
27999 #endif
28000 
28001 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
28002  uint64x1x4_t __s1 = __p1; \
28003  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
28004 })
28005 #ifdef __LITTLE_ENDIAN__
28006 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
28007  uint16x4x4_t __s1 = __p1; \
28008  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
28009 })
28010 #else
28011 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
28012  uint16x4x4_t __s1 = __p1; \
28013  uint16x4x4_t __rev1; \
28014  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28015  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28016  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28017  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
28018  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
28019 })
28020 #endif
28021 
28022 #ifdef __LITTLE_ENDIAN__
28023 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
28024  int8x8x4_t __s1 = __p1; \
28025  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
28026 })
28027 #else
28028 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
28029  int8x8x4_t __s1 = __p1; \
28030  int8x8x4_t __rev1; \
28031  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28032  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28033  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28034  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
28035  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
28036 })
28037 #endif
28038 
28039 #ifdef __LITTLE_ENDIAN__
28040 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
28041  float32x2x4_t __s1 = __p1; \
28042  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 9); \
28043 })
28044 #else
28045 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
28046  float32x2x4_t __s1 = __p1; \
28047  float32x2x4_t __rev1; \
28048  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28049  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28050  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28051  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
28052  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 9); \
28053 })
28054 #endif
28055 
28056 #ifdef __LITTLE_ENDIAN__
28057 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
28058  int32x2x4_t __s1 = __p1; \
28059  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 2); \
28060 })
28061 #else
28062 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
28063  int32x2x4_t __s1 = __p1; \
28064  int32x2x4_t __rev1; \
28065  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28066  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28067  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28068  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
28069  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 2); \
28070 })
28071 #endif
28072 
28073 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
28074  int64x1x4_t __s1 = __p1; \
28075  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 3); \
28076 })
28077 #ifdef __LITTLE_ENDIAN__
28078 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
28079  int16x4x4_t __s1 = __p1; \
28080  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 1); \
28081 })
28082 #else
28083 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
28084  int16x4x4_t __s1 = __p1; \
28085  int16x4x4_t __rev1; \
28086  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28087  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28088  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28089  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
28090  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 1); \
28091 })
28092 #endif
28093 
28094 #ifdef __LITTLE_ENDIAN__
28095 #define vst2_p8(__p0, __p1) __extension__ ({ \
28096  poly8x8x2_t __s1 = __p1; \
28097  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
28098 })
28099 #else
28100 #define vst2_p8(__p0, __p1) __extension__ ({ \
28101  poly8x8x2_t __s1 = __p1; \
28102  poly8x8x2_t __rev1; \
28103  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28104  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28105  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
28106 })
28107 #endif
28108 
28109 #ifdef __LITTLE_ENDIAN__
28110 #define vst2_p16(__p0, __p1) __extension__ ({ \
28111  poly16x4x2_t __s1 = __p1; \
28112  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
28113 })
28114 #else
28115 #define vst2_p16(__p0, __p1) __extension__ ({ \
28116  poly16x4x2_t __s1 = __p1; \
28117  poly16x4x2_t __rev1; \
28118  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28119  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28120  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
28121 })
28122 #endif
28123 
28124 #ifdef __LITTLE_ENDIAN__
28125 #define vst2q_p8(__p0, __p1) __extension__ ({ \
28126  poly8x16x2_t __s1 = __p1; \
28127  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
28128 })
28129 #else
28130 #define vst2q_p8(__p0, __p1) __extension__ ({ \
28131  poly8x16x2_t __s1 = __p1; \
28132  poly8x16x2_t __rev1; \
28133  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28134  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28135  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
28136 })
28137 #endif
28138 
28139 #ifdef __LITTLE_ENDIAN__
28140 #define vst2q_p16(__p0, __p1) __extension__ ({ \
28141  poly16x8x2_t __s1 = __p1; \
28142  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
28143 })
28144 #else
28145 #define vst2q_p16(__p0, __p1) __extension__ ({ \
28146  poly16x8x2_t __s1 = __p1; \
28147  poly16x8x2_t __rev1; \
28148  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28149  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28150  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
28151 })
28152 #endif
28153 
28154 #ifdef __LITTLE_ENDIAN__
28155 #define vst2q_u8(__p0, __p1) __extension__ ({ \
28156  uint8x16x2_t __s1 = __p1; \
28157  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
28158 })
28159 #else
28160 #define vst2q_u8(__p0, __p1) __extension__ ({ \
28161  uint8x16x2_t __s1 = __p1; \
28162  uint8x16x2_t __rev1; \
28163  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28164  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28165  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
28166 })
28167 #endif
28168 
28169 #ifdef __LITTLE_ENDIAN__
28170 #define vst2q_u32(__p0, __p1) __extension__ ({ \
28171  uint32x4x2_t __s1 = __p1; \
28172  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
28173 })
28174 #else
28175 #define vst2q_u32(__p0, __p1) __extension__ ({ \
28176  uint32x4x2_t __s1 = __p1; \
28177  uint32x4x2_t __rev1; \
28178  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28179  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28180  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
28181 })
28182 #endif
28183 
28184 #ifdef __LITTLE_ENDIAN__
28185 #define vst2q_u16(__p0, __p1) __extension__ ({ \
28186  uint16x8x2_t __s1 = __p1; \
28187  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
28188 })
28189 #else
28190 #define vst2q_u16(__p0, __p1) __extension__ ({ \
28191  uint16x8x2_t __s1 = __p1; \
28192  uint16x8x2_t __rev1; \
28193  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28194  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28195  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
28196 })
28197 #endif
28198 
28199 #ifdef __LITTLE_ENDIAN__
28200 #define vst2q_s8(__p0, __p1) __extension__ ({ \
28201  int8x16x2_t __s1 = __p1; \
28202  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
28203 })
28204 #else
28205 #define vst2q_s8(__p0, __p1) __extension__ ({ \
28206  int8x16x2_t __s1 = __p1; \
28207  int8x16x2_t __rev1; \
28208  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28209  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28210  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
28211 })
28212 #endif
28213 
28214 #ifdef __LITTLE_ENDIAN__
28215 #define vst2q_f32(__p0, __p1) __extension__ ({ \
28216  float32x4x2_t __s1 = __p1; \
28217  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 41); \
28218 })
28219 #else
28220 #define vst2q_f32(__p0, __p1) __extension__ ({ \
28221  float32x4x2_t __s1 = __p1; \
28222  float32x4x2_t __rev1; \
28223  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28224  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28225  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 41); \
28226 })
28227 #endif
28228 
28229 #ifdef __LITTLE_ENDIAN__
28230 #define vst2q_s32(__p0, __p1) __extension__ ({ \
28231  int32x4x2_t __s1 = __p1; \
28232  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 34); \
28233 })
28234 #else
28235 #define vst2q_s32(__p0, __p1) __extension__ ({ \
28236  int32x4x2_t __s1 = __p1; \
28237  int32x4x2_t __rev1; \
28238  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28239  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28240  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 34); \
28241 })
28242 #endif
28243 
28244 #ifdef __LITTLE_ENDIAN__
28245 #define vst2q_s16(__p0, __p1) __extension__ ({ \
28246  int16x8x2_t __s1 = __p1; \
28247  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 33); \
28248 })
28249 #else
28250 #define vst2q_s16(__p0, __p1) __extension__ ({ \
28251  int16x8x2_t __s1 = __p1; \
28252  int16x8x2_t __rev1; \
28253  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28254  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28255  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 33); \
28256 })
28257 #endif
28258 
28259 #ifdef __LITTLE_ENDIAN__
28260 #define vst2_u8(__p0, __p1) __extension__ ({ \
28261  uint8x8x2_t __s1 = __p1; \
28262  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
28263 })
28264 #else
28265 #define vst2_u8(__p0, __p1) __extension__ ({ \
28266  uint8x8x2_t __s1 = __p1; \
28267  uint8x8x2_t __rev1; \
28268  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28269  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28270  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
28271 })
28272 #endif
28273 
28274 #ifdef __LITTLE_ENDIAN__
28275 #define vst2_u32(__p0, __p1) __extension__ ({ \
28276  uint32x2x2_t __s1 = __p1; \
28277  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
28278 })
28279 #else
28280 #define vst2_u32(__p0, __p1) __extension__ ({ \
28281  uint32x2x2_t __s1 = __p1; \
28282  uint32x2x2_t __rev1; \
28283  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28284  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28285  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
28286 })
28287 #endif
28288 
28289 #define vst2_u64(__p0, __p1) __extension__ ({ \
28290  uint64x1x2_t __s1 = __p1; \
28291  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
28292 })
28293 #ifdef __LITTLE_ENDIAN__
28294 #define vst2_u16(__p0, __p1) __extension__ ({ \
28295  uint16x4x2_t __s1 = __p1; \
28296  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
28297 })
28298 #else
28299 #define vst2_u16(__p0, __p1) __extension__ ({ \
28300  uint16x4x2_t __s1 = __p1; \
28301  uint16x4x2_t __rev1; \
28302  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28303  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28304  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
28305 })
28306 #endif
28307 
28308 #ifdef __LITTLE_ENDIAN__
28309 #define vst2_s8(__p0, __p1) __extension__ ({ \
28310  int8x8x2_t __s1 = __p1; \
28311  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
28312 })
28313 #else
28314 #define vst2_s8(__p0, __p1) __extension__ ({ \
28315  int8x8x2_t __s1 = __p1; \
28316  int8x8x2_t __rev1; \
28317  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28318  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28319  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
28320 })
28321 #endif
28322 
28323 #ifdef __LITTLE_ENDIAN__
28324 #define vst2_f32(__p0, __p1) __extension__ ({ \
28325  float32x2x2_t __s1 = __p1; \
28326  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 9); \
28327 })
28328 #else
28329 #define vst2_f32(__p0, __p1) __extension__ ({ \
28330  float32x2x2_t __s1 = __p1; \
28331  float32x2x2_t __rev1; \
28332  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28333  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28334  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 9); \
28335 })
28336 #endif
28337 
28338 #ifdef __LITTLE_ENDIAN__
28339 #define vst2_s32(__p0, __p1) __extension__ ({ \
28340  int32x2x2_t __s1 = __p1; \
28341  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 2); \
28342 })
28343 #else
28344 #define vst2_s32(__p0, __p1) __extension__ ({ \
28345  int32x2x2_t __s1 = __p1; \
28346  int32x2x2_t __rev1; \
28347  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28348  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28349  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 2); \
28350 })
28351 #endif
28352 
28353 #define vst2_s64(__p0, __p1) __extension__ ({ \
28354  int64x1x2_t __s1 = __p1; \
28355  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 3); \
28356 })
28357 #ifdef __LITTLE_ENDIAN__
28358 #define vst2_s16(__p0, __p1) __extension__ ({ \
28359  int16x4x2_t __s1 = __p1; \
28360  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 1); \
28361 })
28362 #else
28363 #define vst2_s16(__p0, __p1) __extension__ ({ \
28364  int16x4x2_t __s1 = __p1; \
28365  int16x4x2_t __rev1; \
28366  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28367  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28368  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 1); \
28369 })
28370 #endif
28371 
28372 #ifdef __LITTLE_ENDIAN__
28373 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28374  poly8x8x2_t __s1 = __p1; \
28375  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
28376 })
28377 #else
28378 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28379  poly8x8x2_t __s1 = __p1; \
28380  poly8x8x2_t __rev1; \
28381  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28382  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28383  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
28384 })
28385 #endif
28386 
28387 #ifdef __LITTLE_ENDIAN__
28388 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28389  poly16x4x2_t __s1 = __p1; \
28390  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
28391 })
28392 #else
28393 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28394  poly16x4x2_t __s1 = __p1; \
28395  poly16x4x2_t __rev1; \
28396  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28397  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28398  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
28399 })
28400 #endif
28401 
28402 #ifdef __LITTLE_ENDIAN__
28403 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28404  poly16x8x2_t __s1 = __p1; \
28405  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
28406 })
28407 #else
28408 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28409  poly16x8x2_t __s1 = __p1; \
28410  poly16x8x2_t __rev1; \
28411  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28412  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28413  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
28414 })
28415 #endif
28416 
28417 #ifdef __LITTLE_ENDIAN__
28418 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28419  uint32x4x2_t __s1 = __p1; \
28420  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
28421 })
28422 #else
28423 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28424  uint32x4x2_t __s1 = __p1; \
28425  uint32x4x2_t __rev1; \
28426  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28427  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28428  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
28429 })
28430 #endif
28431 
28432 #ifdef __LITTLE_ENDIAN__
28433 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28434  uint16x8x2_t __s1 = __p1; \
28435  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
28436 })
28437 #else
28438 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28439  uint16x8x2_t __s1 = __p1; \
28440  uint16x8x2_t __rev1; \
28441  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28442  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28443  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
28444 })
28445 #endif
28446 
28447 #ifdef __LITTLE_ENDIAN__
28448 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28449  float32x4x2_t __s1 = __p1; \
28450  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 41); \
28451 })
28452 #else
28453 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28454  float32x4x2_t __s1 = __p1; \
28455  float32x4x2_t __rev1; \
28456  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28457  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28458  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 41); \
28459 })
28460 #endif
28461 
28462 #ifdef __LITTLE_ENDIAN__
28463 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28464  int32x4x2_t __s1 = __p1; \
28465  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 34); \
28466 })
28467 #else
28468 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28469  int32x4x2_t __s1 = __p1; \
28470  int32x4x2_t __rev1; \
28471  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28472  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28473  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 34); \
28474 })
28475 #endif
28476 
28477 #ifdef __LITTLE_ENDIAN__
28478 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28479  int16x8x2_t __s1 = __p1; \
28480  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 33); \
28481 })
28482 #else
28483 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28484  int16x8x2_t __s1 = __p1; \
28485  int16x8x2_t __rev1; \
28486  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28487  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28488  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 33); \
28489 })
28490 #endif
28491 
28492 #ifdef __LITTLE_ENDIAN__
28493 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28494  uint8x8x2_t __s1 = __p1; \
28495  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
28496 })
28497 #else
28498 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
28499  uint8x8x2_t __s1 = __p1; \
28500  uint8x8x2_t __rev1; \
28501  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28502  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28503  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
28504 })
28505 #endif
28506 
28507 #ifdef __LITTLE_ENDIAN__
28508 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28509  uint32x2x2_t __s1 = __p1; \
28510  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
28511 })
28512 #else
28513 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28514  uint32x2x2_t __s1 = __p1; \
28515  uint32x2x2_t __rev1; \
28516  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28517  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28518  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
28519 })
28520 #endif
28521 
28522 #ifdef __LITTLE_ENDIAN__
28523 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28524  uint16x4x2_t __s1 = __p1; \
28525  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
28526 })
28527 #else
28528 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28529  uint16x4x2_t __s1 = __p1; \
28530  uint16x4x2_t __rev1; \
28531  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28532  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28533  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
28534 })
28535 #endif
28536 
28537 #ifdef __LITTLE_ENDIAN__
28538 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28539  int8x8x2_t __s1 = __p1; \
28540  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
28541 })
28542 #else
28543 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
28544  int8x8x2_t __s1 = __p1; \
28545  int8x8x2_t __rev1; \
28546  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28547  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28548  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
28549 })
28550 #endif
28551 
28552 #ifdef __LITTLE_ENDIAN__
28553 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28554  float32x2x2_t __s1 = __p1; \
28555  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 9); \
28556 })
28557 #else
28558 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28559  float32x2x2_t __s1 = __p1; \
28560  float32x2x2_t __rev1; \
28561  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28562  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28563  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 9); \
28564 })
28565 #endif
28566 
28567 #ifdef __LITTLE_ENDIAN__
28568 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28569  int32x2x2_t __s1 = __p1; \
28570  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 2); \
28571 })
28572 #else
28573 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28574  int32x2x2_t __s1 = __p1; \
28575  int32x2x2_t __rev1; \
28576  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28577  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28578  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 2); \
28579 })
28580 #endif
28581 
28582 #ifdef __LITTLE_ENDIAN__
28583 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28584  int16x4x2_t __s1 = __p1; \
28585  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 1); \
28586 })
28587 #else
28588 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
28589  int16x4x2_t __s1 = __p1; \
28590  int16x4x2_t __rev1; \
28591  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28592  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28593  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 1); \
28594 })
28595 #endif
28596 
28597 #ifdef __LITTLE_ENDIAN__
28598 #define vst3_p8(__p0, __p1) __extension__ ({ \
28599  poly8x8x3_t __s1 = __p1; \
28600  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
28601 })
28602 #else
28603 #define vst3_p8(__p0, __p1) __extension__ ({ \
28604  poly8x8x3_t __s1 = __p1; \
28605  poly8x8x3_t __rev1; \
28606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28608  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28609  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
28610 })
28611 #endif
28612 
28613 #ifdef __LITTLE_ENDIAN__
28614 #define vst3_p16(__p0, __p1) __extension__ ({ \
28615  poly16x4x3_t __s1 = __p1; \
28616  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
28617 })
28618 #else
28619 #define vst3_p16(__p0, __p1) __extension__ ({ \
28620  poly16x4x3_t __s1 = __p1; \
28621  poly16x4x3_t __rev1; \
28622  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28623  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28624  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28625  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
28626 })
28627 #endif
28628 
28629 #ifdef __LITTLE_ENDIAN__
28630 #define vst3q_p8(__p0, __p1) __extension__ ({ \
28631  poly8x16x3_t __s1 = __p1; \
28632  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
28633 })
28634 #else
28635 #define vst3q_p8(__p0, __p1) __extension__ ({ \
28636  poly8x16x3_t __s1 = __p1; \
28637  poly8x16x3_t __rev1; \
28638  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28639  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28640  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28641  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
28642 })
28643 #endif
28644 
28645 #ifdef __LITTLE_ENDIAN__
28646 #define vst3q_p16(__p0, __p1) __extension__ ({ \
28647  poly16x8x3_t __s1 = __p1; \
28648  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
28649 })
28650 #else
28651 #define vst3q_p16(__p0, __p1) __extension__ ({ \
28652  poly16x8x3_t __s1 = __p1; \
28653  poly16x8x3_t __rev1; \
28654  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28655  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28656  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28657  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
28658 })
28659 #endif
28660 
28661 #ifdef __LITTLE_ENDIAN__
28662 #define vst3q_u8(__p0, __p1) __extension__ ({ \
28663  uint8x16x3_t __s1 = __p1; \
28664  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
28665 })
28666 #else
28667 #define vst3q_u8(__p0, __p1) __extension__ ({ \
28668  uint8x16x3_t __s1 = __p1; \
28669  uint8x16x3_t __rev1; \
28670  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28671  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28672  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28673  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
28674 })
28675 #endif
28676 
28677 #ifdef __LITTLE_ENDIAN__
28678 #define vst3q_u32(__p0, __p1) __extension__ ({ \
28679  uint32x4x3_t __s1 = __p1; \
28680  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
28681 })
28682 #else
28683 #define vst3q_u32(__p0, __p1) __extension__ ({ \
28684  uint32x4x3_t __s1 = __p1; \
28685  uint32x4x3_t __rev1; \
28686  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28687  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28688  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28689  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
28690 })
28691 #endif
28692 
28693 #ifdef __LITTLE_ENDIAN__
28694 #define vst3q_u16(__p0, __p1) __extension__ ({ \
28695  uint16x8x3_t __s1 = __p1; \
28696  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
28697 })
28698 #else
28699 #define vst3q_u16(__p0, __p1) __extension__ ({ \
28700  uint16x8x3_t __s1 = __p1; \
28701  uint16x8x3_t __rev1; \
28702  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28703  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28704  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28705  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
28706 })
28707 #endif
28708 
28709 #ifdef __LITTLE_ENDIAN__
28710 #define vst3q_s8(__p0, __p1) __extension__ ({ \
28711  int8x16x3_t __s1 = __p1; \
28712  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
28713 })
28714 #else
28715 #define vst3q_s8(__p0, __p1) __extension__ ({ \
28716  int8x16x3_t __s1 = __p1; \
28717  int8x16x3_t __rev1; \
28718  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28719  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28720  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
28721  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
28722 })
28723 #endif
28724 
28725 #ifdef __LITTLE_ENDIAN__
28726 #define vst3q_f32(__p0, __p1) __extension__ ({ \
28727  float32x4x3_t __s1 = __p1; \
28728  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 41); \
28729 })
28730 #else
28731 #define vst3q_f32(__p0, __p1) __extension__ ({ \
28732  float32x4x3_t __s1 = __p1; \
28733  float32x4x3_t __rev1; \
28734  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28735  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28736  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28737  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 41); \
28738 })
28739 #endif
28740 
28741 #ifdef __LITTLE_ENDIAN__
28742 #define vst3q_s32(__p0, __p1) __extension__ ({ \
28743  int32x4x3_t __s1 = __p1; \
28744  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 34); \
28745 })
28746 #else
28747 #define vst3q_s32(__p0, __p1) __extension__ ({ \
28748  int32x4x3_t __s1 = __p1; \
28749  int32x4x3_t __rev1; \
28750  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28751  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28752  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28753  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 34); \
28754 })
28755 #endif
28756 
28757 #ifdef __LITTLE_ENDIAN__
28758 #define vst3q_s16(__p0, __p1) __extension__ ({ \
28759  int16x8x3_t __s1 = __p1; \
28760  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 33); \
28761 })
28762 #else
28763 #define vst3q_s16(__p0, __p1) __extension__ ({ \
28764  int16x8x3_t __s1 = __p1; \
28765  int16x8x3_t __rev1; \
28766  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28767  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28768  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28769  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 33); \
28770 })
28771 #endif
28772 
28773 #ifdef __LITTLE_ENDIAN__
28774 #define vst3_u8(__p0, __p1) __extension__ ({ \
28775  uint8x8x3_t __s1 = __p1; \
28776  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
28777 })
28778 #else
28779 #define vst3_u8(__p0, __p1) __extension__ ({ \
28780  uint8x8x3_t __s1 = __p1; \
28781  uint8x8x3_t __rev1; \
28782  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28783  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28784  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28785  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
28786 })
28787 #endif
28788 
28789 #ifdef __LITTLE_ENDIAN__
28790 #define vst3_u32(__p0, __p1) __extension__ ({ \
28791  uint32x2x3_t __s1 = __p1; \
28792  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
28793 })
28794 #else
28795 #define vst3_u32(__p0, __p1) __extension__ ({ \
28796  uint32x2x3_t __s1 = __p1; \
28797  uint32x2x3_t __rev1; \
28798  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28799  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28800  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28801  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
28802 })
28803 #endif
28804 
28805 #define vst3_u64(__p0, __p1) __extension__ ({ \
28806  uint64x1x3_t __s1 = __p1; \
28807  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
28808 })
28809 #ifdef __LITTLE_ENDIAN__
28810 #define vst3_u16(__p0, __p1) __extension__ ({ \
28811  uint16x4x3_t __s1 = __p1; \
28812  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
28813 })
28814 #else
28815 #define vst3_u16(__p0, __p1) __extension__ ({ \
28816  uint16x4x3_t __s1 = __p1; \
28817  uint16x4x3_t __rev1; \
28818  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28819  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28820  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28821  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
28822 })
28823 #endif
28824 
28825 #ifdef __LITTLE_ENDIAN__
28826 #define vst3_s8(__p0, __p1) __extension__ ({ \
28827  int8x8x3_t __s1 = __p1; \
28828  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
28829 })
28830 #else
28831 #define vst3_s8(__p0, __p1) __extension__ ({ \
28832  int8x8x3_t __s1 = __p1; \
28833  int8x8x3_t __rev1; \
28834  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28835  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28836  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28837  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
28838 })
28839 #endif
28840 
28841 #ifdef __LITTLE_ENDIAN__
28842 #define vst3_f32(__p0, __p1) __extension__ ({ \
28843  float32x2x3_t __s1 = __p1; \
28844  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 9); \
28845 })
28846 #else
28847 #define vst3_f32(__p0, __p1) __extension__ ({ \
28848  float32x2x3_t __s1 = __p1; \
28849  float32x2x3_t __rev1; \
28850  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28851  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28852  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28853  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 9); \
28854 })
28855 #endif
28856 
28857 #ifdef __LITTLE_ENDIAN__
28858 #define vst3_s32(__p0, __p1) __extension__ ({ \
28859  int32x2x3_t __s1 = __p1; \
28860  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 2); \
28861 })
28862 #else
28863 #define vst3_s32(__p0, __p1) __extension__ ({ \
28864  int32x2x3_t __s1 = __p1; \
28865  int32x2x3_t __rev1; \
28866  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
28867  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
28868  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
28869  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 2); \
28870 })
28871 #endif
28872 
28873 #define vst3_s64(__p0, __p1) __extension__ ({ \
28874  int64x1x3_t __s1 = __p1; \
28875  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 3); \
28876 })
28877 #ifdef __LITTLE_ENDIAN__
28878 #define vst3_s16(__p0, __p1) __extension__ ({ \
28879  int16x4x3_t __s1 = __p1; \
28880  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 1); \
28881 })
28882 #else
28883 #define vst3_s16(__p0, __p1) __extension__ ({ \
28884  int16x4x3_t __s1 = __p1; \
28885  int16x4x3_t __rev1; \
28886  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28887  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28888  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28889  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 1); \
28890 })
28891 #endif
28892 
28893 #ifdef __LITTLE_ENDIAN__
28894 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28895  poly8x8x3_t __s1 = __p1; \
28896  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
28897 })
28898 #else
28899 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
28900  poly8x8x3_t __s1 = __p1; \
28901  poly8x8x3_t __rev1; \
28902  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28903  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28904  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28905  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
28906 })
28907 #endif
28908 
28909 #ifdef __LITTLE_ENDIAN__
28910 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28911  poly16x4x3_t __s1 = __p1; \
28912  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
28913 })
28914 #else
28915 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28916  poly16x4x3_t __s1 = __p1; \
28917  poly16x4x3_t __rev1; \
28918  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28919  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28920  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28921  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
28922 })
28923 #endif
28924 
28925 #ifdef __LITTLE_ENDIAN__
28926 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28927  poly16x8x3_t __s1 = __p1; \
28928  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
28929 })
28930 #else
28931 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
28932  poly16x8x3_t __s1 = __p1; \
28933  poly16x8x3_t __rev1; \
28934  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28935  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28936  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28937  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
28938 })
28939 #endif
28940 
28941 #ifdef __LITTLE_ENDIAN__
28942 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28943  uint32x4x3_t __s1 = __p1; \
28944  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
28945 })
28946 #else
28947 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
28948  uint32x4x3_t __s1 = __p1; \
28949  uint32x4x3_t __rev1; \
28950  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28951  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28952  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28953  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
28954 })
28955 #endif
28956 
28957 #ifdef __LITTLE_ENDIAN__
28958 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28959  uint16x8x3_t __s1 = __p1; \
28960  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
28961 })
28962 #else
28963 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
28964  uint16x8x3_t __s1 = __p1; \
28965  uint16x8x3_t __rev1; \
28966  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
28967  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
28968  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
28969  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
28970 })
28971 #endif
28972 
28973 #ifdef __LITTLE_ENDIAN__
28974 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28975  float32x4x3_t __s1 = __p1; \
28976  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 41); \
28977 })
28978 #else
28979 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
28980  float32x4x3_t __s1 = __p1; \
28981  float32x4x3_t __rev1; \
28982  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28983  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
28984  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
28985  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 41); \
28986 })
28987 #endif
28988 
28989 #ifdef __LITTLE_ENDIAN__
28990 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28991  int32x4x3_t __s1 = __p1; \
28992  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 34); \
28993 })
28994 #else
28995 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
28996  int32x4x3_t __s1 = __p1; \
28997  int32x4x3_t __rev1; \
28998  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
28999  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29000  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29001  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 34); \
29002 })
29003 #endif
29004 
29005 #ifdef __LITTLE_ENDIAN__
29006 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29007  int16x8x3_t __s1 = __p1; \
29008  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 33); \
29009 })
29010 #else
29011 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29012  int16x8x3_t __s1 = __p1; \
29013  int16x8x3_t __rev1; \
29014  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29015  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29016  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29017  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 33); \
29018 })
29019 #endif
29020 
29021 #ifdef __LITTLE_ENDIAN__
29022 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29023  uint8x8x3_t __s1 = __p1; \
29024  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
29025 })
29026 #else
29027 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29028  uint8x8x3_t __s1 = __p1; \
29029  uint8x8x3_t __rev1; \
29030  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29031  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29032  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29033  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
29034 })
29035 #endif
29036 
29037 #ifdef __LITTLE_ENDIAN__
29038 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29039  uint32x2x3_t __s1 = __p1; \
29040  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
29041 })
29042 #else
29043 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29044  uint32x2x3_t __s1 = __p1; \
29045  uint32x2x3_t __rev1; \
29046  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29047  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29048  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29049  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
29050 })
29051 #endif
29052 
29053 #ifdef __LITTLE_ENDIAN__
29054 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29055  uint16x4x3_t __s1 = __p1; \
29056  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
29057 })
29058 #else
29059 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29060  uint16x4x3_t __s1 = __p1; \
29061  uint16x4x3_t __rev1; \
29062  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29063  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29064  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29065  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
29066 })
29067 #endif
29068 
29069 #ifdef __LITTLE_ENDIAN__
29070 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29071  int8x8x3_t __s1 = __p1; \
29072  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
29073 })
29074 #else
29075 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29076  int8x8x3_t __s1 = __p1; \
29077  int8x8x3_t __rev1; \
29078  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29079  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29080  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29081  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
29082 })
29083 #endif
29084 
29085 #ifdef __LITTLE_ENDIAN__
29086 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29087  float32x2x3_t __s1 = __p1; \
29088  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 9); \
29089 })
29090 #else
29091 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29092  float32x2x3_t __s1 = __p1; \
29093  float32x2x3_t __rev1; \
29094  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29095  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29096  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29097  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 9); \
29098 })
29099 #endif
29100 
29101 #ifdef __LITTLE_ENDIAN__
29102 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29103  int32x2x3_t __s1 = __p1; \
29104  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 2); \
29105 })
29106 #else
29107 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29108  int32x2x3_t __s1 = __p1; \
29109  int32x2x3_t __rev1; \
29110  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29111  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29112  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29113  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 2); \
29114 })
29115 #endif
29116 
29117 #ifdef __LITTLE_ENDIAN__
29118 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29119  int16x4x3_t __s1 = __p1; \
29120  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 1); \
29121 })
29122 #else
29123 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29124  int16x4x3_t __s1 = __p1; \
29125  int16x4x3_t __rev1; \
29126  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29127  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29128  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29129  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 1); \
29130 })
29131 #endif
29132 
29133 #ifdef __LITTLE_ENDIAN__
29134 #define vst4_p8(__p0, __p1) __extension__ ({ \
29135  poly8x8x4_t __s1 = __p1; \
29136  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
29137 })
29138 #else
29139 #define vst4_p8(__p0, __p1) __extension__ ({ \
29140  poly8x8x4_t __s1 = __p1; \
29141  poly8x8x4_t __rev1; \
29142  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29143  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29144  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29145  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29146  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
29147 })
29148 #endif
29149 
29150 #ifdef __LITTLE_ENDIAN__
29151 #define vst4_p16(__p0, __p1) __extension__ ({ \
29152  poly16x4x4_t __s1 = __p1; \
29153  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
29154 })
29155 #else
29156 #define vst4_p16(__p0, __p1) __extension__ ({ \
29157  poly16x4x4_t __s1 = __p1; \
29158  poly16x4x4_t __rev1; \
29159  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29160  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29161  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29162  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29163  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
29164 })
29165 #endif
29166 
29167 #ifdef __LITTLE_ENDIAN__
29168 #define vst4q_p8(__p0, __p1) __extension__ ({ \
29169  poly8x16x4_t __s1 = __p1; \
29170  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
29171 })
29172 #else
29173 #define vst4q_p8(__p0, __p1) __extension__ ({ \
29174  poly8x16x4_t __s1 = __p1; \
29175  poly8x16x4_t __rev1; \
29176  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29177  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29178  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29179  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29180  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
29181 })
29182 #endif
29183 
29184 #ifdef __LITTLE_ENDIAN__
29185 #define vst4q_p16(__p0, __p1) __extension__ ({ \
29186  poly16x8x4_t __s1 = __p1; \
29187  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
29188 })
29189 #else
29190 #define vst4q_p16(__p0, __p1) __extension__ ({ \
29191  poly16x8x4_t __s1 = __p1; \
29192  poly16x8x4_t __rev1; \
29193  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29194  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29195  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29196  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29197  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
29198 })
29199 #endif
29200 
29201 #ifdef __LITTLE_ENDIAN__
29202 #define vst4q_u8(__p0, __p1) __extension__ ({ \
29203  uint8x16x4_t __s1 = __p1; \
29204  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
29205 })
29206 #else
29207 #define vst4q_u8(__p0, __p1) __extension__ ({ \
29208  uint8x16x4_t __s1 = __p1; \
29209  uint8x16x4_t __rev1; \
29210  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29211  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29212  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29213  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29214  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
29215 })
29216 #endif
29217 
29218 #ifdef __LITTLE_ENDIAN__
29219 #define vst4q_u32(__p0, __p1) __extension__ ({ \
29220  uint32x4x4_t __s1 = __p1; \
29221  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
29222 })
29223 #else
29224 #define vst4q_u32(__p0, __p1) __extension__ ({ \
29225  uint32x4x4_t __s1 = __p1; \
29226  uint32x4x4_t __rev1; \
29227  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29228  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29229  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29230  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29231  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
29232 })
29233 #endif
29234 
29235 #ifdef __LITTLE_ENDIAN__
29236 #define vst4q_u16(__p0, __p1) __extension__ ({ \
29237  uint16x8x4_t __s1 = __p1; \
29238  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
29239 })
29240 #else
29241 #define vst4q_u16(__p0, __p1) __extension__ ({ \
29242  uint16x8x4_t __s1 = __p1; \
29243  uint16x8x4_t __rev1; \
29244  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29245  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29246  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29247  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29248  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
29249 })
29250 #endif
29251 
29252 #ifdef __LITTLE_ENDIAN__
29253 #define vst4q_s8(__p0, __p1) __extension__ ({ \
29254  int8x16x4_t __s1 = __p1; \
29255  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
29256 })
29257 #else
29258 #define vst4q_s8(__p0, __p1) __extension__ ({ \
29259  int8x16x4_t __s1 = __p1; \
29260  int8x16x4_t __rev1; \
29261  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29262  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29263  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29264  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
29265  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
29266 })
29267 #endif
29268 
29269 #ifdef __LITTLE_ENDIAN__
29270 #define vst4q_f32(__p0, __p1) __extension__ ({ \
29271  float32x4x4_t __s1 = __p1; \
29272  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 41); \
29273 })
29274 #else
29275 #define vst4q_f32(__p0, __p1) __extension__ ({ \
29276  float32x4x4_t __s1 = __p1; \
29277  float32x4x4_t __rev1; \
29278  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29279  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29280  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29281  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29282  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 41); \
29283 })
29284 #endif
29285 
29286 #ifdef __LITTLE_ENDIAN__
29287 #define vst4q_s32(__p0, __p1) __extension__ ({ \
29288  int32x4x4_t __s1 = __p1; \
29289  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 34); \
29290 })
29291 #else
29292 #define vst4q_s32(__p0, __p1) __extension__ ({ \
29293  int32x4x4_t __s1 = __p1; \
29294  int32x4x4_t __rev1; \
29295  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29296  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29297  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29298  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29299  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 34); \
29300 })
29301 #endif
29302 
29303 #ifdef __LITTLE_ENDIAN__
29304 #define vst4q_s16(__p0, __p1) __extension__ ({ \
29305  int16x8x4_t __s1 = __p1; \
29306  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 33); \
29307 })
29308 #else
29309 #define vst4q_s16(__p0, __p1) __extension__ ({ \
29310  int16x8x4_t __s1 = __p1; \
29311  int16x8x4_t __rev1; \
29312  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29313  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29314  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29315  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29316  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 33); \
29317 })
29318 #endif
29319 
29320 #ifdef __LITTLE_ENDIAN__
29321 #define vst4_u8(__p0, __p1) __extension__ ({ \
29322  uint8x8x4_t __s1 = __p1; \
29323  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
29324 })
29325 #else
29326 #define vst4_u8(__p0, __p1) __extension__ ({ \
29327  uint8x8x4_t __s1 = __p1; \
29328  uint8x8x4_t __rev1; \
29329  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29330  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29331  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29332  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29333  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
29334 })
29335 #endif
29336 
29337 #ifdef __LITTLE_ENDIAN__
29338 #define vst4_u32(__p0, __p1) __extension__ ({ \
29339  uint32x2x4_t __s1 = __p1; \
29340  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
29341 })
29342 #else
29343 #define vst4_u32(__p0, __p1) __extension__ ({ \
29344  uint32x2x4_t __s1 = __p1; \
29345  uint32x2x4_t __rev1; \
29346  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29347  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29348  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29349  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29350  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
29351 })
29352 #endif
29353 
29354 #define vst4_u64(__p0, __p1) __extension__ ({ \
29355  uint64x1x4_t __s1 = __p1; \
29356  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
29357 })
29358 #ifdef __LITTLE_ENDIAN__
29359 #define vst4_u16(__p0, __p1) __extension__ ({ \
29360  uint16x4x4_t __s1 = __p1; \
29361  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
29362 })
29363 #else
29364 #define vst4_u16(__p0, __p1) __extension__ ({ \
29365  uint16x4x4_t __s1 = __p1; \
29366  uint16x4x4_t __rev1; \
29367  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29368  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29369  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29370  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29371  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
29372 })
29373 #endif
29374 
29375 #ifdef __LITTLE_ENDIAN__
29376 #define vst4_s8(__p0, __p1) __extension__ ({ \
29377  int8x8x4_t __s1 = __p1; \
29378  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
29379 })
29380 #else
29381 #define vst4_s8(__p0, __p1) __extension__ ({ \
29382  int8x8x4_t __s1 = __p1; \
29383  int8x8x4_t __rev1; \
29384  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29385  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29386  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29387  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29388  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
29389 })
29390 #endif
29391 
29392 #ifdef __LITTLE_ENDIAN__
29393 #define vst4_f32(__p0, __p1) __extension__ ({ \
29394  float32x2x4_t __s1 = __p1; \
29395  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 9); \
29396 })
29397 #else
29398 #define vst4_f32(__p0, __p1) __extension__ ({ \
29399  float32x2x4_t __s1 = __p1; \
29400  float32x2x4_t __rev1; \
29401  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29402  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29403  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29404  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29405  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 9); \
29406 })
29407 #endif
29408 
29409 #ifdef __LITTLE_ENDIAN__
29410 #define vst4_s32(__p0, __p1) __extension__ ({ \
29411  int32x2x4_t __s1 = __p1; \
29412  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 2); \
29413 })
29414 #else
29415 #define vst4_s32(__p0, __p1) __extension__ ({ \
29416  int32x2x4_t __s1 = __p1; \
29417  int32x2x4_t __rev1; \
29418  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29419  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29420  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29421  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29422  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 2); \
29423 })
29424 #endif
29425 
29426 #define vst4_s64(__p0, __p1) __extension__ ({ \
29427  int64x1x4_t __s1 = __p1; \
29428  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 3); \
29429 })
29430 #ifdef __LITTLE_ENDIAN__
29431 #define vst4_s16(__p0, __p1) __extension__ ({ \
29432  int16x4x4_t __s1 = __p1; \
29433  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 1); \
29434 })
29435 #else
29436 #define vst4_s16(__p0, __p1) __extension__ ({ \
29437  int16x4x4_t __s1 = __p1; \
29438  int16x4x4_t __rev1; \
29439  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29440  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29441  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29442  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29443  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 1); \
29444 })
29445 #endif
29446 
29447 #ifdef __LITTLE_ENDIAN__
29448 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
29449  poly8x8x4_t __s1 = __p1; \
29450  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
29451 })
29452 #else
29453 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
29454  poly8x8x4_t __s1 = __p1; \
29455  poly8x8x4_t __rev1; \
29456  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29457  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29458  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29459  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29460  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
29461 })
29462 #endif
29463 
29464 #ifdef __LITTLE_ENDIAN__
29465 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29466  poly16x4x4_t __s1 = __p1; \
29467  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
29468 })
29469 #else
29470 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29471  poly16x4x4_t __s1 = __p1; \
29472  poly16x4x4_t __rev1; \
29473  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29474  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29475  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29476  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29477  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
29478 })
29479 #endif
29480 
29481 #ifdef __LITTLE_ENDIAN__
29482 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29483  poly16x8x4_t __s1 = __p1; \
29484  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
29485 })
29486 #else
29487 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
29488  poly16x8x4_t __s1 = __p1; \
29489  poly16x8x4_t __rev1; \
29490  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29491  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29492  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29493  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29494  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
29495 })
29496 #endif
29497 
29498 #ifdef __LITTLE_ENDIAN__
29499 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29500  uint32x4x4_t __s1 = __p1; \
29501  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
29502 })
29503 #else
29504 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29505  uint32x4x4_t __s1 = __p1; \
29506  uint32x4x4_t __rev1; \
29507  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29508  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29509  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29510  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29511  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
29512 })
29513 #endif
29514 
29515 #ifdef __LITTLE_ENDIAN__
29516 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29517  uint16x8x4_t __s1 = __p1; \
29518  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
29519 })
29520 #else
29521 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29522  uint16x8x4_t __s1 = __p1; \
29523  uint16x8x4_t __rev1; \
29524  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29525  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29526  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29527  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29528  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
29529 })
29530 #endif
29531 
29532 #ifdef __LITTLE_ENDIAN__
29533 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29534  float32x4x4_t __s1 = __p1; \
29535  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 41); \
29536 })
29537 #else
29538 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29539  float32x4x4_t __s1 = __p1; \
29540  float32x4x4_t __rev1; \
29541  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29542  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29543  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29544  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29545  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 41); \
29546 })
29547 #endif
29548 
29549 #ifdef __LITTLE_ENDIAN__
29550 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29551  int32x4x4_t __s1 = __p1; \
29552  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 34); \
29553 })
29554 #else
29555 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29556  int32x4x4_t __s1 = __p1; \
29557  int32x4x4_t __rev1; \
29558  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29559  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29560  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29561  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29562  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 34); \
29563 })
29564 #endif
29565 
29566 #ifdef __LITTLE_ENDIAN__
29567 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29568  int16x8x4_t __s1 = __p1; \
29569  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 33); \
29570 })
29571 #else
29572 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29573  int16x8x4_t __s1 = __p1; \
29574  int16x8x4_t __rev1; \
29575  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29576  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29577  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29578  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29579  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 33); \
29580 })
29581 #endif
29582 
29583 #ifdef __LITTLE_ENDIAN__
29584 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29585  uint8x8x4_t __s1 = __p1; \
29586  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
29587 })
29588 #else
29589 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
29590  uint8x8x4_t __s1 = __p1; \
29591  uint8x8x4_t __rev1; \
29592  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29593  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29594  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29595  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29596  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
29597 })
29598 #endif
29599 
29600 #ifdef __LITTLE_ENDIAN__
29601 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29602  uint32x2x4_t __s1 = __p1; \
29603  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
29604 })
29605 #else
29606 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
29607  uint32x2x4_t __s1 = __p1; \
29608  uint32x2x4_t __rev1; \
29609  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29610  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29611  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29612  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29613  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
29614 })
29615 #endif
29616 
29617 #ifdef __LITTLE_ENDIAN__
29618 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29619  uint16x4x4_t __s1 = __p1; \
29620  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
29621 })
29622 #else
29623 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
29624  uint16x4x4_t __s1 = __p1; \
29625  uint16x4x4_t __rev1; \
29626  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29627  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29628  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29629  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29630  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
29631 })
29632 #endif
29633 
29634 #ifdef __LITTLE_ENDIAN__
29635 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29636  int8x8x4_t __s1 = __p1; \
29637  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
29638 })
29639 #else
29640 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
29641  int8x8x4_t __s1 = __p1; \
29642  int8x8x4_t __rev1; \
29643  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
29644  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
29645  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
29646  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
29647  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
29648 })
29649 #endif
29650 
29651 #ifdef __LITTLE_ENDIAN__
29652 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29653  float32x2x4_t __s1 = __p1; \
29654  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 9); \
29655 })
29656 #else
29657 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
29658  float32x2x4_t __s1 = __p1; \
29659  float32x2x4_t __rev1; \
29660  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29661  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29662  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29663  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29664  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 9); \
29665 })
29666 #endif
29667 
29668 #ifdef __LITTLE_ENDIAN__
29669 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29670  int32x2x4_t __s1 = __p1; \
29671  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 2); \
29672 })
29673 #else
29674 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
29675  int32x2x4_t __s1 = __p1; \
29676  int32x2x4_t __rev1; \
29677  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
29678  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
29679  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
29680  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
29681  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 2); \
29682 })
29683 #endif
29684 
29685 #ifdef __LITTLE_ENDIAN__
29686 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29687  int16x4x4_t __s1 = __p1; \
29688  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 1); \
29689 })
29690 #else
29691 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
29692  int16x4x4_t __s1 = __p1; \
29693  int16x4x4_t __rev1; \
29694  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
29695  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
29696  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
29697  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
29698  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 1); \
29699 })
29700 #endif
29701 
29702 #ifdef __LITTLE_ENDIAN__
29703 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29704  uint8x16_t __ret;
29705  __ret = __p0 - __p1;
29706  return __ret;
29707 }
29708 #else
29709 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29710  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29711  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29712  uint8x16_t __ret;
29713  __ret = __rev0 - __rev1;
29714  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29715  return __ret;
29716 }
29717 #endif
29718 
29719 #ifdef __LITTLE_ENDIAN__
29720 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29721  uint32x4_t __ret;
29722  __ret = __p0 - __p1;
29723  return __ret;
29724 }
29725 #else
29726 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29727  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29728  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29729  uint32x4_t __ret;
29730  __ret = __rev0 - __rev1;
29731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29732  return __ret;
29733 }
29734 #endif
29735 
29736 #ifdef __LITTLE_ENDIAN__
29737 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
29738  uint64x2_t __ret;
29739  __ret = __p0 - __p1;
29740  return __ret;
29741 }
29742 #else
29743 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
29744  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29745  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29746  uint64x2_t __ret;
29747  __ret = __rev0 - __rev1;
29748  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29749  return __ret;
29750 }
29751 #endif
29752 
29753 #ifdef __LITTLE_ENDIAN__
29754 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29755  uint16x8_t __ret;
29756  __ret = __p0 - __p1;
29757  return __ret;
29758 }
29759 #else
29760 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29761  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29762  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29763  uint16x8_t __ret;
29764  __ret = __rev0 - __rev1;
29765  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29766  return __ret;
29767 }
29768 #endif
29769 
29770 #ifdef __LITTLE_ENDIAN__
29771 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
29772  int8x16_t __ret;
29773  __ret = __p0 - __p1;
29774  return __ret;
29775 }
29776 #else
29777 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
29778  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29779  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29780  int8x16_t __ret;
29781  __ret = __rev0 - __rev1;
29782  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29783  return __ret;
29784 }
29785 #endif
29786 
29787 #ifdef __LITTLE_ENDIAN__
29788 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
29789  float32x4_t __ret;
29790  __ret = __p0 - __p1;
29791  return __ret;
29792 }
29793 #else
29794 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
29795  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29796  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29797  float32x4_t __ret;
29798  __ret = __rev0 - __rev1;
29799  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29800  return __ret;
29801 }
29802 #endif
29803 
29804 #ifdef __LITTLE_ENDIAN__
29805 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
29806  int32x4_t __ret;
29807  __ret = __p0 - __p1;
29808  return __ret;
29809 }
29810 #else
29811 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
29812  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29813  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29814  int32x4_t __ret;
29815  __ret = __rev0 - __rev1;
29816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29817  return __ret;
29818 }
29819 #endif
29820 
29821 #ifdef __LITTLE_ENDIAN__
29822 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
29823  int64x2_t __ret;
29824  __ret = __p0 - __p1;
29825  return __ret;
29826 }
29827 #else
29828 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
29829  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29830  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29831  int64x2_t __ret;
29832  __ret = __rev0 - __rev1;
29833  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29834  return __ret;
29835 }
29836 #endif
29837 
29838 #ifdef __LITTLE_ENDIAN__
29839 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
29840  int16x8_t __ret;
29841  __ret = __p0 - __p1;
29842  return __ret;
29843 }
29844 #else
29845 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
29846  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29847  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29848  int16x8_t __ret;
29849  __ret = __rev0 - __rev1;
29850  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29851  return __ret;
29852 }
29853 #endif
29854 
29855 #ifdef __LITTLE_ENDIAN__
29856 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
29857  uint8x8_t __ret;
29858  __ret = __p0 - __p1;
29859  return __ret;
29860 }
29861 #else
29862 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
29863  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29864  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29865  uint8x8_t __ret;
29866  __ret = __rev0 - __rev1;
29867  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29868  return __ret;
29869 }
29870 #endif
29871 
29872 #ifdef __LITTLE_ENDIAN__
29873 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
29874  uint32x2_t __ret;
29875  __ret = __p0 - __p1;
29876  return __ret;
29877 }
29878 #else
29879 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
29880  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29881  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29882  uint32x2_t __ret;
29883  __ret = __rev0 - __rev1;
29884  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29885  return __ret;
29886 }
29887 #endif
29888 
29889 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
29890  uint64x1_t __ret;
29891  __ret = __p0 - __p1;
29892  return __ret;
29893 }
29894 #ifdef __LITTLE_ENDIAN__
29895 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
29896  uint16x4_t __ret;
29897  __ret = __p0 - __p1;
29898  return __ret;
29899 }
29900 #else
29901 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
29902  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29903  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29904  uint16x4_t __ret;
29905  __ret = __rev0 - __rev1;
29906  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29907  return __ret;
29908 }
29909 #endif
29910 
29911 #ifdef __LITTLE_ENDIAN__
29912 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
29913  int8x8_t __ret;
29914  __ret = __p0 - __p1;
29915  return __ret;
29916 }
29917 #else
29918 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
29919  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29920  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29921  int8x8_t __ret;
29922  __ret = __rev0 - __rev1;
29923  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29924  return __ret;
29925 }
29926 #endif
29927 
29928 #ifdef __LITTLE_ENDIAN__
29929 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
29930  float32x2_t __ret;
29931  __ret = __p0 - __p1;
29932  return __ret;
29933 }
29934 #else
29935 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
29936  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29937  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29938  float32x2_t __ret;
29939  __ret = __rev0 - __rev1;
29940  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29941  return __ret;
29942 }
29943 #endif
29944 
29945 #ifdef __LITTLE_ENDIAN__
29946 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
29947  int32x2_t __ret;
29948  __ret = __p0 - __p1;
29949  return __ret;
29950 }
29951 #else
29952 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
29953  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29954  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29955  int32x2_t __ret;
29956  __ret = __rev0 - __rev1;
29957  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29958  return __ret;
29959 }
29960 #endif
29961 
29962 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
29963  int64x1_t __ret;
29964  __ret = __p0 - __p1;
29965  return __ret;
29966 }
29967 #ifdef __LITTLE_ENDIAN__
29968 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
29969  int16x4_t __ret;
29970  __ret = __p0 - __p1;
29971  return __ret;
29972 }
29973 #else
29974 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
29975  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29976  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29977  int16x4_t __ret;
29978  __ret = __rev0 - __rev1;
29979  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29980  return __ret;
29981 }
29982 #endif
29983 
29984 #ifdef __LITTLE_ENDIAN__
29985 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
29986  uint16x4_t __ret;
29987  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
29988  return __ret;
29989 }
29990 #else
29991 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
29992  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29993  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29994  uint16x4_t __ret;
29995  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
29996  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29997  return __ret;
29998 }
29999 __ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
30000  uint16x4_t __ret;
30001  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
30002  return __ret;
30003 }
30004 #endif
30005 
30006 #ifdef __LITTLE_ENDIAN__
30007 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
30008  uint32x2_t __ret;
30009  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
30010  return __ret;
30011 }
30012 #else
30013 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
30014  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30015  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30016  uint32x2_t __ret;
30017  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
30018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30019  return __ret;
30020 }
30021 __ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
30022  uint32x2_t __ret;
30023  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
30024  return __ret;
30025 }
30026 #endif
30027 
30028 #ifdef __LITTLE_ENDIAN__
30029 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
30030  uint8x8_t __ret;
30031  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
30032  return __ret;
30033 }
30034 #else
30035 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
30036  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30037  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30038  uint8x8_t __ret;
30039  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
30040  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30041  return __ret;
30042 }
30043 __ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
30044  uint8x8_t __ret;
30045  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
30046  return __ret;
30047 }
30048 #endif
30049 
30050 #ifdef __LITTLE_ENDIAN__
30051 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
30052  int16x4_t __ret;
30053  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
30054  return __ret;
30055 }
30056 #else
30057 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
30058  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30059  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30060  int16x4_t __ret;
30061  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
30062  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30063  return __ret;
30064 }
30065 __ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
30066  int16x4_t __ret;
30067  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
30068  return __ret;
30069 }
30070 #endif
30071 
30072 #ifdef __LITTLE_ENDIAN__
30073 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
30074  int32x2_t __ret;
30075  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
30076  return __ret;
30077 }
30078 #else
30079 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
30080  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30081  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30082  int32x2_t __ret;
30083  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
30084  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30085  return __ret;
30086 }
30087 __ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
30088  int32x2_t __ret;
30089  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
30090  return __ret;
30091 }
30092 #endif
30093 
30094 #ifdef __LITTLE_ENDIAN__
30095 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
30096  int8x8_t __ret;
30097  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
30098  return __ret;
30099 }
30100 #else
30101 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
30102  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30103  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30104  int8x8_t __ret;
30105  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
30106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30107  return __ret;
30108 }
30109 __ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
30110  int8x8_t __ret;
30111  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
30112  return __ret;
30113 }
30114 #endif
30115 
30116 #ifdef __LITTLE_ENDIAN__
30117 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
30118  uint16x8_t __ret;
30119  __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
30120  return __ret;
30121 }
30122 #else
30123 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
30124  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30125  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30126  uint16x8_t __ret;
30127  __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
30128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30129  return __ret;
30130 }
30131 #endif
30132 
30133 #ifdef __LITTLE_ENDIAN__
30134 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
30135  uint64x2_t __ret;
30136  __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
30137  return __ret;
30138 }
30139 #else
30140 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
30141  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30142  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30143  uint64x2_t __ret;
30144  __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
30145  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30146  return __ret;
30147 }
30148 #endif
30149 
30150 #ifdef __LITTLE_ENDIAN__
30151 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
30152  uint32x4_t __ret;
30153  __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
30154  return __ret;
30155 }
30156 #else
30157 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
30158  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30159  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30160  uint32x4_t __ret;
30161  __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
30162  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30163  return __ret;
30164 }
30165 #endif
30166 
30167 #ifdef __LITTLE_ENDIAN__
30168 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
30169  int16x8_t __ret;
30170  __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
30171  return __ret;
30172 }
30173 #else
30174 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
30175  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30176  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30177  int16x8_t __ret;
30178  __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
30179  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30180  return __ret;
30181 }
30182 #endif
30183 
30184 #ifdef __LITTLE_ENDIAN__
30185 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
30186  int64x2_t __ret;
30187  __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
30188  return __ret;
30189 }
30190 #else
30191 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
30192  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30193  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30194  int64x2_t __ret;
30195  __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
30196  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30197  return __ret;
30198 }
30199 #endif
30200 
30201 #ifdef __LITTLE_ENDIAN__
30202 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
30203  int32x4_t __ret;
30204  __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
30205  return __ret;
30206 }
30207 #else
30208 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
30209  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30210  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30211  int32x4_t __ret;
30212  __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
30213  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30214  return __ret;
30215 }
30216 #endif
30217 
30218 #ifdef __LITTLE_ENDIAN__
30219 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
30220  uint16x8_t __ret;
30221  __ret = __p0 - vmovl_u8(__p1);
30222  return __ret;
30223 }
30224 #else
30225 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
30226  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30227  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30228  uint16x8_t __ret;
30229  __ret = __rev0 - __noswap_vmovl_u8(__rev1);
30230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30231  return __ret;
30232 }
30233 #endif
30234 
30235 #ifdef __LITTLE_ENDIAN__
30236 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
30237  uint64x2_t __ret;
30238  __ret = __p0 - vmovl_u32(__p1);
30239  return __ret;
30240 }
30241 #else
30242 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
30243  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30244  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30245  uint64x2_t __ret;
30246  __ret = __rev0 - __noswap_vmovl_u32(__rev1);
30247  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30248  return __ret;
30249 }
30250 #endif
30251 
30252 #ifdef __LITTLE_ENDIAN__
30253 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
30254  uint32x4_t __ret;
30255  __ret = __p0 - vmovl_u16(__p1);
30256  return __ret;
30257 }
30258 #else
30259 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
30260  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30261  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30262  uint32x4_t __ret;
30263  __ret = __rev0 - __noswap_vmovl_u16(__rev1);
30264  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30265  return __ret;
30266 }
30267 #endif
30268 
30269 #ifdef __LITTLE_ENDIAN__
30270 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
30271  int16x8_t __ret;
30272  __ret = __p0 - vmovl_s8(__p1);
30273  return __ret;
30274 }
30275 #else
30276 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
30277  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30278  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30279  int16x8_t __ret;
30280  __ret = __rev0 - __noswap_vmovl_s8(__rev1);
30281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30282  return __ret;
30283 }
30284 #endif
30285 
30286 #ifdef __LITTLE_ENDIAN__
30287 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
30288  int64x2_t __ret;
30289  __ret = __p0 - vmovl_s32(__p1);
30290  return __ret;
30291 }
30292 #else
30293 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
30294  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
30295  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
30296  int64x2_t __ret;
30297  __ret = __rev0 - __noswap_vmovl_s32(__rev1);
30298  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
30299  return __ret;
30300 }
30301 #endif
30302 
30303 #ifdef __LITTLE_ENDIAN__
30304 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
30305  int32x4_t __ret;
30306  __ret = __p0 - vmovl_s16(__p1);
30307  return __ret;
30308 }
30309 #else
30310 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
30311  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30312  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30313  int32x4_t __ret;
30314  __ret = __rev0 - __noswap_vmovl_s16(__rev1);
30315  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
30316  return __ret;
30317 }
30318 #endif
30319 
30320 #ifdef __LITTLE_ENDIAN__
30321 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
30322  poly8x8_t __ret;
30323  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
30324  return __ret;
30325 }
30326 #else
30327 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
30328  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30329  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30330  poly8x8_t __ret;
30331  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
30332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30333  return __ret;
30334 }
30335 #endif
30336 
30337 #ifdef __LITTLE_ENDIAN__
30338 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
30339  uint8x8_t __ret;
30340  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
30341  return __ret;
30342 }
30343 #else
30344 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
30345  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30346  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30347  uint8x8_t __ret;
30348  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
30349  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30350  return __ret;
30351 }
30352 #endif
30353 
30354 #ifdef __LITTLE_ENDIAN__
30355 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
30356  int8x8_t __ret;
30357  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
30358  return __ret;
30359 }
30360 #else
30361 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
30362  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30363  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30364  int8x8_t __ret;
30365  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
30366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30367  return __ret;
30368 }
30369 #endif
30370 
30371 #ifdef __LITTLE_ENDIAN__
30372 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
30373  poly8x8_t __ret;
30374  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
30375  return __ret;
30376 }
30377 #else
30378 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
30379  poly8x8x2_t __rev0;
30380  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30381  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30382  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30383  poly8x8_t __ret;
30384  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
30385  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30386  return __ret;
30387 }
30388 #endif
30389 
30390 #ifdef __LITTLE_ENDIAN__
30391 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
30392  uint8x8_t __ret;
30393  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
30394  return __ret;
30395 }
30396 #else
30397 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
30398  uint8x8x2_t __rev0;
30399  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30400  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30401  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30402  uint8x8_t __ret;
30403  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
30404  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30405  return __ret;
30406 }
30407 #endif
30408 
30409 #ifdef __LITTLE_ENDIAN__
30410 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
30411  int8x8_t __ret;
30412  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
30413  return __ret;
30414 }
30415 #else
30416 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
30417  int8x8x2_t __rev0;
30418  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30419  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30420  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30421  int8x8_t __ret;
30422  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
30423  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30424  return __ret;
30425 }
30426 #endif
30427 
30428 #ifdef __LITTLE_ENDIAN__
30429 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
30430  poly8x8_t __ret;
30431  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
30432  return __ret;
30433 }
30434 #else
30435 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
30436  poly8x8x3_t __rev0;
30437  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30438  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30439  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30440  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30441  poly8x8_t __ret;
30442  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
30443  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30444  return __ret;
30445 }
30446 #endif
30447 
30448 #ifdef __LITTLE_ENDIAN__
30449 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
30450  uint8x8_t __ret;
30451  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
30452  return __ret;
30453 }
30454 #else
30455 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
30456  uint8x8x3_t __rev0;
30457  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30458  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30459  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30460  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30461  uint8x8_t __ret;
30462  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
30463  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30464  return __ret;
30465 }
30466 #endif
30467 
30468 #ifdef __LITTLE_ENDIAN__
30469 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
30470  int8x8_t __ret;
30471  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
30472  return __ret;
30473 }
30474 #else
30475 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
30476  int8x8x3_t __rev0;
30477  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30478  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30479  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30480  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30481  int8x8_t __ret;
30482  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
30483  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30484  return __ret;
30485 }
30486 #endif
30487 
30488 #ifdef __LITTLE_ENDIAN__
30489 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
30490  poly8x8_t __ret;
30491  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
30492  return __ret;
30493 }
30494 #else
30495 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
30496  poly8x8x4_t __rev0;
30497  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30498  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30499  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30500  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30501  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30502  poly8x8_t __ret;
30503  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
30504  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30505  return __ret;
30506 }
30507 #endif
30508 
30509 #ifdef __LITTLE_ENDIAN__
30510 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
30511  uint8x8_t __ret;
30512  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
30513  return __ret;
30514 }
30515 #else
30516 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
30517  uint8x8x4_t __rev0;
30518  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30519  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30520  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30521  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30522  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30523  uint8x8_t __ret;
30524  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
30525  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30526  return __ret;
30527 }
30528 #endif
30529 
30530 #ifdef __LITTLE_ENDIAN__
30531 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
30532  int8x8_t __ret;
30533  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
30534  return __ret;
30535 }
30536 #else
30537 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
30538  int8x8x4_t __rev0;
30539  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30540  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30541  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30542  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30543  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30544  int8x8_t __ret;
30545  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
30546  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30547  return __ret;
30548 }
30549 #endif
30550 
30551 #ifdef __LITTLE_ENDIAN__
30552 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
30553  poly8x8_t __ret;
30554  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
30555  return __ret;
30556 }
30557 #else
30558 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
30559  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30560  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30561  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30562  poly8x8_t __ret;
30563  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
30564  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30565  return __ret;
30566 }
30567 #endif
30568 
30569 #ifdef __LITTLE_ENDIAN__
30570 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
30571  uint8x8_t __ret;
30572  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
30573  return __ret;
30574 }
30575 #else
30576 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
30577  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30578  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30579  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30580  uint8x8_t __ret;
30581  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
30582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30583  return __ret;
30584 }
30585 #endif
30586 
30587 #ifdef __LITTLE_ENDIAN__
30588 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
30589  int8x8_t __ret;
30590  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
30591  return __ret;
30592 }
30593 #else
30594 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
30595  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30596  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30597  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30598  int8x8_t __ret;
30599  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
30600  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30601  return __ret;
30602 }
30603 #endif
30604 
30605 #ifdef __LITTLE_ENDIAN__
30606 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
30607  poly8x8_t __ret;
30608  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
30609  return __ret;
30610 }
30611 #else
30612 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
30613  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30614  poly8x8x2_t __rev1;
30615  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30616  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30617  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30618  poly8x8_t __ret;
30619  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
30620  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30621  return __ret;
30622 }
30623 #endif
30624 
30625 #ifdef __LITTLE_ENDIAN__
30626 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
30627  uint8x8_t __ret;
30628  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
30629  return __ret;
30630 }
30631 #else
30632 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
30633  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30634  uint8x8x2_t __rev1;
30635  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30636  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30637  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30638  uint8x8_t __ret;
30639  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
30640  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30641  return __ret;
30642 }
30643 #endif
30644 
30645 #ifdef __LITTLE_ENDIAN__
30646 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
30647  int8x8_t __ret;
30648  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
30649  return __ret;
30650 }
30651 #else
30652 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
30653  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30654  int8x8x2_t __rev1;
30655  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30656  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30657  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30658  int8x8_t __ret;
30659  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
30660  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30661  return __ret;
30662 }
30663 #endif
30664 
30665 #ifdef __LITTLE_ENDIAN__
30666 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
30667  poly8x8_t __ret;
30668  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
30669  return __ret;
30670 }
30671 #else
30672 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
30673  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30674  poly8x8x3_t __rev1;
30675  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30676  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30677  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30678  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30679  poly8x8_t __ret;
30680  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
30681  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30682  return __ret;
30683 }
30684 #endif
30685 
30686 #ifdef __LITTLE_ENDIAN__
30687 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
30688  uint8x8_t __ret;
30689  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
30690  return __ret;
30691 }
30692 #else
30693 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
30694  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30695  uint8x8x3_t __rev1;
30696  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30697  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30698  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30699  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30700  uint8x8_t __ret;
30701  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
30702  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30703  return __ret;
30704 }
30705 #endif
30706 
30707 #ifdef __LITTLE_ENDIAN__
30708 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
30709  int8x8_t __ret;
30710  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
30711  return __ret;
30712 }
30713 #else
30714 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
30715  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30716  int8x8x3_t __rev1;
30717  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30718  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30719  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30720  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30721  int8x8_t __ret;
30722  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
30723  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30724  return __ret;
30725 }
30726 #endif
30727 
30728 #ifdef __LITTLE_ENDIAN__
30729 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
30730  poly8x8_t __ret;
30731  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
30732  return __ret;
30733 }
30734 #else
30735 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
30736  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30737  poly8x8x4_t __rev1;
30738  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30739  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30740  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30741  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30742  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30743  poly8x8_t __ret;
30744  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
30745  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30746  return __ret;
30747 }
30748 #endif
30749 
30750 #ifdef __LITTLE_ENDIAN__
30751 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
30752  uint8x8_t __ret;
30753  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
30754  return __ret;
30755 }
30756 #else
30757 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
30758  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30759  uint8x8x4_t __rev1;
30760  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30761  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30762  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30763  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30764  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30765  uint8x8_t __ret;
30766  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
30767  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30768  return __ret;
30769 }
30770 #endif
30771 
30772 #ifdef __LITTLE_ENDIAN__
30773 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
30774  int8x8_t __ret;
30775  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
30776  return __ret;
30777 }
30778 #else
30779 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
30780  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30781  int8x8x4_t __rev1;
30782  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30783  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30784  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
30785  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
30786  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
30787  int8x8_t __ret;
30788  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
30789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
30790  return __ret;
30791 }
30792 #endif
30793 
30794 #ifdef __LITTLE_ENDIAN__
30795 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
30796  poly8x8x2_t __ret;
30797  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
30798  return __ret;
30799 }
30800 #else
30801 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
30802  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30803  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30804  poly8x8x2_t __ret;
30805  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
30806 
30807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30809  return __ret;
30810 }
30811 #endif
30812 
30813 #ifdef __LITTLE_ENDIAN__
30814 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
30815  poly16x4x2_t __ret;
30816  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
30817  return __ret;
30818 }
30819 #else
30820 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
30821  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30822  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30823  poly16x4x2_t __ret;
30824  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
30825 
30826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30828  return __ret;
30829 }
30830 #endif
30831 
30832 #ifdef __LITTLE_ENDIAN__
30833 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
30834  poly8x16x2_t __ret;
30835  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
30836  return __ret;
30837 }
30838 #else
30839 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
30840  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30841  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30842  poly8x16x2_t __ret;
30843  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
30844 
30845  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30846  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30847  return __ret;
30848 }
30849 #endif
30850 
30851 #ifdef __LITTLE_ENDIAN__
30852 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
30853  poly16x8x2_t __ret;
30854  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
30855  return __ret;
30856 }
30857 #else
30858 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
30859  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30860  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30861  poly16x8x2_t __ret;
30862  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
30863 
30864  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30865  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30866  return __ret;
30867 }
30868 #endif
30869 
30870 #ifdef __LITTLE_ENDIAN__
30871 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
30872  uint8x16x2_t __ret;
30873  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
30874  return __ret;
30875 }
30876 #else
30877 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
30878  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30879  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30880  uint8x16x2_t __ret;
30881  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
30882 
30883  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30884  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30885  return __ret;
30886 }
30887 #endif
30888 
30889 #ifdef __LITTLE_ENDIAN__
30890 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
30891  uint32x4x2_t __ret;
30892  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
30893  return __ret;
30894 }
30895 #else
30896 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
30897  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30898  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30899  uint32x4x2_t __ret;
30900  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
30901 
30902  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30903  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30904  return __ret;
30905 }
30906 #endif
30907 
30908 #ifdef __LITTLE_ENDIAN__
30909 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
30910  uint16x8x2_t __ret;
30911  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
30912  return __ret;
30913 }
30914 #else
30915 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
30916  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30917  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30918  uint16x8x2_t __ret;
30919  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
30920 
30921  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30922  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30923  return __ret;
30924 }
30925 #endif
30926 
30927 #ifdef __LITTLE_ENDIAN__
30928 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
30929  int8x16x2_t __ret;
30930  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
30931  return __ret;
30932 }
30933 #else
30934 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
30935  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30936  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30937  int8x16x2_t __ret;
30938  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
30939 
30940  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30941  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
30942  return __ret;
30943 }
30944 #endif
30945 
30946 #ifdef __LITTLE_ENDIAN__
30947 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
30948  float32x4x2_t __ret;
30949  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
30950  return __ret;
30951 }
30952 #else
30953 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
30954  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30955  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30956  float32x4x2_t __ret;
30957  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
30958 
30959  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30960  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30961  return __ret;
30962 }
30963 #endif
30964 
30965 #ifdef __LITTLE_ENDIAN__
30966 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
30967  int32x4x2_t __ret;
30968  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
30969  return __ret;
30970 }
30971 #else
30972 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
30973  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
30974  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
30975  int32x4x2_t __ret;
30976  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
30977 
30978  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
30979  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
30980  return __ret;
30981 }
30982 #endif
30983 
30984 #ifdef __LITTLE_ENDIAN__
30985 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
30986  int16x8x2_t __ret;
30987  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
30988  return __ret;
30989 }
30990 #else
30991 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
30992  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
30993  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
30994  int16x8x2_t __ret;
30995  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
30996 
30997  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
30998  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
30999  return __ret;
31000 }
31001 #endif
31002 
31003 #ifdef __LITTLE_ENDIAN__
31004 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
31005  uint8x8x2_t __ret;
31006  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
31007  return __ret;
31008 }
31009 #else
31010 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
31011  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31012  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31013  uint8x8x2_t __ret;
31014  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31015 
31016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31018  return __ret;
31019 }
31020 #endif
31021 
31022 #ifdef __LITTLE_ENDIAN__
31023 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
31024  uint32x2x2_t __ret;
31025  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
31026  return __ret;
31027 }
31028 #else
31029 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
31030  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31031  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31032  uint32x2x2_t __ret;
31033  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31034 
31035  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31036  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31037  return __ret;
31038 }
31039 #endif
31040 
31041 #ifdef __LITTLE_ENDIAN__
31042 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
31043  uint16x4x2_t __ret;
31044  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
31045  return __ret;
31046 }
31047 #else
31048 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
31049  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31050  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31051  uint16x4x2_t __ret;
31052  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31053 
31054  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31055  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31056  return __ret;
31057 }
31058 #endif
31059 
31060 #ifdef __LITTLE_ENDIAN__
31061 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
31062  int8x8x2_t __ret;
31063  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
31064  return __ret;
31065 }
31066 #else
31067 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
31068  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31069  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31070  int8x8x2_t __ret;
31071  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
31072 
31073  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31074  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31075  return __ret;
31076 }
31077 #endif
31078 
31079 #ifdef __LITTLE_ENDIAN__
31080 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
31081  float32x2x2_t __ret;
31082  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
31083  return __ret;
31084 }
31085 #else
31086 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
31087  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31088  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31089  float32x2x2_t __ret;
31090  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
31091 
31092  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31093  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31094  return __ret;
31095 }
31096 #endif
31097 
31098 #ifdef __LITTLE_ENDIAN__
31099 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
31100  int32x2x2_t __ret;
31101  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
31102  return __ret;
31103 }
31104 #else
31105 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
31106  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31107  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31108  int32x2x2_t __ret;
31109  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
31110 
31111  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31112  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31113  return __ret;
31114 }
31115 #endif
31116 
31117 #ifdef __LITTLE_ENDIAN__
31118 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
31119  int16x4x2_t __ret;
31120  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
31121  return __ret;
31122 }
31123 #else
31124 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
31125  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31126  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31127  int16x4x2_t __ret;
31128  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
31129 
31130  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31131  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31132  return __ret;
31133 }
31134 #endif
31135 
31136 #ifdef __LITTLE_ENDIAN__
31137 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
31138  uint8x8_t __ret;
31139  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
31140  return __ret;
31141 }
31142 #else
31143 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
31144  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31145  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31146  uint8x8_t __ret;
31147  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31148  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31149  return __ret;
31150 }
31151 #endif
31152 
31153 #ifdef __LITTLE_ENDIAN__
31154 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
31155  uint16x4_t __ret;
31156  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31157  return __ret;
31158 }
31159 #else
31160 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
31161  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31162  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31163  uint16x4_t __ret;
31164  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31165  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31166  return __ret;
31167 }
31168 #endif
31169 
31170 #ifdef __LITTLE_ENDIAN__
31171 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31172  uint8x16_t __ret;
31173  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31174  return __ret;
31175 }
31176 #else
31177 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31178  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31179  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31180  uint8x16_t __ret;
31181  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31182  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31183  return __ret;
31184 }
31185 #endif
31186 
31187 #ifdef __LITTLE_ENDIAN__
31188 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31189  uint16x8_t __ret;
31190  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31191  return __ret;
31192 }
31193 #else
31194 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31195  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31196  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31197  uint16x8_t __ret;
31198  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31199  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31200  return __ret;
31201 }
31202 #endif
31203 
31204 #ifdef __LITTLE_ENDIAN__
31205 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31206  uint8x16_t __ret;
31207  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31208  return __ret;
31209 }
31210 #else
31211 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31212  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31213  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31214  uint8x16_t __ret;
31215  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31216  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31217  return __ret;
31218 }
31219 #endif
31220 
31221 #ifdef __LITTLE_ENDIAN__
31222 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31223  uint32x4_t __ret;
31224  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
31225  return __ret;
31226 }
31227 #else
31228 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31229  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31230  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31231  uint32x4_t __ret;
31232  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31233  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31234  return __ret;
31235 }
31236 #endif
31237 
31238 #ifdef __LITTLE_ENDIAN__
31239 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31240  uint16x8_t __ret;
31241  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31242  return __ret;
31243 }
31244 #else
31245 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31246  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31247  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31248  uint16x8_t __ret;
31249  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31250  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31251  return __ret;
31252 }
31253 #endif
31254 
31255 #ifdef __LITTLE_ENDIAN__
31256 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
31257  uint8x16_t __ret;
31258  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
31259  return __ret;
31260 }
31261 #else
31262 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
31263  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31264  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31265  uint8x16_t __ret;
31266  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31267  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31268  return __ret;
31269 }
31270 #endif
31271 
31272 #ifdef __LITTLE_ENDIAN__
31273 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
31274  uint32x4_t __ret;
31275  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
31276  return __ret;
31277 }
31278 #else
31279 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
31280  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31281  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31282  uint32x4_t __ret;
31283  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31284  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31285  return __ret;
31286 }
31287 #endif
31288 
31289 #ifdef __LITTLE_ENDIAN__
31290 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
31291  uint16x8_t __ret;
31292  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
31293  return __ret;
31294 }
31295 #else
31296 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
31297  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31298  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31299  uint16x8_t __ret;
31300  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31302  return __ret;
31303 }
31304 #endif
31305 
31306 #ifdef __LITTLE_ENDIAN__
31307 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
31308  uint8x8_t __ret;
31309  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
31310  return __ret;
31311 }
31312 #else
31313 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
31314  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31315  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31316  uint8x8_t __ret;
31317  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31319  return __ret;
31320 }
31321 #endif
31322 
31323 #ifdef __LITTLE_ENDIAN__
31324 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
31325  uint32x2_t __ret;
31326  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
31327  return __ret;
31328 }
31329 #else
31330 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
31331  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31332  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31333  uint32x2_t __ret;
31334  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31335  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
31336  return __ret;
31337 }
31338 #endif
31339 
31340 #ifdef __LITTLE_ENDIAN__
31341 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
31342  uint16x4_t __ret;
31343  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31344  return __ret;
31345 }
31346 #else
31347 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
31348  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31349  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31350  uint16x4_t __ret;
31351  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31352  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31353  return __ret;
31354 }
31355 #endif
31356 
31357 #ifdef __LITTLE_ENDIAN__
31358 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
31359  uint8x8_t __ret;
31360  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
31361  return __ret;
31362 }
31363 #else
31364 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
31365  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31366  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31367  uint8x8_t __ret;
31368  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31369  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
31370  return __ret;
31371 }
31372 #endif
31373 
31374 #ifdef __LITTLE_ENDIAN__
31375 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
31376  uint32x2_t __ret;
31377  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
31378  return __ret;
31379 }
31380 #else
31381 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
31382  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31383  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31384  uint32x2_t __ret;
31385  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31386  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
31387  return __ret;
31388 }
31389 #endif
31390 
31391 #ifdef __LITTLE_ENDIAN__
31392 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
31393  uint16x4_t __ret;
31394  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
31395  return __ret;
31396 }
31397 #else
31398 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
31399  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31400  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31401  uint16x4_t __ret;
31402  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31403  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
31404  return __ret;
31405 }
31406 #endif
31407 
31408 #ifdef __LITTLE_ENDIAN__
31409 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
31410  poly8x8x2_t __ret;
31411  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
31412  return __ret;
31413 }
31414 #else
31415 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
31416  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31417  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31418  poly8x8x2_t __ret;
31419  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
31420 
31421  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31422  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31423  return __ret;
31424 }
31425 #endif
31426 
31427 #ifdef __LITTLE_ENDIAN__
31428 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
31429  poly16x4x2_t __ret;
31430  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
31431  return __ret;
31432 }
31433 #else
31434 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
31435  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31436  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31437  poly16x4x2_t __ret;
31438  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
31439 
31440  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31441  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31442  return __ret;
31443 }
31444 #endif
31445 
31446 #ifdef __LITTLE_ENDIAN__
31447 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31448  poly8x16x2_t __ret;
31449  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
31450  return __ret;
31451 }
31452 #else
31453 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31454  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31455  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31456  poly8x16x2_t __ret;
31457  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
31458 
31459  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31460  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31461  return __ret;
31462 }
31463 #endif
31464 
31465 #ifdef __LITTLE_ENDIAN__
31466 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31467  poly16x8x2_t __ret;
31468  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
31469  return __ret;
31470 }
31471 #else
31472 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31473  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31474  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31475  poly16x8x2_t __ret;
31476  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
31477 
31478  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31479  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31480  return __ret;
31481 }
31482 #endif
31483 
31484 #ifdef __LITTLE_ENDIAN__
31485 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31486  uint8x16x2_t __ret;
31487  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
31488  return __ret;
31489 }
31490 #else
31491 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31492  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31493  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31494  uint8x16x2_t __ret;
31495  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31496 
31497  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31498  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31499  return __ret;
31500 }
31501 #endif
31502 
31503 #ifdef __LITTLE_ENDIAN__
31504 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31505  uint32x4x2_t __ret;
31506  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
31507  return __ret;
31508 }
31509 #else
31510 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31511  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31512  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31513  uint32x4x2_t __ret;
31514  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31515 
31516  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31517  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31518  return __ret;
31519 }
31520 #endif
31521 
31522 #ifdef __LITTLE_ENDIAN__
31523 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31524  uint16x8x2_t __ret;
31525  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
31526  return __ret;
31527 }
31528 #else
31529 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31530  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31531  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31532  uint16x8x2_t __ret;
31533  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31534 
31535  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31536  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31537  return __ret;
31538 }
31539 #endif
31540 
31541 #ifdef __LITTLE_ENDIAN__
31542 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
31543  int8x16x2_t __ret;
31544  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
31545  return __ret;
31546 }
31547 #else
31548 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
31549  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31550  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31551  int8x16x2_t __ret;
31552  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
31553 
31554  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31555  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31556  return __ret;
31557 }
31558 #endif
31559 
31560 #ifdef __LITTLE_ENDIAN__
31561 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
31562  float32x4x2_t __ret;
31563  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
31564  return __ret;
31565 }
31566 #else
31567 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
31568  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31569  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31570  float32x4x2_t __ret;
31571  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
31572 
31573  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31574  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31575  return __ret;
31576 }
31577 #endif
31578 
31579 #ifdef __LITTLE_ENDIAN__
31580 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
31581  int32x4x2_t __ret;
31582  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
31583  return __ret;
31584 }
31585 #else
31586 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
31587  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31588  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31589  int32x4x2_t __ret;
31590  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
31591 
31592  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31593  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31594  return __ret;
31595 }
31596 #endif
31597 
31598 #ifdef __LITTLE_ENDIAN__
31599 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
31600  int16x8x2_t __ret;
31601  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
31602  return __ret;
31603 }
31604 #else
31605 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
31606  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31607  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31608  int16x8x2_t __ret;
31609  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
31610 
31611  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31612  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31613  return __ret;
31614 }
31615 #endif
31616 
31617 #ifdef __LITTLE_ENDIAN__
31618 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
31619  uint8x8x2_t __ret;
31620  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
31621  return __ret;
31622 }
31623 #else
31624 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
31625  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31626  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31627  uint8x8x2_t __ret;
31628  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31629 
31630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31632  return __ret;
31633 }
31634 #endif
31635 
31636 #ifdef __LITTLE_ENDIAN__
31637 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
31638  uint32x2x2_t __ret;
31639  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
31640  return __ret;
31641 }
31642 #else
31643 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
31644  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31645  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31646  uint32x2x2_t __ret;
31647  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31648 
31649  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31650  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31651  return __ret;
31652 }
31653 #endif
31654 
31655 #ifdef __LITTLE_ENDIAN__
31656 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
31657  uint16x4x2_t __ret;
31658  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
31659  return __ret;
31660 }
31661 #else
31662 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
31663  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31664  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31665  uint16x4x2_t __ret;
31666  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
31667 
31668  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31669  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31670  return __ret;
31671 }
31672 #endif
31673 
31674 #ifdef __LITTLE_ENDIAN__
31675 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
31676  int8x8x2_t __ret;
31677  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
31678  return __ret;
31679 }
31680 #else
31681 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
31682  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31683  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31684  int8x8x2_t __ret;
31685  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
31686 
31687  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31688  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31689  return __ret;
31690 }
31691 #endif
31692 
31693 #ifdef __LITTLE_ENDIAN__
31694 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
31695  float32x2x2_t __ret;
31696  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
31697  return __ret;
31698 }
31699 #else
31700 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
31701  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31702  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31703  float32x2x2_t __ret;
31704  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
31705 
31706  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31707  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31708  return __ret;
31709 }
31710 #endif
31711 
31712 #ifdef __LITTLE_ENDIAN__
31713 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
31714  int32x2x2_t __ret;
31715  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
31716  return __ret;
31717 }
31718 #else
31719 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
31720  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31721  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31722  int32x2x2_t __ret;
31723  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
31724 
31725  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31726  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31727  return __ret;
31728 }
31729 #endif
31730 
31731 #ifdef __LITTLE_ENDIAN__
31732 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
31733  int16x4x2_t __ret;
31734  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
31735  return __ret;
31736 }
31737 #else
31738 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
31739  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31740  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31741  int16x4x2_t __ret;
31742  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
31743 
31744  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31745  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31746  return __ret;
31747 }
31748 #endif
31749 
31750 #ifdef __LITTLE_ENDIAN__
31751 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
31752  poly8x8x2_t __ret;
31753  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
31754  return __ret;
31755 }
31756 #else
31757 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
31758  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31759  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31760  poly8x8x2_t __ret;
31761  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
31762 
31763  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31764  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31765  return __ret;
31766 }
31767 #endif
31768 
31769 #ifdef __LITTLE_ENDIAN__
31770 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
31771  poly16x4x2_t __ret;
31772  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
31773  return __ret;
31774 }
31775 #else
31776 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
31777  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31778  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31779  poly16x4x2_t __ret;
31780  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
31781 
31782  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31783  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31784  return __ret;
31785 }
31786 #endif
31787 
31788 #ifdef __LITTLE_ENDIAN__
31789 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31790  poly8x16x2_t __ret;
31791  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
31792  return __ret;
31793 }
31794 #else
31795 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
31796  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31797  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31798  poly8x16x2_t __ret;
31799  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
31800 
31801  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31802  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31803  return __ret;
31804 }
31805 #endif
31806 
31807 #ifdef __LITTLE_ENDIAN__
31808 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31809  poly16x8x2_t __ret;
31810  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
31811  return __ret;
31812 }
31813 #else
31814 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
31815  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31816  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31817  poly16x8x2_t __ret;
31818  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
31819 
31820  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31821  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31822  return __ret;
31823 }
31824 #endif
31825 
31826 #ifdef __LITTLE_ENDIAN__
31827 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31828  uint8x16x2_t __ret;
31829  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
31830  return __ret;
31831 }
31832 #else
31833 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
31834  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31835  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31836  uint8x16x2_t __ret;
31837  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
31838 
31839  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31840  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31841  return __ret;
31842 }
31843 #endif
31844 
31845 #ifdef __LITTLE_ENDIAN__
31846 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31847  uint32x4x2_t __ret;
31848  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
31849  return __ret;
31850 }
31851 #else
31852 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
31853  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31854  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31855  uint32x4x2_t __ret;
31856  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
31857 
31858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31860  return __ret;
31861 }
31862 #endif
31863 
31864 #ifdef __LITTLE_ENDIAN__
31865 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31866  uint16x8x2_t __ret;
31867  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
31868  return __ret;
31869 }
31870 #else
31871 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
31872  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31873  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31874  uint16x8x2_t __ret;
31875  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
31876 
31877  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31878  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31879  return __ret;
31880 }
31881 #endif
31882 
31883 #ifdef __LITTLE_ENDIAN__
31884 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
31885  int8x16x2_t __ret;
31886  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
31887  return __ret;
31888 }
31889 #else
31890 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
31891  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31892  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31893  int8x16x2_t __ret;
31894  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
31895 
31896  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31897  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
31898  return __ret;
31899 }
31900 #endif
31901 
31902 #ifdef __LITTLE_ENDIAN__
31903 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
31904  float32x4x2_t __ret;
31905  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
31906  return __ret;
31907 }
31908 #else
31909 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
31910  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31911  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31912  float32x4x2_t __ret;
31913  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
31914 
31915  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31916  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31917  return __ret;
31918 }
31919 #endif
31920 
31921 #ifdef __LITTLE_ENDIAN__
31922 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
31923  int32x4x2_t __ret;
31924  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
31925  return __ret;
31926 }
31927 #else
31928 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
31929  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
31930  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
31931  int32x4x2_t __ret;
31932  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
31933 
31934  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
31935  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
31936  return __ret;
31937 }
31938 #endif
31939 
31940 #ifdef __LITTLE_ENDIAN__
31941 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
31942  int16x8x2_t __ret;
31943  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
31944  return __ret;
31945 }
31946 #else
31947 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
31948  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31949  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31950  int16x8x2_t __ret;
31951  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
31952 
31953  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31954  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31955  return __ret;
31956 }
31957 #endif
31958 
31959 #ifdef __LITTLE_ENDIAN__
31960 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
31961  uint8x8x2_t __ret;
31962  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
31963  return __ret;
31964 }
31965 #else
31966 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
31967  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
31968  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
31969  uint8x8x2_t __ret;
31970  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
31971 
31972  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
31973  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
31974  return __ret;
31975 }
31976 #endif
31977 
31978 #ifdef __LITTLE_ENDIAN__
31979 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
31980  uint32x2x2_t __ret;
31981  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
31982  return __ret;
31983 }
31984 #else
31985 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
31986  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
31987  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
31988  uint32x2x2_t __ret;
31989  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
31990 
31991  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
31992  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
31993  return __ret;
31994 }
31995 #endif
31996 
31997 #ifdef __LITTLE_ENDIAN__
31998 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
31999  uint16x4x2_t __ret;
32000  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
32001  return __ret;
32002 }
32003 #else
32004 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
32005  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32006  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
32007  uint16x4x2_t __ret;
32008  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
32009 
32010  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
32011  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
32012  return __ret;
32013 }
32014 #endif
32015 
32016 #ifdef __LITTLE_ENDIAN__
32017 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
32018  int8x8x2_t __ret;
32019  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
32020  return __ret;
32021 }
32022 #else
32023 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
32024  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
32025  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
32026  int8x8x2_t __ret;
32027  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
32028 
32029  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
32030  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
32031  return __ret;
32032 }
32033 #endif
32034 
32035 #ifdef __LITTLE_ENDIAN__
32036 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
32037  float32x2x2_t __ret;
32038  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
32039  return __ret;
32040 }
32041 #else
32042 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
32043  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
32044  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
32045  float32x2x2_t __ret;
32046  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
32047 
32048  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
32049  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
32050  return __ret;
32051 }
32052 #endif
32053 
32054 #ifdef __LITTLE_ENDIAN__
32055 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
32056  int32x2x2_t __ret;
32057  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
32058  return __ret;
32059 }
32060 #else
32061 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
32062  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
32063  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
32064  int32x2x2_t __ret;
32065  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
32066 
32067  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
32068  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
32069  return __ret;
32070 }
32071 #endif
32072 
32073 #ifdef __LITTLE_ENDIAN__
32074 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
32075  int16x4x2_t __ret;
32076  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
32077  return __ret;
32078 }
32079 #else
32080 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
32081  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
32082  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
32083  int16x4x2_t __ret;
32084  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
32085 
32086  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
32087  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
32088  return __ret;
32089 }
32090 #endif
32091 
32092 #if !defined(__aarch64__)
32093 #ifdef __LITTLE_ENDIAN__
32094 #define vdupq_lane_f16(__p0_122, __p1_122) __extension__ ({ \
32095  float16x4_t __s0_122 = __p0_122; \
32096  float16x8_t __ret_122; \
32097  __ret_122 = splatq_lane_f16(__s0_122, __p1_122); \
32098  __ret_122; \
32099 })
32100 #else
32101 #define vdupq_lane_f16(__p0_123, __p1_123) __extension__ ({ \
32102  float16x4_t __s0_123 = __p0_123; \
32103  float16x4_t __rev0_123; __rev0_123 = __builtin_shufflevector(__s0_123, __s0_123, 3, 2, 1, 0); \
32104  float16x8_t __ret_123; \
32105  __ret_123 = __noswap_splatq_lane_f16(__rev0_123, __p1_123); \
32106  __ret_123 = __builtin_shufflevector(__ret_123, __ret_123, 7, 6, 5, 4, 3, 2, 1, 0); \
32107  __ret_123; \
32108 })
32109 #endif
32110 
32111 #ifdef __LITTLE_ENDIAN__
32112 #define vdup_lane_f16(__p0_124, __p1_124) __extension__ ({ \
32113  float16x4_t __s0_124 = __p0_124; \
32114  float16x4_t __ret_124; \
32115  __ret_124 = splat_lane_f16(__s0_124, __p1_124); \
32116  __ret_124; \
32117 })
32118 #else
32119 #define vdup_lane_f16(__p0_125, __p1_125) __extension__ ({ \
32120  float16x4_t __s0_125 = __p0_125; \
32121  float16x4_t __rev0_125; __rev0_125 = __builtin_shufflevector(__s0_125, __s0_125, 3, 2, 1, 0); \
32122  float16x4_t __ret_125; \
32123  __ret_125 = __noswap_splat_lane_f16(__rev0_125, __p1_125); \
32124  __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 3, 2, 1, 0); \
32125  __ret_125; \
32126 })
32127 #endif
32128 
32129 #ifdef __LITTLE_ENDIAN__
32130 #define vdupq_n_f16(__p0) __extension__ ({ \
32131  float16_t __s0 = __p0; \
32132  float16x8_t __ret; \
32133  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
32134  __ret; \
32135 })
32136 #else
32137 #define vdupq_n_f16(__p0) __extension__ ({ \
32138  float16_t __s0 = __p0; \
32139  float16x8_t __ret; \
32140  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
32141  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32142  __ret; \
32143 })
32144 #endif
32145 
32146 #ifdef __LITTLE_ENDIAN__
32147 #define vdup_n_f16(__p0) __extension__ ({ \
32148  float16_t __s0 = __p0; \
32149  float16x4_t __ret; \
32150  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
32151  __ret; \
32152 })
32153 #else
32154 #define vdup_n_f16(__p0) __extension__ ({ \
32155  float16_t __s0 = __p0; \
32156  float16x4_t __ret; \
32157  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
32158  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32159  __ret; \
32160 })
32161 #endif
32162 
32163 #ifdef __LITTLE_ENDIAN__
32164 #define vmovq_n_f16(__p0) __extension__ ({ \
32165  float16_t __s0 = __p0; \
32166  float16x8_t __ret; \
32167  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
32168  __ret; \
32169 })
32170 #else
32171 #define vmovq_n_f16(__p0) __extension__ ({ \
32172  float16_t __s0 = __p0; \
32173  float16x8_t __ret; \
32174  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
32175  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
32176  __ret; \
32177 })
32178 #endif
32179 
32180 #ifdef __LITTLE_ENDIAN__
32181 #define vmov_n_f16(__p0) __extension__ ({ \
32182  float16_t __s0 = __p0; \
32183  float16x4_t __ret; \
32184  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
32185  __ret; \
32186 })
32187 #else
32188 #define vmov_n_f16(__p0) __extension__ ({ \
32189  float16_t __s0 = __p0; \
32190  float16x4_t __ret; \
32191  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
32192  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
32193  __ret; \
32194 })
32195 #endif
32196 
32197 #ifdef __LITTLE_ENDIAN__
32198 #define vqdmulhq_lane_s32(__p0_126, __p1_126, __p2_126) __extension__ ({ \
32199  int32x4_t __s0_126 = __p0_126; \
32200  int32x2_t __s1_126 = __p1_126; \
32201  int32x4_t __ret_126; \
32202  __ret_126 = vqdmulhq_s32(__s0_126, splatq_lane_s32(__s1_126, __p2_126)); \
32203  __ret_126; \
32204 })
32205 #else
32206 #define vqdmulhq_lane_s32(__p0_127, __p1_127, __p2_127) __extension__ ({ \
32207  int32x4_t __s0_127 = __p0_127; \
32208  int32x2_t __s1_127 = __p1_127; \
32209  int32x4_t __rev0_127; __rev0_127 = __builtin_shufflevector(__s0_127, __s0_127, 3, 2, 1, 0); \
32210  int32x2_t __rev1_127; __rev1_127 = __builtin_shufflevector(__s1_127, __s1_127, 1, 0); \
32211  int32x4_t __ret_127; \
32212  __ret_127 = __noswap_vqdmulhq_s32(__rev0_127, __noswap_splatq_lane_s32(__rev1_127, __p2_127)); \
32213  __ret_127 = __builtin_shufflevector(__ret_127, __ret_127, 3, 2, 1, 0); \
32214  __ret_127; \
32215 })
32216 #endif
32217 
32218 #ifdef __LITTLE_ENDIAN__
32219 #define vqdmulhq_lane_s16(__p0_128, __p1_128, __p2_128) __extension__ ({ \
32220  int16x8_t __s0_128 = __p0_128; \
32221  int16x4_t __s1_128 = __p1_128; \
32222  int16x8_t __ret_128; \
32223  __ret_128 = vqdmulhq_s16(__s0_128, splatq_lane_s16(__s1_128, __p2_128)); \
32224  __ret_128; \
32225 })
32226 #else
32227 #define vqdmulhq_lane_s16(__p0_129, __p1_129, __p2_129) __extension__ ({ \
32228  int16x8_t __s0_129 = __p0_129; \
32229  int16x4_t __s1_129 = __p1_129; \
32230  int16x8_t __rev0_129; __rev0_129 = __builtin_shufflevector(__s0_129, __s0_129, 7, 6, 5, 4, 3, 2, 1, 0); \
32231  int16x4_t __rev1_129; __rev1_129 = __builtin_shufflevector(__s1_129, __s1_129, 3, 2, 1, 0); \
32232  int16x8_t __ret_129; \
32233  __ret_129 = __noswap_vqdmulhq_s16(__rev0_129, __noswap_splatq_lane_s16(__rev1_129, __p2_129)); \
32234  __ret_129 = __builtin_shufflevector(__ret_129, __ret_129, 7, 6, 5, 4, 3, 2, 1, 0); \
32235  __ret_129; \
32236 })
32237 #endif
32238 
32239 #ifdef __LITTLE_ENDIAN__
32240 #define vqdmulh_lane_s32(__p0_130, __p1_130, __p2_130) __extension__ ({ \
32241  int32x2_t __s0_130 = __p0_130; \
32242  int32x2_t __s1_130 = __p1_130; \
32243  int32x2_t __ret_130; \
32244  __ret_130 = vqdmulh_s32(__s0_130, splat_lane_s32(__s1_130, __p2_130)); \
32245  __ret_130; \
32246 })
32247 #else
32248 #define vqdmulh_lane_s32(__p0_131, __p1_131, __p2_131) __extension__ ({ \
32249  int32x2_t __s0_131 = __p0_131; \
32250  int32x2_t __s1_131 = __p1_131; \
32251  int32x2_t __rev0_131; __rev0_131 = __builtin_shufflevector(__s0_131, __s0_131, 1, 0); \
32252  int32x2_t __rev1_131; __rev1_131 = __builtin_shufflevector(__s1_131, __s1_131, 1, 0); \
32253  int32x2_t __ret_131; \
32254  __ret_131 = __noswap_vqdmulh_s32(__rev0_131, __noswap_splat_lane_s32(__rev1_131, __p2_131)); \
32255  __ret_131 = __builtin_shufflevector(__ret_131, __ret_131, 1, 0); \
32256  __ret_131; \
32257 })
32258 #endif
32259 
32260 #ifdef __LITTLE_ENDIAN__
32261 #define vqdmulh_lane_s16(__p0_132, __p1_132, __p2_132) __extension__ ({ \
32262  int16x4_t __s0_132 = __p0_132; \
32263  int16x4_t __s1_132 = __p1_132; \
32264  int16x4_t __ret_132; \
32265  __ret_132 = vqdmulh_s16(__s0_132, splat_lane_s16(__s1_132, __p2_132)); \
32266  __ret_132; \
32267 })
32268 #else
32269 #define vqdmulh_lane_s16(__p0_133, __p1_133, __p2_133) __extension__ ({ \
32270  int16x4_t __s0_133 = __p0_133; \
32271  int16x4_t __s1_133 = __p1_133; \
32272  int16x4_t __rev0_133; __rev0_133 = __builtin_shufflevector(__s0_133, __s0_133, 3, 2, 1, 0); \
32273  int16x4_t __rev1_133; __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 3, 2, 1, 0); \
32274  int16x4_t __ret_133; \
32275  __ret_133 = __noswap_vqdmulh_s16(__rev0_133, __noswap_splat_lane_s16(__rev1_133, __p2_133)); \
32276  __ret_133 = __builtin_shufflevector(__ret_133, __ret_133, 3, 2, 1, 0); \
32277  __ret_133; \
32278 })
32279 #endif
32280 
32281 #ifdef __LITTLE_ENDIAN__
32282 #define vqrdmulhq_lane_s32(__p0_134, __p1_134, __p2_134) __extension__ ({ \
32283  int32x4_t __s0_134 = __p0_134; \
32284  int32x2_t __s1_134 = __p1_134; \
32285  int32x4_t __ret_134; \
32286  __ret_134 = vqrdmulhq_s32(__s0_134, splatq_lane_s32(__s1_134, __p2_134)); \
32287  __ret_134; \
32288 })
32289 #else
32290 #define vqrdmulhq_lane_s32(__p0_135, __p1_135, __p2_135) __extension__ ({ \
32291  int32x4_t __s0_135 = __p0_135; \
32292  int32x2_t __s1_135 = __p1_135; \
32293  int32x4_t __rev0_135; __rev0_135 = __builtin_shufflevector(__s0_135, __s0_135, 3, 2, 1, 0); \
32294  int32x2_t __rev1_135; __rev1_135 = __builtin_shufflevector(__s1_135, __s1_135, 1, 0); \
32295  int32x4_t __ret_135; \
32296  __ret_135 = __noswap_vqrdmulhq_s32(__rev0_135, __noswap_splatq_lane_s32(__rev1_135, __p2_135)); \
32297  __ret_135 = __builtin_shufflevector(__ret_135, __ret_135, 3, 2, 1, 0); \
32298  __ret_135; \
32299 })
32300 #endif
32301 
32302 #ifdef __LITTLE_ENDIAN__
32303 #define vqrdmulhq_lane_s16(__p0_136, __p1_136, __p2_136) __extension__ ({ \
32304  int16x8_t __s0_136 = __p0_136; \
32305  int16x4_t __s1_136 = __p1_136; \
32306  int16x8_t __ret_136; \
32307  __ret_136 = vqrdmulhq_s16(__s0_136, splatq_lane_s16(__s1_136, __p2_136)); \
32308  __ret_136; \
32309 })
32310 #else
32311 #define vqrdmulhq_lane_s16(__p0_137, __p1_137, __p2_137) __extension__ ({ \
32312  int16x8_t __s0_137 = __p0_137; \
32313  int16x4_t __s1_137 = __p1_137; \
32314  int16x8_t __rev0_137; __rev0_137 = __builtin_shufflevector(__s0_137, __s0_137, 7, 6, 5, 4, 3, 2, 1, 0); \
32315  int16x4_t __rev1_137; __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 3, 2, 1, 0); \
32316  int16x8_t __ret_137; \
32317  __ret_137 = __noswap_vqrdmulhq_s16(__rev0_137, __noswap_splatq_lane_s16(__rev1_137, __p2_137)); \
32318  __ret_137 = __builtin_shufflevector(__ret_137, __ret_137, 7, 6, 5, 4, 3, 2, 1, 0); \
32319  __ret_137; \
32320 })
32321 #endif
32322 
32323 #ifdef __LITTLE_ENDIAN__
32324 #define vqrdmulh_lane_s32(__p0_138, __p1_138, __p2_138) __extension__ ({ \
32325  int32x2_t __s0_138 = __p0_138; \
32326  int32x2_t __s1_138 = __p1_138; \
32327  int32x2_t __ret_138; \
32328  __ret_138 = vqrdmulh_s32(__s0_138, splat_lane_s32(__s1_138, __p2_138)); \
32329  __ret_138; \
32330 })
32331 #else
32332 #define vqrdmulh_lane_s32(__p0_139, __p1_139, __p2_139) __extension__ ({ \
32333  int32x2_t __s0_139 = __p0_139; \
32334  int32x2_t __s1_139 = __p1_139; \
32335  int32x2_t __rev0_139; __rev0_139 = __builtin_shufflevector(__s0_139, __s0_139, 1, 0); \
32336  int32x2_t __rev1_139; __rev1_139 = __builtin_shufflevector(__s1_139, __s1_139, 1, 0); \
32337  int32x2_t __ret_139; \
32338  __ret_139 = __noswap_vqrdmulh_s32(__rev0_139, __noswap_splat_lane_s32(__rev1_139, __p2_139)); \
32339  __ret_139 = __builtin_shufflevector(__ret_139, __ret_139, 1, 0); \
32340  __ret_139; \
32341 })
32342 #endif
32343 
32344 #ifdef __LITTLE_ENDIAN__
32345 #define vqrdmulh_lane_s16(__p0_140, __p1_140, __p2_140) __extension__ ({ \
32346  int16x4_t __s0_140 = __p0_140; \
32347  int16x4_t __s1_140 = __p1_140; \
32348  int16x4_t __ret_140; \
32349  __ret_140 = vqrdmulh_s16(__s0_140, splat_lane_s16(__s1_140, __p2_140)); \
32350  __ret_140; \
32351 })
32352 #else
32353 #define vqrdmulh_lane_s16(__p0_141, __p1_141, __p2_141) __extension__ ({ \
32354  int16x4_t __s0_141 = __p0_141; \
32355  int16x4_t __s1_141 = __p1_141; \
32356  int16x4_t __rev0_141; __rev0_141 = __builtin_shufflevector(__s0_141, __s0_141, 3, 2, 1, 0); \
32357  int16x4_t __rev1_141; __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 3, 2, 1, 0); \
32358  int16x4_t __ret_141; \
32359  __ret_141 = __noswap_vqrdmulh_s16(__rev0_141, __noswap_splat_lane_s16(__rev1_141, __p2_141)); \
32360  __ret_141 = __builtin_shufflevector(__ret_141, __ret_141, 3, 2, 1, 0); \
32361  __ret_141; \
32362 })
32363 #endif
32364 
32365 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
32366  poly8x8_t __ret;
32367  __ret = (poly8x8_t)(__p0);
32368  return __ret;
32369 }
32370 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
32371  poly8x8_t __ret;
32372  __ret = (poly8x8_t)(__p0);
32373  return __ret;
32374 }
32375 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
32376  poly8x8_t __ret;
32377  __ret = (poly8x8_t)(__p0);
32378  return __ret;
32379 }
32380 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
32381  poly8x8_t __ret;
32382  __ret = (poly8x8_t)(__p0);
32383  return __ret;
32384 }
32385 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
32386  poly8x8_t __ret;
32387  __ret = (poly8x8_t)(__p0);
32388  return __ret;
32389 }
32390 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
32391  poly8x8_t __ret;
32392  __ret = (poly8x8_t)(__p0);
32393  return __ret;
32394 }
32395 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
32396  poly8x8_t __ret;
32397  __ret = (poly8x8_t)(__p0);
32398  return __ret;
32399 }
32400 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
32401  poly8x8_t __ret;
32402  __ret = (poly8x8_t)(__p0);
32403  return __ret;
32404 }
32405 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
32406  poly8x8_t __ret;
32407  __ret = (poly8x8_t)(__p0);
32408  return __ret;
32409 }
32410 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
32411  poly8x8_t __ret;
32412  __ret = (poly8x8_t)(__p0);
32413  return __ret;
32414 }
32415 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
32416  poly8x8_t __ret;
32417  __ret = (poly8x8_t)(__p0);
32418  return __ret;
32419 }
32420 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
32421  poly16x4_t __ret;
32422  __ret = (poly16x4_t)(__p0);
32423  return __ret;
32424 }
32425 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
32426  poly16x4_t __ret;
32427  __ret = (poly16x4_t)(__p0);
32428  return __ret;
32429 }
32430 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
32431  poly16x4_t __ret;
32432  __ret = (poly16x4_t)(__p0);
32433  return __ret;
32434 }
32435 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
32436  poly16x4_t __ret;
32437  __ret = (poly16x4_t)(__p0);
32438  return __ret;
32439 }
32440 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
32441  poly16x4_t __ret;
32442  __ret = (poly16x4_t)(__p0);
32443  return __ret;
32444 }
32445 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
32446  poly16x4_t __ret;
32447  __ret = (poly16x4_t)(__p0);
32448  return __ret;
32449 }
32450 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
32451  poly16x4_t __ret;
32452  __ret = (poly16x4_t)(__p0);
32453  return __ret;
32454 }
32455 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
32456  poly16x4_t __ret;
32457  __ret = (poly16x4_t)(__p0);
32458  return __ret;
32459 }
32460 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
32461  poly16x4_t __ret;
32462  __ret = (poly16x4_t)(__p0);
32463  return __ret;
32464 }
32465 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
32466  poly16x4_t __ret;
32467  __ret = (poly16x4_t)(__p0);
32468  return __ret;
32469 }
32470 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
32471  poly16x4_t __ret;
32472  __ret = (poly16x4_t)(__p0);
32473  return __ret;
32474 }
32475 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
32476  poly8x16_t __ret;
32477  __ret = (poly8x16_t)(__p0);
32478  return __ret;
32479 }
32480 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
32481  poly8x16_t __ret;
32482  __ret = (poly8x16_t)(__p0);
32483  return __ret;
32484 }
32485 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
32486  poly8x16_t __ret;
32487  __ret = (poly8x16_t)(__p0);
32488  return __ret;
32489 }
32490 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
32491  poly8x16_t __ret;
32492  __ret = (poly8x16_t)(__p0);
32493  return __ret;
32494 }
32495 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
32496  poly8x16_t __ret;
32497  __ret = (poly8x16_t)(__p0);
32498  return __ret;
32499 }
32500 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
32501  poly8x16_t __ret;
32502  __ret = (poly8x16_t)(__p0);
32503  return __ret;
32504 }
32505 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
32506  poly8x16_t __ret;
32507  __ret = (poly8x16_t)(__p0);
32508  return __ret;
32509 }
32510 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
32511  poly8x16_t __ret;
32512  __ret = (poly8x16_t)(__p0);
32513  return __ret;
32514 }
32515 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
32516  poly8x16_t __ret;
32517  __ret = (poly8x16_t)(__p0);
32518  return __ret;
32519 }
32520 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
32521  poly8x16_t __ret;
32522  __ret = (poly8x16_t)(__p0);
32523  return __ret;
32524 }
32525 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
32526  poly8x16_t __ret;
32527  __ret = (poly8x16_t)(__p0);
32528  return __ret;
32529 }
32530 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
32531  poly16x8_t __ret;
32532  __ret = (poly16x8_t)(__p0);
32533  return __ret;
32534 }
32535 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
32536  poly16x8_t __ret;
32537  __ret = (poly16x8_t)(__p0);
32538  return __ret;
32539 }
32540 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
32541  poly16x8_t __ret;
32542  __ret = (poly16x8_t)(__p0);
32543  return __ret;
32544 }
32545 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
32546  poly16x8_t __ret;
32547  __ret = (poly16x8_t)(__p0);
32548  return __ret;
32549 }
32550 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
32551  poly16x8_t __ret;
32552  __ret = (poly16x8_t)(__p0);
32553  return __ret;
32554 }
32555 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
32556  poly16x8_t __ret;
32557  __ret = (poly16x8_t)(__p0);
32558  return __ret;
32559 }
32560 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
32561  poly16x8_t __ret;
32562  __ret = (poly16x8_t)(__p0);
32563  return __ret;
32564 }
32565 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
32566  poly16x8_t __ret;
32567  __ret = (poly16x8_t)(__p0);
32568  return __ret;
32569 }
32570 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
32571  poly16x8_t __ret;
32572  __ret = (poly16x8_t)(__p0);
32573  return __ret;
32574 }
32575 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
32576  poly16x8_t __ret;
32577  __ret = (poly16x8_t)(__p0);
32578  return __ret;
32579 }
32580 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
32581  poly16x8_t __ret;
32582  __ret = (poly16x8_t)(__p0);
32583  return __ret;
32584 }
32585 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
32586  uint8x16_t __ret;
32587  __ret = (uint8x16_t)(__p0);
32588  return __ret;
32589 }
32590 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
32591  uint8x16_t __ret;
32592  __ret = (uint8x16_t)(__p0);
32593  return __ret;
32594 }
32595 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
32596  uint8x16_t __ret;
32597  __ret = (uint8x16_t)(__p0);
32598  return __ret;
32599 }
32600 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
32601  uint8x16_t __ret;
32602  __ret = (uint8x16_t)(__p0);
32603  return __ret;
32604 }
32605 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
32606  uint8x16_t __ret;
32607  __ret = (uint8x16_t)(__p0);
32608  return __ret;
32609 }
32610 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
32611  uint8x16_t __ret;
32612  __ret = (uint8x16_t)(__p0);
32613  return __ret;
32614 }
32615 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
32616  uint8x16_t __ret;
32617  __ret = (uint8x16_t)(__p0);
32618  return __ret;
32619 }
32620 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
32621  uint8x16_t __ret;
32622  __ret = (uint8x16_t)(__p0);
32623  return __ret;
32624 }
32625 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
32626  uint8x16_t __ret;
32627  __ret = (uint8x16_t)(__p0);
32628  return __ret;
32629 }
32630 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
32631  uint8x16_t __ret;
32632  __ret = (uint8x16_t)(__p0);
32633  return __ret;
32634 }
32635 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
32636  uint8x16_t __ret;
32637  __ret = (uint8x16_t)(__p0);
32638  return __ret;
32639 }
32640 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
32641  uint32x4_t __ret;
32642  __ret = (uint32x4_t)(__p0);
32643  return __ret;
32644 }
32645 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
32646  uint32x4_t __ret;
32647  __ret = (uint32x4_t)(__p0);
32648  return __ret;
32649 }
32650 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
32651  uint32x4_t __ret;
32652  __ret = (uint32x4_t)(__p0);
32653  return __ret;
32654 }
32655 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
32656  uint32x4_t __ret;
32657  __ret = (uint32x4_t)(__p0);
32658  return __ret;
32659 }
32660 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
32661  uint32x4_t __ret;
32662  __ret = (uint32x4_t)(__p0);
32663  return __ret;
32664 }
32665 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
32666  uint32x4_t __ret;
32667  __ret = (uint32x4_t)(__p0);
32668  return __ret;
32669 }
32670 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
32671  uint32x4_t __ret;
32672  __ret = (uint32x4_t)(__p0);
32673  return __ret;
32674 }
32675 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
32676  uint32x4_t __ret;
32677  __ret = (uint32x4_t)(__p0);
32678  return __ret;
32679 }
32680 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
32681  uint32x4_t __ret;
32682  __ret = (uint32x4_t)(__p0);
32683  return __ret;
32684 }
32685 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
32686  uint32x4_t __ret;
32687  __ret = (uint32x4_t)(__p0);
32688  return __ret;
32689 }
32690 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
32691  uint32x4_t __ret;
32692  __ret = (uint32x4_t)(__p0);
32693  return __ret;
32694 }
32695 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
32696  uint64x2_t __ret;
32697  __ret = (uint64x2_t)(__p0);
32698  return __ret;
32699 }
32700 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
32701  uint64x2_t __ret;
32702  __ret = (uint64x2_t)(__p0);
32703  return __ret;
32704 }
32705 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
32706  uint64x2_t __ret;
32707  __ret = (uint64x2_t)(__p0);
32708  return __ret;
32709 }
32710 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
32711  uint64x2_t __ret;
32712  __ret = (uint64x2_t)(__p0);
32713  return __ret;
32714 }
32715 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
32716  uint64x2_t __ret;
32717  __ret = (uint64x2_t)(__p0);
32718  return __ret;
32719 }
32720 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
32721  uint64x2_t __ret;
32722  __ret = (uint64x2_t)(__p0);
32723  return __ret;
32724 }
32725 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
32726  uint64x2_t __ret;
32727  __ret = (uint64x2_t)(__p0);
32728  return __ret;
32729 }
32730 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
32731  uint64x2_t __ret;
32732  __ret = (uint64x2_t)(__p0);
32733  return __ret;
32734 }
32735 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
32736  uint64x2_t __ret;
32737  __ret = (uint64x2_t)(__p0);
32738  return __ret;
32739 }
32740 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
32741  uint64x2_t __ret;
32742  __ret = (uint64x2_t)(__p0);
32743  return __ret;
32744 }
32745 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
32746  uint64x2_t __ret;
32747  __ret = (uint64x2_t)(__p0);
32748  return __ret;
32749 }
32750 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
32751  uint16x8_t __ret;
32752  __ret = (uint16x8_t)(__p0);
32753  return __ret;
32754 }
32755 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
32756  uint16x8_t __ret;
32757  __ret = (uint16x8_t)(__p0);
32758  return __ret;
32759 }
32760 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
32761  uint16x8_t __ret;
32762  __ret = (uint16x8_t)(__p0);
32763  return __ret;
32764 }
32765 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
32766  uint16x8_t __ret;
32767  __ret = (uint16x8_t)(__p0);
32768  return __ret;
32769 }
32770 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
32771  uint16x8_t __ret;
32772  __ret = (uint16x8_t)(__p0);
32773  return __ret;
32774 }
32775 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
32776  uint16x8_t __ret;
32777  __ret = (uint16x8_t)(__p0);
32778  return __ret;
32779 }
32780 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
32781  uint16x8_t __ret;
32782  __ret = (uint16x8_t)(__p0);
32783  return __ret;
32784 }
32785 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
32786  uint16x8_t __ret;
32787  __ret = (uint16x8_t)(__p0);
32788  return __ret;
32789 }
32790 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
32791  uint16x8_t __ret;
32792  __ret = (uint16x8_t)(__p0);
32793  return __ret;
32794 }
32795 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
32796  uint16x8_t __ret;
32797  __ret = (uint16x8_t)(__p0);
32798  return __ret;
32799 }
32800 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
32801  uint16x8_t __ret;
32802  __ret = (uint16x8_t)(__p0);
32803  return __ret;
32804 }
32805 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
32806  int8x16_t __ret;
32807  __ret = (int8x16_t)(__p0);
32808  return __ret;
32809 }
32810 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
32811  int8x16_t __ret;
32812  __ret = (int8x16_t)(__p0);
32813  return __ret;
32814 }
32815 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
32816  int8x16_t __ret;
32817  __ret = (int8x16_t)(__p0);
32818  return __ret;
32819 }
32820 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
32821  int8x16_t __ret;
32822  __ret = (int8x16_t)(__p0);
32823  return __ret;
32824 }
32825 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
32826  int8x16_t __ret;
32827  __ret = (int8x16_t)(__p0);
32828  return __ret;
32829 }
32830 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
32831  int8x16_t __ret;
32832  __ret = (int8x16_t)(__p0);
32833  return __ret;
32834 }
32835 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
32836  int8x16_t __ret;
32837  __ret = (int8x16_t)(__p0);
32838  return __ret;
32839 }
32840 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
32841  int8x16_t __ret;
32842  __ret = (int8x16_t)(__p0);
32843  return __ret;
32844 }
32845 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
32846  int8x16_t __ret;
32847  __ret = (int8x16_t)(__p0);
32848  return __ret;
32849 }
32850 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
32851  int8x16_t __ret;
32852  __ret = (int8x16_t)(__p0);
32853  return __ret;
32854 }
32855 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
32856  int8x16_t __ret;
32857  __ret = (int8x16_t)(__p0);
32858  return __ret;
32859 }
32860 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
32861  float32x4_t __ret;
32862  __ret = (float32x4_t)(__p0);
32863  return __ret;
32864 }
32865 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
32866  float32x4_t __ret;
32867  __ret = (float32x4_t)(__p0);
32868  return __ret;
32869 }
32870 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
32871  float32x4_t __ret;
32872  __ret = (float32x4_t)(__p0);
32873  return __ret;
32874 }
32875 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
32876  float32x4_t __ret;
32877  __ret = (float32x4_t)(__p0);
32878  return __ret;
32879 }
32880 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
32881  float32x4_t __ret;
32882  __ret = (float32x4_t)(__p0);
32883  return __ret;
32884 }
32885 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
32886  float32x4_t __ret;
32887  __ret = (float32x4_t)(__p0);
32888  return __ret;
32889 }
32890 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
32891  float32x4_t __ret;
32892  __ret = (float32x4_t)(__p0);
32893  return __ret;
32894 }
32895 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
32896  float32x4_t __ret;
32897  __ret = (float32x4_t)(__p0);
32898  return __ret;
32899 }
32900 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
32901  float32x4_t __ret;
32902  __ret = (float32x4_t)(__p0);
32903  return __ret;
32904 }
32905 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
32906  float32x4_t __ret;
32907  __ret = (float32x4_t)(__p0);
32908  return __ret;
32909 }
32910 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
32911  float32x4_t __ret;
32912  __ret = (float32x4_t)(__p0);
32913  return __ret;
32914 }
32915 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
32916  float16x8_t __ret;
32917  __ret = (float16x8_t)(__p0);
32918  return __ret;
32919 }
32920 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
32921  float16x8_t __ret;
32922  __ret = (float16x8_t)(__p0);
32923  return __ret;
32924 }
32925 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
32926  float16x8_t __ret;
32927  __ret = (float16x8_t)(__p0);
32928  return __ret;
32929 }
32930 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
32931  float16x8_t __ret;
32932  __ret = (float16x8_t)(__p0);
32933  return __ret;
32934 }
32935 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
32936  float16x8_t __ret;
32937  __ret = (float16x8_t)(__p0);
32938  return __ret;
32939 }
32940 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
32941  float16x8_t __ret;
32942  __ret = (float16x8_t)(__p0);
32943  return __ret;
32944 }
32945 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
32946  float16x8_t __ret;
32947  __ret = (float16x8_t)(__p0);
32948  return __ret;
32949 }
32950 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
32951  float16x8_t __ret;
32952  __ret = (float16x8_t)(__p0);
32953  return __ret;
32954 }
32955 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
32956  float16x8_t __ret;
32957  __ret = (float16x8_t)(__p0);
32958  return __ret;
32959 }
32960 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
32961  float16x8_t __ret;
32962  __ret = (float16x8_t)(__p0);
32963  return __ret;
32964 }
32965 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
32966  float16x8_t __ret;
32967  __ret = (float16x8_t)(__p0);
32968  return __ret;
32969 }
32970 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
32971  int32x4_t __ret;
32972  __ret = (int32x4_t)(__p0);
32973  return __ret;
32974 }
32975 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
32976  int32x4_t __ret;
32977  __ret = (int32x4_t)(__p0);
32978  return __ret;
32979 }
32980 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
32981  int32x4_t __ret;
32982  __ret = (int32x4_t)(__p0);
32983  return __ret;
32984 }
32985 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
32986  int32x4_t __ret;
32987  __ret = (int32x4_t)(__p0);
32988  return __ret;
32989 }
32990 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
32991  int32x4_t __ret;
32992  __ret = (int32x4_t)(__p0);
32993  return __ret;
32994 }
32995 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
32996  int32x4_t __ret;
32997  __ret = (int32x4_t)(__p0);
32998  return __ret;
32999 }
33000 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
33001  int32x4_t __ret;
33002  __ret = (int32x4_t)(__p0);
33003  return __ret;
33004 }
33005 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
33006  int32x4_t __ret;
33007  __ret = (int32x4_t)(__p0);
33008  return __ret;
33009 }
33010 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
33011  int32x4_t __ret;
33012  __ret = (int32x4_t)(__p0);
33013  return __ret;
33014 }
33015 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
33016  int32x4_t __ret;
33017  __ret = (int32x4_t)(__p0);
33018  return __ret;
33019 }
33020 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
33021  int32x4_t __ret;
33022  __ret = (int32x4_t)(__p0);
33023  return __ret;
33024 }
33025 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
33026  int64x2_t __ret;
33027  __ret = (int64x2_t)(__p0);
33028  return __ret;
33029 }
33030 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
33031  int64x2_t __ret;
33032  __ret = (int64x2_t)(__p0);
33033  return __ret;
33034 }
33035 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
33036  int64x2_t __ret;
33037  __ret = (int64x2_t)(__p0);
33038  return __ret;
33039 }
33040 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
33041  int64x2_t __ret;
33042  __ret = (int64x2_t)(__p0);
33043  return __ret;
33044 }
33045 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
33046  int64x2_t __ret;
33047  __ret = (int64x2_t)(__p0);
33048  return __ret;
33049 }
33050 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
33051  int64x2_t __ret;
33052  __ret = (int64x2_t)(__p0);
33053  return __ret;
33054 }
33055 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
33056  int64x2_t __ret;
33057  __ret = (int64x2_t)(__p0);
33058  return __ret;
33059 }
33060 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
33061  int64x2_t __ret;
33062  __ret = (int64x2_t)(__p0);
33063  return __ret;
33064 }
33065 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
33066  int64x2_t __ret;
33067  __ret = (int64x2_t)(__p0);
33068  return __ret;
33069 }
33070 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
33071  int64x2_t __ret;
33072  __ret = (int64x2_t)(__p0);
33073  return __ret;
33074 }
33075 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
33076  int64x2_t __ret;
33077  __ret = (int64x2_t)(__p0);
33078  return __ret;
33079 }
33080 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
33081  int16x8_t __ret;
33082  __ret = (int16x8_t)(__p0);
33083  return __ret;
33084 }
33085 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
33086  int16x8_t __ret;
33087  __ret = (int16x8_t)(__p0);
33088  return __ret;
33089 }
33090 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
33091  int16x8_t __ret;
33092  __ret = (int16x8_t)(__p0);
33093  return __ret;
33094 }
33095 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
33096  int16x8_t __ret;
33097  __ret = (int16x8_t)(__p0);
33098  return __ret;
33099 }
33100 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
33101  int16x8_t __ret;
33102  __ret = (int16x8_t)(__p0);
33103  return __ret;
33104 }
33105 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
33106  int16x8_t __ret;
33107  __ret = (int16x8_t)(__p0);
33108  return __ret;
33109 }
33110 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
33111  int16x8_t __ret;
33112  __ret = (int16x8_t)(__p0);
33113  return __ret;
33114 }
33115 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
33116  int16x8_t __ret;
33117  __ret = (int16x8_t)(__p0);
33118  return __ret;
33119 }
33120 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
33121  int16x8_t __ret;
33122  __ret = (int16x8_t)(__p0);
33123  return __ret;
33124 }
33125 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
33126  int16x8_t __ret;
33127  __ret = (int16x8_t)(__p0);
33128  return __ret;
33129 }
33130 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
33131  int16x8_t __ret;
33132  __ret = (int16x8_t)(__p0);
33133  return __ret;
33134 }
33135 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
33136  uint8x8_t __ret;
33137  __ret = (uint8x8_t)(__p0);
33138  return __ret;
33139 }
33140 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
33141  uint8x8_t __ret;
33142  __ret = (uint8x8_t)(__p0);
33143  return __ret;
33144 }
33145 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
33146  uint8x8_t __ret;
33147  __ret = (uint8x8_t)(__p0);
33148  return __ret;
33149 }
33150 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
33151  uint8x8_t __ret;
33152  __ret = (uint8x8_t)(__p0);
33153  return __ret;
33154 }
33155 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
33156  uint8x8_t __ret;
33157  __ret = (uint8x8_t)(__p0);
33158  return __ret;
33159 }
33160 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
33161  uint8x8_t __ret;
33162  __ret = (uint8x8_t)(__p0);
33163  return __ret;
33164 }
33165 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
33166  uint8x8_t __ret;
33167  __ret = (uint8x8_t)(__p0);
33168  return __ret;
33169 }
33170 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
33171  uint8x8_t __ret;
33172  __ret = (uint8x8_t)(__p0);
33173  return __ret;
33174 }
33175 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
33176  uint8x8_t __ret;
33177  __ret = (uint8x8_t)(__p0);
33178  return __ret;
33179 }
33180 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
33181  uint8x8_t __ret;
33182  __ret = (uint8x8_t)(__p0);
33183  return __ret;
33184 }
33185 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
33186  uint8x8_t __ret;
33187  __ret = (uint8x8_t)(__p0);
33188  return __ret;
33189 }
33190 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
33191  uint32x2_t __ret;
33192  __ret = (uint32x2_t)(__p0);
33193  return __ret;
33194 }
33195 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
33196  uint32x2_t __ret;
33197  __ret = (uint32x2_t)(__p0);
33198  return __ret;
33199 }
33200 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
33201  uint32x2_t __ret;
33202  __ret = (uint32x2_t)(__p0);
33203  return __ret;
33204 }
33205 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
33206  uint32x2_t __ret;
33207  __ret = (uint32x2_t)(__p0);
33208  return __ret;
33209 }
33210 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
33211  uint32x2_t __ret;
33212  __ret = (uint32x2_t)(__p0);
33213  return __ret;
33214 }
33215 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
33216  uint32x2_t __ret;
33217  __ret = (uint32x2_t)(__p0);
33218  return __ret;
33219 }
33220 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
33221  uint32x2_t __ret;
33222  __ret = (uint32x2_t)(__p0);
33223  return __ret;
33224 }
33225 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
33226  uint32x2_t __ret;
33227  __ret = (uint32x2_t)(__p0);
33228  return __ret;
33229 }
33230 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
33231  uint32x2_t __ret;
33232  __ret = (uint32x2_t)(__p0);
33233  return __ret;
33234 }
33235 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
33236  uint32x2_t __ret;
33237  __ret = (uint32x2_t)(__p0);
33238  return __ret;
33239 }
33240 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
33241  uint32x2_t __ret;
33242  __ret = (uint32x2_t)(__p0);
33243  return __ret;
33244 }
33245 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
33246  uint64x1_t __ret;
33247  __ret = (uint64x1_t)(__p0);
33248  return __ret;
33249 }
33250 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
33251  uint64x1_t __ret;
33252  __ret = (uint64x1_t)(__p0);
33253  return __ret;
33254 }
33255 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
33256  uint64x1_t __ret;
33257  __ret = (uint64x1_t)(__p0);
33258  return __ret;
33259 }
33260 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
33261  uint64x1_t __ret;
33262  __ret = (uint64x1_t)(__p0);
33263  return __ret;
33264 }
33265 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
33266  uint64x1_t __ret;
33267  __ret = (uint64x1_t)(__p0);
33268  return __ret;
33269 }
33270 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
33271  uint64x1_t __ret;
33272  __ret = (uint64x1_t)(__p0);
33273  return __ret;
33274 }
33275 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
33276  uint64x1_t __ret;
33277  __ret = (uint64x1_t)(__p0);
33278  return __ret;
33279 }
33280 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
33281  uint64x1_t __ret;
33282  __ret = (uint64x1_t)(__p0);
33283  return __ret;
33284 }
33285 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
33286  uint64x1_t __ret;
33287  __ret = (uint64x1_t)(__p0);
33288  return __ret;
33289 }
33290 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
33291  uint64x1_t __ret;
33292  __ret = (uint64x1_t)(__p0);
33293  return __ret;
33294 }
33295 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
33296  uint64x1_t __ret;
33297  __ret = (uint64x1_t)(__p0);
33298  return __ret;
33299 }
33300 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
33301  uint16x4_t __ret;
33302  __ret = (uint16x4_t)(__p0);
33303  return __ret;
33304 }
33305 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
33306  uint16x4_t __ret;
33307  __ret = (uint16x4_t)(__p0);
33308  return __ret;
33309 }
33310 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
33311  uint16x4_t __ret;
33312  __ret = (uint16x4_t)(__p0);
33313  return __ret;
33314 }
33315 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
33316  uint16x4_t __ret;
33317  __ret = (uint16x4_t)(__p0);
33318  return __ret;
33319 }
33320 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
33321  uint16x4_t __ret;
33322  __ret = (uint16x4_t)(__p0);
33323  return __ret;
33324 }
33325 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
33326  uint16x4_t __ret;
33327  __ret = (uint16x4_t)(__p0);
33328  return __ret;
33329 }
33330 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
33331  uint16x4_t __ret;
33332  __ret = (uint16x4_t)(__p0);
33333  return __ret;
33334 }
33335 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
33336  uint16x4_t __ret;
33337  __ret = (uint16x4_t)(__p0);
33338  return __ret;
33339 }
33340 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
33341  uint16x4_t __ret;
33342  __ret = (uint16x4_t)(__p0);
33343  return __ret;
33344 }
33345 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
33346  uint16x4_t __ret;
33347  __ret = (uint16x4_t)(__p0);
33348  return __ret;
33349 }
33350 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
33351  uint16x4_t __ret;
33352  __ret = (uint16x4_t)(__p0);
33353  return __ret;
33354 }
33355 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
33356  int8x8_t __ret;
33357  __ret = (int8x8_t)(__p0);
33358  return __ret;
33359 }
33360 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
33361  int8x8_t __ret;
33362  __ret = (int8x8_t)(__p0);
33363  return __ret;
33364 }
33365 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
33366  int8x8_t __ret;
33367  __ret = (int8x8_t)(__p0);
33368  return __ret;
33369 }
33370 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
33371  int8x8_t __ret;
33372  __ret = (int8x8_t)(__p0);
33373  return __ret;
33374 }
33375 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
33376  int8x8_t __ret;
33377  __ret = (int8x8_t)(__p0);
33378  return __ret;
33379 }
33380 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
33381  int8x8_t __ret;
33382  __ret = (int8x8_t)(__p0);
33383  return __ret;
33384 }
33385 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
33386  int8x8_t __ret;
33387  __ret = (int8x8_t)(__p0);
33388  return __ret;
33389 }
33390 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
33391  int8x8_t __ret;
33392  __ret = (int8x8_t)(__p0);
33393  return __ret;
33394 }
33395 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
33396  int8x8_t __ret;
33397  __ret = (int8x8_t)(__p0);
33398  return __ret;
33399 }
33400 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
33401  int8x8_t __ret;
33402  __ret = (int8x8_t)(__p0);
33403  return __ret;
33404 }
33405 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
33406  int8x8_t __ret;
33407  __ret = (int8x8_t)(__p0);
33408  return __ret;
33409 }
33410 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
33411  float32x2_t __ret;
33412  __ret = (float32x2_t)(__p0);
33413  return __ret;
33414 }
33415 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
33416  float32x2_t __ret;
33417  __ret = (float32x2_t)(__p0);
33418  return __ret;
33419 }
33420 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
33421  float32x2_t __ret;
33422  __ret = (float32x2_t)(__p0);
33423  return __ret;
33424 }
33425 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
33426  float32x2_t __ret;
33427  __ret = (float32x2_t)(__p0);
33428  return __ret;
33429 }
33430 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
33431  float32x2_t __ret;
33432  __ret = (float32x2_t)(__p0);
33433  return __ret;
33434 }
33435 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
33436  float32x2_t __ret;
33437  __ret = (float32x2_t)(__p0);
33438  return __ret;
33439 }
33440 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
33441  float32x2_t __ret;
33442  __ret = (float32x2_t)(__p0);
33443  return __ret;
33444 }
33445 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
33446  float32x2_t __ret;
33447  __ret = (float32x2_t)(__p0);
33448  return __ret;
33449 }
33450 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
33451  float32x2_t __ret;
33452  __ret = (float32x2_t)(__p0);
33453  return __ret;
33454 }
33455 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
33456  float32x2_t __ret;
33457  __ret = (float32x2_t)(__p0);
33458  return __ret;
33459 }
33460 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
33461  float32x2_t __ret;
33462  __ret = (float32x2_t)(__p0);
33463  return __ret;
33464 }
33465 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
33466  float16x4_t __ret;
33467  __ret = (float16x4_t)(__p0);
33468  return __ret;
33469 }
33470 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
33471  float16x4_t __ret;
33472  __ret = (float16x4_t)(__p0);
33473  return __ret;
33474 }
33475 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33476  float16x4_t __ret;
33477  __ret = (float16x4_t)(__p0);
33478  return __ret;
33479 }
33480 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33481  float16x4_t __ret;
33482  __ret = (float16x4_t)(__p0);
33483  return __ret;
33484 }
33485 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33486  float16x4_t __ret;
33487  __ret = (float16x4_t)(__p0);
33488  return __ret;
33489 }
33490 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33491  float16x4_t __ret;
33492  __ret = (float16x4_t)(__p0);
33493  return __ret;
33494 }
33495 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33496  float16x4_t __ret;
33497  __ret = (float16x4_t)(__p0);
33498  return __ret;
33499 }
33500 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33501  float16x4_t __ret;
33502  __ret = (float16x4_t)(__p0);
33503  return __ret;
33504 }
33505 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33506  float16x4_t __ret;
33507  __ret = (float16x4_t)(__p0);
33508  return __ret;
33509 }
33510 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33511  float16x4_t __ret;
33512  __ret = (float16x4_t)(__p0);
33513  return __ret;
33514 }
33515 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33516  float16x4_t __ret;
33517  __ret = (float16x4_t)(__p0);
33518  return __ret;
33519 }
33520 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33521  int32x2_t __ret;
33522  __ret = (int32x2_t)(__p0);
33523  return __ret;
33524 }
33525 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33526  int32x2_t __ret;
33527  __ret = (int32x2_t)(__p0);
33528  return __ret;
33529 }
33530 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33531  int32x2_t __ret;
33532  __ret = (int32x2_t)(__p0);
33533  return __ret;
33534 }
33535 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33536  int32x2_t __ret;
33537  __ret = (int32x2_t)(__p0);
33538  return __ret;
33539 }
33540 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33541  int32x2_t __ret;
33542  __ret = (int32x2_t)(__p0);
33543  return __ret;
33544 }
33545 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33546  int32x2_t __ret;
33547  __ret = (int32x2_t)(__p0);
33548  return __ret;
33549 }
33550 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33551  int32x2_t __ret;
33552  __ret = (int32x2_t)(__p0);
33553  return __ret;
33554 }
33555 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33556  int32x2_t __ret;
33557  __ret = (int32x2_t)(__p0);
33558  return __ret;
33559 }
33560 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33561  int32x2_t __ret;
33562  __ret = (int32x2_t)(__p0);
33563  return __ret;
33564 }
33565 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33566  int32x2_t __ret;
33567  __ret = (int32x2_t)(__p0);
33568  return __ret;
33569 }
33570 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33571  int32x2_t __ret;
33572  __ret = (int32x2_t)(__p0);
33573  return __ret;
33574 }
33575 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33576  int64x1_t __ret;
33577  __ret = (int64x1_t)(__p0);
33578  return __ret;
33579 }
33580 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33581  int64x1_t __ret;
33582  __ret = (int64x1_t)(__p0);
33583  return __ret;
33584 }
33585 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33586  int64x1_t __ret;
33587  __ret = (int64x1_t)(__p0);
33588  return __ret;
33589 }
33590 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33591  int64x1_t __ret;
33592  __ret = (int64x1_t)(__p0);
33593  return __ret;
33594 }
33595 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33596  int64x1_t __ret;
33597  __ret = (int64x1_t)(__p0);
33598  return __ret;
33599 }
33600 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33601  int64x1_t __ret;
33602  __ret = (int64x1_t)(__p0);
33603  return __ret;
33604 }
33605 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33606  int64x1_t __ret;
33607  __ret = (int64x1_t)(__p0);
33608  return __ret;
33609 }
33610 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33611  int64x1_t __ret;
33612  __ret = (int64x1_t)(__p0);
33613  return __ret;
33614 }
33615 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33616  int64x1_t __ret;
33617  __ret = (int64x1_t)(__p0);
33618  return __ret;
33619 }
33620 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33621  int64x1_t __ret;
33622  __ret = (int64x1_t)(__p0);
33623  return __ret;
33624 }
33625 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33626  int64x1_t __ret;
33627  __ret = (int64x1_t)(__p0);
33628  return __ret;
33629 }
33630 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33631  int16x4_t __ret;
33632  __ret = (int16x4_t)(__p0);
33633  return __ret;
33634 }
33635 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33636  int16x4_t __ret;
33637  __ret = (int16x4_t)(__p0);
33638  return __ret;
33639 }
33640 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33641  int16x4_t __ret;
33642  __ret = (int16x4_t)(__p0);
33643  return __ret;
33644 }
33645 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33646  int16x4_t __ret;
33647  __ret = (int16x4_t)(__p0);
33648  return __ret;
33649 }
33650 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33651  int16x4_t __ret;
33652  __ret = (int16x4_t)(__p0);
33653  return __ret;
33654 }
33655 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33656  int16x4_t __ret;
33657  __ret = (int16x4_t)(__p0);
33658  return __ret;
33659 }
33660 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33661  int16x4_t __ret;
33662  __ret = (int16x4_t)(__p0);
33663  return __ret;
33664 }
33665 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33666  int16x4_t __ret;
33667  __ret = (int16x4_t)(__p0);
33668  return __ret;
33669 }
33670 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33671  int16x4_t __ret;
33672  __ret = (int16x4_t)(__p0);
33673  return __ret;
33674 }
33675 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33676  int16x4_t __ret;
33677  __ret = (int16x4_t)(__p0);
33678  return __ret;
33679 }
33680 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33681  int16x4_t __ret;
33682  __ret = (int16x4_t)(__p0);
33683  return __ret;
33684 }
33685 #endif
33686 #if (__ARM_FP & 2)
33687 #ifdef __LITTLE_ENDIAN__
33688 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
33689  float16x4_t __ret;
33690  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 41);
33691  return __ret;
33692 }
33693 #else
33694 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
33695  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33696  float16x4_t __ret;
33697  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 41);
33698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33699  return __ret;
33700 }
33701 __ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
33702  float16x4_t __ret;
33703  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 41);
33704  return __ret;
33705 }
33706 #endif
33707 
33708 #ifdef __LITTLE_ENDIAN__
33709 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
33710  float32x4_t __ret;
33711  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
33712  return __ret;
33713 }
33714 #else
33715 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
33716  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33717  float32x4_t __ret;
33718  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 8);
33719  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33720  return __ret;
33721 }
33722 __ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
33723  float32x4_t __ret;
33724  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
33725  return __ret;
33726 }
33727 #endif
33728 
33729 #ifdef __LITTLE_ENDIAN__
33730 #define vld1q_f16(__p0) __extension__ ({ \
33731  float16x8_t __ret; \
33732  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
33733  __ret; \
33734 })
33735 #else
33736 #define vld1q_f16(__p0) __extension__ ({ \
33737  float16x8_t __ret; \
33738  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
33739  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
33740  __ret; \
33741 })
33742 #endif
33743 
33744 #ifdef __LITTLE_ENDIAN__
33745 #define vld1_f16(__p0) __extension__ ({ \
33746  float16x4_t __ret; \
33747  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
33748  __ret; \
33749 })
33750 #else
33751 #define vld1_f16(__p0) __extension__ ({ \
33752  float16x4_t __ret; \
33753  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
33754  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
33755  __ret; \
33756 })
33757 #endif
33758 
33759 #ifdef __LITTLE_ENDIAN__
33760 #define vld1q_dup_f16(__p0) __extension__ ({ \
33761  float16x8_t __ret; \
33762  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
33763  __ret; \
33764 })
33765 #else
33766 #define vld1q_dup_f16(__p0) __extension__ ({ \
33767  float16x8_t __ret; \
33768  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
33769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
33770  __ret; \
33771 })
33772 #endif
33773 
33774 #ifdef __LITTLE_ENDIAN__
33775 #define vld1_dup_f16(__p0) __extension__ ({ \
33776  float16x4_t __ret; \
33777  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
33778  __ret; \
33779 })
33780 #else
33781 #define vld1_dup_f16(__p0) __extension__ ({ \
33782  float16x4_t __ret; \
33783  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
33784  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
33785  __ret; \
33786 })
33787 #endif
33788 
33789 #ifdef __LITTLE_ENDIAN__
33790 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
33791  float16x8_t __s1 = __p1; \
33792  float16x8_t __ret; \
33793  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
33794  __ret; \
33795 })
33796 #else
33797 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
33798  float16x8_t __s1 = __p1; \
33799  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
33800  float16x8_t __ret; \
33801  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
33802  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
33803  __ret; \
33804 })
33805 #endif
33806 
33807 #ifdef __LITTLE_ENDIAN__
33808 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
33809  float16x4_t __s1 = __p1; \
33810  float16x4_t __ret; \
33811  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
33812  __ret; \
33813 })
33814 #else
33815 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
33816  float16x4_t __s1 = __p1; \
33817  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
33818  float16x4_t __ret; \
33819  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
33820  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
33821  __ret; \
33822 })
33823 #endif
33824 
33825 #ifdef __LITTLE_ENDIAN__
33826 #define vld1q_f16_x2(__p0) __extension__ ({ \
33827  float16x8x2_t __ret; \
33828  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
33829  __ret; \
33830 })
33831 #else
33832 #define vld1q_f16_x2(__p0) __extension__ ({ \
33833  float16x8x2_t __ret; \
33834  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
33835  \
33836  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33837  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33838  __ret; \
33839 })
33840 #endif
33841 
33842 #ifdef __LITTLE_ENDIAN__
33843 #define vld1_f16_x2(__p0) __extension__ ({ \
33844  float16x4x2_t __ret; \
33845  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
33846  __ret; \
33847 })
33848 #else
33849 #define vld1_f16_x2(__p0) __extension__ ({ \
33850  float16x4x2_t __ret; \
33851  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
33852  \
33853  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33854  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33855  __ret; \
33856 })
33857 #endif
33858 
33859 #ifdef __LITTLE_ENDIAN__
33860 #define vld1q_f16_x3(__p0) __extension__ ({ \
33861  float16x8x3_t __ret; \
33862  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
33863  __ret; \
33864 })
33865 #else
33866 #define vld1q_f16_x3(__p0) __extension__ ({ \
33867  float16x8x3_t __ret; \
33868  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
33869  \
33870  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33871  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33872  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33873  __ret; \
33874 })
33875 #endif
33876 
33877 #ifdef __LITTLE_ENDIAN__
33878 #define vld1_f16_x3(__p0) __extension__ ({ \
33879  float16x4x3_t __ret; \
33880  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
33881  __ret; \
33882 })
33883 #else
33884 #define vld1_f16_x3(__p0) __extension__ ({ \
33885  float16x4x3_t __ret; \
33886  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
33887  \
33888  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33889  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33890  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
33891  __ret; \
33892 })
33893 #endif
33894 
33895 #ifdef __LITTLE_ENDIAN__
33896 #define vld1q_f16_x4(__p0) __extension__ ({ \
33897  float16x8x4_t __ret; \
33898  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
33899  __ret; \
33900 })
33901 #else
33902 #define vld1q_f16_x4(__p0) __extension__ ({ \
33903  float16x8x4_t __ret; \
33904  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
33905  \
33906  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33907  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33908  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
33909  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
33910  __ret; \
33911 })
33912 #endif
33913 
33914 #ifdef __LITTLE_ENDIAN__
33915 #define vld1_f16_x4(__p0) __extension__ ({ \
33916  float16x4x4_t __ret; \
33917  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
33918  __ret; \
33919 })
33920 #else
33921 #define vld1_f16_x4(__p0) __extension__ ({ \
33922  float16x4x4_t __ret; \
33923  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
33924  \
33925  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33926  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33927  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
33928  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
33929  __ret; \
33930 })
33931 #endif
33932 
33933 #ifdef __LITTLE_ENDIAN__
33934 #define vld2q_f16(__p0) __extension__ ({ \
33935  float16x8x2_t __ret; \
33936  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
33937  __ret; \
33938 })
33939 #else
33940 #define vld2q_f16(__p0) __extension__ ({ \
33941  float16x8x2_t __ret; \
33942  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
33943  \
33944  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33945  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33946  __ret; \
33947 })
33948 #endif
33949 
33950 #ifdef __LITTLE_ENDIAN__
33951 #define vld2_f16(__p0) __extension__ ({ \
33952  float16x4x2_t __ret; \
33953  __builtin_neon_vld2_v(&__ret, __p0, 8); \
33954  __ret; \
33955 })
33956 #else
33957 #define vld2_f16(__p0) __extension__ ({ \
33958  float16x4x2_t __ret; \
33959  __builtin_neon_vld2_v(&__ret, __p0, 8); \
33960  \
33961  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33962  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33963  __ret; \
33964 })
33965 #endif
33966 
33967 #ifdef __LITTLE_ENDIAN__
33968 #define vld2q_dup_f16(__p0) __extension__ ({ \
33969  float16x8x2_t __ret; \
33970  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
33971  __ret; \
33972 })
33973 #else
33974 #define vld2q_dup_f16(__p0) __extension__ ({ \
33975  float16x8x2_t __ret; \
33976  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
33977  \
33978  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
33979  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
33980  __ret; \
33981 })
33982 #endif
33983 
33984 #ifdef __LITTLE_ENDIAN__
33985 #define vld2_dup_f16(__p0) __extension__ ({ \
33986  float16x4x2_t __ret; \
33987  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
33988  __ret; \
33989 })
33990 #else
33991 #define vld2_dup_f16(__p0) __extension__ ({ \
33992  float16x4x2_t __ret; \
33993  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
33994  \
33995  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
33996  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
33997  __ret; \
33998 })
33999 #endif
34000 
34001 #ifdef __LITTLE_ENDIAN__
34002 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34003  float16x8x2_t __s1 = __p1; \
34004  float16x8x2_t __ret; \
34005  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 40); \
34006  __ret; \
34007 })
34008 #else
34009 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34010  float16x8x2_t __s1 = __p1; \
34011  float16x8x2_t __rev1; \
34012  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34013  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34014  float16x8x2_t __ret; \
34015  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 40); \
34016  \
34017  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34018  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34019  __ret; \
34020 })
34021 #endif
34022 
34023 #ifdef __LITTLE_ENDIAN__
34024 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34025  float16x4x2_t __s1 = __p1; \
34026  float16x4x2_t __ret; \
34027  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 8); \
34028  __ret; \
34029 })
34030 #else
34031 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34032  float16x4x2_t __s1 = __p1; \
34033  float16x4x2_t __rev1; \
34034  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34035  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34036  float16x4x2_t __ret; \
34037  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 8); \
34038  \
34039  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34040  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34041  __ret; \
34042 })
34043 #endif
34044 
34045 #ifdef __LITTLE_ENDIAN__
34046 #define vld3q_f16(__p0) __extension__ ({ \
34047  float16x8x3_t __ret; \
34048  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
34049  __ret; \
34050 })
34051 #else
34052 #define vld3q_f16(__p0) __extension__ ({ \
34053  float16x8x3_t __ret; \
34054  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
34055  \
34056  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34057  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34058  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34059  __ret; \
34060 })
34061 #endif
34062 
34063 #ifdef __LITTLE_ENDIAN__
34064 #define vld3_f16(__p0) __extension__ ({ \
34065  float16x4x3_t __ret; \
34066  __builtin_neon_vld3_v(&__ret, __p0, 8); \
34067  __ret; \
34068 })
34069 #else
34070 #define vld3_f16(__p0) __extension__ ({ \
34071  float16x4x3_t __ret; \
34072  __builtin_neon_vld3_v(&__ret, __p0, 8); \
34073  \
34074  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34075  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34076  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34077  __ret; \
34078 })
34079 #endif
34080 
34081 #ifdef __LITTLE_ENDIAN__
34082 #define vld3q_dup_f16(__p0) __extension__ ({ \
34083  float16x8x3_t __ret; \
34084  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
34085  __ret; \
34086 })
34087 #else
34088 #define vld3q_dup_f16(__p0) __extension__ ({ \
34089  float16x8x3_t __ret; \
34090  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
34091  \
34092  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34093  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34094  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34095  __ret; \
34096 })
34097 #endif
34098 
34099 #ifdef __LITTLE_ENDIAN__
34100 #define vld3_dup_f16(__p0) __extension__ ({ \
34101  float16x4x3_t __ret; \
34102  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
34103  __ret; \
34104 })
34105 #else
34106 #define vld3_dup_f16(__p0) __extension__ ({ \
34107  float16x4x3_t __ret; \
34108  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
34109  \
34110  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34111  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34112  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34113  __ret; \
34114 })
34115 #endif
34116 
34117 #ifdef __LITTLE_ENDIAN__
34118 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34119  float16x8x3_t __s1 = __p1; \
34120  float16x8x3_t __ret; \
34121  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 40); \
34122  __ret; \
34123 })
34124 #else
34125 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34126  float16x8x3_t __s1 = __p1; \
34127  float16x8x3_t __rev1; \
34128  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34129  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34130  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34131  float16x8x3_t __ret; \
34132  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 40); \
34133  \
34134  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34135  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34136  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34137  __ret; \
34138 })
34139 #endif
34140 
34141 #ifdef __LITTLE_ENDIAN__
34142 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34143  float16x4x3_t __s1 = __p1; \
34144  float16x4x3_t __ret; \
34145  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 8); \
34146  __ret; \
34147 })
34148 #else
34149 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34150  float16x4x3_t __s1 = __p1; \
34151  float16x4x3_t __rev1; \
34152  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34153  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34154  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34155  float16x4x3_t __ret; \
34156  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 8); \
34157  \
34158  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34159  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34160  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34161  __ret; \
34162 })
34163 #endif
34164 
34165 #ifdef __LITTLE_ENDIAN__
34166 #define vld4q_f16(__p0) __extension__ ({ \
34167  float16x8x4_t __ret; \
34168  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
34169  __ret; \
34170 })
34171 #else
34172 #define vld4q_f16(__p0) __extension__ ({ \
34173  float16x8x4_t __ret; \
34174  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
34175  \
34176  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34177  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34178  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34179  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34180  __ret; \
34181 })
34182 #endif
34183 
34184 #ifdef __LITTLE_ENDIAN__
34185 #define vld4_f16(__p0) __extension__ ({ \
34186  float16x4x4_t __ret; \
34187  __builtin_neon_vld4_v(&__ret, __p0, 8); \
34188  __ret; \
34189 })
34190 #else
34191 #define vld4_f16(__p0) __extension__ ({ \
34192  float16x4x4_t __ret; \
34193  __builtin_neon_vld4_v(&__ret, __p0, 8); \
34194  \
34195  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34196  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34197  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34198  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
34199  __ret; \
34200 })
34201 #endif
34202 
34203 #ifdef __LITTLE_ENDIAN__
34204 #define vld4q_dup_f16(__p0) __extension__ ({ \
34205  float16x8x4_t __ret; \
34206  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
34207  __ret; \
34208 })
34209 #else
34210 #define vld4q_dup_f16(__p0) __extension__ ({ \
34211  float16x8x4_t __ret; \
34212  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
34213  \
34214  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34215  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34216  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34217  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34218  __ret; \
34219 })
34220 #endif
34221 
34222 #ifdef __LITTLE_ENDIAN__
34223 #define vld4_dup_f16(__p0) __extension__ ({ \
34224  float16x4x4_t __ret; \
34225  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
34226  __ret; \
34227 })
34228 #else
34229 #define vld4_dup_f16(__p0) __extension__ ({ \
34230  float16x4x4_t __ret; \
34231  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
34232  \
34233  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34234  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34235  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34236  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
34237  __ret; \
34238 })
34239 #endif
34240 
34241 #ifdef __LITTLE_ENDIAN__
34242 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34243  float16x8x4_t __s1 = __p1; \
34244  float16x8x4_t __ret; \
34245  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 40); \
34246  __ret; \
34247 })
34248 #else
34249 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34250  float16x8x4_t __s1 = __p1; \
34251  float16x8x4_t __rev1; \
34252  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34253  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34254  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34255  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34256  float16x8x4_t __ret; \
34257  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 40); \
34258  \
34259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34261  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34262  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34263  __ret; \
34264 })
34265 #endif
34266 
34267 #ifdef __LITTLE_ENDIAN__
34268 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34269  float16x4x4_t __s1 = __p1; \
34270  float16x4x4_t __ret; \
34271  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 8); \
34272  __ret; \
34273 })
34274 #else
34275 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34276  float16x4x4_t __s1 = __p1; \
34277  float16x4x4_t __rev1; \
34278  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34279  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34280  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34281  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
34282  float16x4x4_t __ret; \
34283  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 8); \
34284  \
34285  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
34286  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
34287  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
34288  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
34289  __ret; \
34290 })
34291 #endif
34292 
34293 #ifdef __LITTLE_ENDIAN__
34294 #define vst1q_f16(__p0, __p1) __extension__ ({ \
34295  float16x8_t __s1 = __p1; \
34296  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
34297 })
34298 #else
34299 #define vst1q_f16(__p0, __p1) __extension__ ({ \
34300  float16x8_t __s1 = __p1; \
34301  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
34302  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
34303 })
34304 #endif
34305 
34306 #ifdef __LITTLE_ENDIAN__
34307 #define vst1_f16(__p0, __p1) __extension__ ({ \
34308  float16x4_t __s1 = __p1; \
34309  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
34310 })
34311 #else
34312 #define vst1_f16(__p0, __p1) __extension__ ({ \
34313  float16x4_t __s1 = __p1; \
34314  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
34315  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
34316 })
34317 #endif
34318 
34319 #ifdef __LITTLE_ENDIAN__
34320 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34321  float16x8_t __s1 = __p1; \
34322  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
34323 })
34324 #else
34325 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34326  float16x8_t __s1 = __p1; \
34327  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
34328  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
34329 })
34330 #endif
34331 
34332 #ifdef __LITTLE_ENDIAN__
34333 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34334  float16x4_t __s1 = __p1; \
34335  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
34336 })
34337 #else
34338 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34339  float16x4_t __s1 = __p1; \
34340  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
34341  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
34342 })
34343 #endif
34344 
34345 #ifdef __LITTLE_ENDIAN__
34346 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
34347  float16x8x2_t __s1 = __p1; \
34348  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 40); \
34349 })
34350 #else
34351 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
34352  float16x8x2_t __s1 = __p1; \
34353  float16x8x2_t __rev1; \
34354  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34355  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34356  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 40); \
34357 })
34358 #endif
34359 
34360 #ifdef __LITTLE_ENDIAN__
34361 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
34362  float16x4x2_t __s1 = __p1; \
34363  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 8); \
34364 })
34365 #else
34366 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
34367  float16x4x2_t __s1 = __p1; \
34368  float16x4x2_t __rev1; \
34369  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34370  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34371  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 8); \
34372 })
34373 #endif
34374 
34375 #ifdef __LITTLE_ENDIAN__
34376 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
34377  float16x8x3_t __s1 = __p1; \
34378  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 40); \
34379 })
34380 #else
34381 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
34382  float16x8x3_t __s1 = __p1; \
34383  float16x8x3_t __rev1; \
34384  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34385  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34386  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34387  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 40); \
34388 })
34389 #endif
34390 
34391 #ifdef __LITTLE_ENDIAN__
34392 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
34393  float16x4x3_t __s1 = __p1; \
34394  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 8); \
34395 })
34396 #else
34397 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
34398  float16x4x3_t __s1 = __p1; \
34399  float16x4x3_t __rev1; \
34400  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34401  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34402  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34403  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 8); \
34404 })
34405 #endif
34406 
34407 #ifdef __LITTLE_ENDIAN__
34408 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
34409  float16x8x4_t __s1 = __p1; \
34410  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 40); \
34411 })
34412 #else
34413 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
34414  float16x8x4_t __s1 = __p1; \
34415  float16x8x4_t __rev1; \
34416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34419  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34420  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 40); \
34421 })
34422 #endif
34423 
34424 #ifdef __LITTLE_ENDIAN__
34425 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
34426  float16x4x4_t __s1 = __p1; \
34427  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 8); \
34428 })
34429 #else
34430 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
34431  float16x4x4_t __s1 = __p1; \
34432  float16x4x4_t __rev1; \
34433  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34434  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34435  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34436  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
34437  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 8); \
34438 })
34439 #endif
34440 
34441 #ifdef __LITTLE_ENDIAN__
34442 #define vst2q_f16(__p0, __p1) __extension__ ({ \
34443  float16x8x2_t __s1 = __p1; \
34444  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 40); \
34445 })
34446 #else
34447 #define vst2q_f16(__p0, __p1) __extension__ ({ \
34448  float16x8x2_t __s1 = __p1; \
34449  float16x8x2_t __rev1; \
34450  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34451  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34452  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 40); \
34453 })
34454 #endif
34455 
34456 #ifdef __LITTLE_ENDIAN__
34457 #define vst2_f16(__p0, __p1) __extension__ ({ \
34458  float16x4x2_t __s1 = __p1; \
34459  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 8); \
34460 })
34461 #else
34462 #define vst2_f16(__p0, __p1) __extension__ ({ \
34463  float16x4x2_t __s1 = __p1; \
34464  float16x4x2_t __rev1; \
34465  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34466  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34467  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 8); \
34468 })
34469 #endif
34470 
34471 #ifdef __LITTLE_ENDIAN__
34472 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34473  float16x8x2_t __s1 = __p1; \
34474  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 40); \
34475 })
34476 #else
34477 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34478  float16x8x2_t __s1 = __p1; \
34479  float16x8x2_t __rev1; \
34480  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34481  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34482  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 40); \
34483 })
34484 #endif
34485 
34486 #ifdef __LITTLE_ENDIAN__
34487 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34488  float16x4x2_t __s1 = __p1; \
34489  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 8); \
34490 })
34491 #else
34492 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34493  float16x4x2_t __s1 = __p1; \
34494  float16x4x2_t __rev1; \
34495  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34496  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34497  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 8); \
34498 })
34499 #endif
34500 
34501 #ifdef __LITTLE_ENDIAN__
34502 #define vst3q_f16(__p0, __p1) __extension__ ({ \
34503  float16x8x3_t __s1 = __p1; \
34504  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 40); \
34505 })
34506 #else
34507 #define vst3q_f16(__p0, __p1) __extension__ ({ \
34508  float16x8x3_t __s1 = __p1; \
34509  float16x8x3_t __rev1; \
34510  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34511  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34512  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34513  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 40); \
34514 })
34515 #endif
34516 
34517 #ifdef __LITTLE_ENDIAN__
34518 #define vst3_f16(__p0, __p1) __extension__ ({ \
34519  float16x4x3_t __s1 = __p1; \
34520  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 8); \
34521 })
34522 #else
34523 #define vst3_f16(__p0, __p1) __extension__ ({ \
34524  float16x4x3_t __s1 = __p1; \
34525  float16x4x3_t __rev1; \
34526  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34527  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34528  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34529  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 8); \
34530 })
34531 #endif
34532 
34533 #ifdef __LITTLE_ENDIAN__
34534 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34535  float16x8x3_t __s1 = __p1; \
34536  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 40); \
34537 })
34538 #else
34539 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34540  float16x8x3_t __s1 = __p1; \
34541  float16x8x3_t __rev1; \
34542  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34543  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34544  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34545  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 40); \
34546 })
34547 #endif
34548 
34549 #ifdef __LITTLE_ENDIAN__
34550 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34551  float16x4x3_t __s1 = __p1; \
34552  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 8); \
34553 })
34554 #else
34555 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34556  float16x4x3_t __s1 = __p1; \
34557  float16x4x3_t __rev1; \
34558  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34559  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34560  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34561  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 8); \
34562 })
34563 #endif
34564 
34565 #ifdef __LITTLE_ENDIAN__
34566 #define vst4q_f16(__p0, __p1) __extension__ ({ \
34567  float16x8x4_t __s1 = __p1; \
34568  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 40); \
34569 })
34570 #else
34571 #define vst4q_f16(__p0, __p1) __extension__ ({ \
34572  float16x8x4_t __s1 = __p1; \
34573  float16x8x4_t __rev1; \
34574  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34575  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34576  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34577  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34578  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 40); \
34579 })
34580 #endif
34581 
34582 #ifdef __LITTLE_ENDIAN__
34583 #define vst4_f16(__p0, __p1) __extension__ ({ \
34584  float16x4x4_t __s1 = __p1; \
34585  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 8); \
34586 })
34587 #else
34588 #define vst4_f16(__p0, __p1) __extension__ ({ \
34589  float16x4x4_t __s1 = __p1; \
34590  float16x4x4_t __rev1; \
34591  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34592  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34593  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34594  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
34595  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 8); \
34596 })
34597 #endif
34598 
34599 #ifdef __LITTLE_ENDIAN__
34600 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34601  float16x8x4_t __s1 = __p1; \
34602  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 40); \
34603 })
34604 #else
34605 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34606  float16x8x4_t __s1 = __p1; \
34607  float16x8x4_t __rev1; \
34608  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
34609  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
34610  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
34611  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
34612  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 40); \
34613 })
34614 #endif
34615 
34616 #ifdef __LITTLE_ENDIAN__
34617 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34618  float16x4x4_t __s1 = __p1; \
34619  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 8); \
34620 })
34621 #else
34622 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
34623  float16x4x4_t __s1 = __p1; \
34624  float16x4x4_t __rev1; \
34625  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
34626  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
34627  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
34628  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
34629  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 8); \
34630 })
34631 #endif
34632 
34633 #endif
34634 #if __ARM_ARCH >= 8
34635 #ifdef __LITTLE_ENDIAN__
34636 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
34637  int32x4_t __ret;
34638  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
34639  return __ret;
34640 }
34641 #else
34642 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
34643  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34644  int32x4_t __ret;
34645  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
34646  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34647  return __ret;
34648 }
34649 #endif
34650 
34651 #ifdef __LITTLE_ENDIAN__
34652 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
34653  int32x2_t __ret;
34654  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
34655  return __ret;
34656 }
34657 #else
34658 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
34659  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34660  int32x2_t __ret;
34661  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
34662  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34663  return __ret;
34664 }
34665 #endif
34666 
34667 #ifdef __LITTLE_ENDIAN__
34668 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
34669  uint32x4_t __ret;
34670  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
34671  return __ret;
34672 }
34673 #else
34674 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
34675  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34676  uint32x4_t __ret;
34677  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
34678  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34679  return __ret;
34680 }
34681 #endif
34682 
34683 #ifdef __LITTLE_ENDIAN__
34684 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
34685  uint32x2_t __ret;
34686  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
34687  return __ret;
34688 }
34689 #else
34690 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
34691  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34692  uint32x2_t __ret;
34693  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
34694  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34695  return __ret;
34696 }
34697 #endif
34698 
34699 #ifdef __LITTLE_ENDIAN__
34700 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
34701  int32x4_t __ret;
34702  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
34703  return __ret;
34704 }
34705 #else
34706 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
34707  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34708  int32x4_t __ret;
34709  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
34710  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34711  return __ret;
34712 }
34713 #endif
34714 
34715 #ifdef __LITTLE_ENDIAN__
34716 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
34717  int32x2_t __ret;
34718  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
34719  return __ret;
34720 }
34721 #else
34722 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
34723  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34724  int32x2_t __ret;
34725  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
34726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34727  return __ret;
34728 }
34729 #endif
34730 
34731 #ifdef __LITTLE_ENDIAN__
34732 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
34733  uint32x4_t __ret;
34734  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
34735  return __ret;
34736 }
34737 #else
34738 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
34739  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34740  uint32x4_t __ret;
34741  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
34742  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34743  return __ret;
34744 }
34745 #endif
34746 
34747 #ifdef __LITTLE_ENDIAN__
34748 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
34749  uint32x2_t __ret;
34750  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
34751  return __ret;
34752 }
34753 #else
34754 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
34755  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34756  uint32x2_t __ret;
34757  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
34758  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34759  return __ret;
34760 }
34761 #endif
34762 
34763 #ifdef __LITTLE_ENDIAN__
34764 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
34765  int32x4_t __ret;
34766  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
34767  return __ret;
34768 }
34769 #else
34770 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
34771  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34772  int32x4_t __ret;
34773  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
34774  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34775  return __ret;
34776 }
34777 #endif
34778 
34779 #ifdef __LITTLE_ENDIAN__
34780 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
34781  int32x2_t __ret;
34782  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
34783  return __ret;
34784 }
34785 #else
34786 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
34787  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34788  int32x2_t __ret;
34789  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
34790  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34791  return __ret;
34792 }
34793 #endif
34794 
34795 #ifdef __LITTLE_ENDIAN__
34796 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
34797  uint32x4_t __ret;
34798  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
34799  return __ret;
34800 }
34801 #else
34802 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
34803  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34804  uint32x4_t __ret;
34805  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
34806  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34807  return __ret;
34808 }
34809 #endif
34810 
34811 #ifdef __LITTLE_ENDIAN__
34812 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
34813  uint32x2_t __ret;
34814  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
34815  return __ret;
34816 }
34817 #else
34818 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
34819  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34820  uint32x2_t __ret;
34821  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
34822  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34823  return __ret;
34824 }
34825 #endif
34826 
34827 #ifdef __LITTLE_ENDIAN__
34828 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
34829  int32x4_t __ret;
34830  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
34831  return __ret;
34832 }
34833 #else
34834 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
34835  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34836  int32x4_t __ret;
34837  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
34838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34839  return __ret;
34840 }
34841 #endif
34842 
34843 #ifdef __LITTLE_ENDIAN__
34844 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
34845  int32x2_t __ret;
34846  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
34847  return __ret;
34848 }
34849 #else
34850 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
34851  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34852  int32x2_t __ret;
34853  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
34854  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34855  return __ret;
34856 }
34857 #endif
34858 
34859 #ifdef __LITTLE_ENDIAN__
34860 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
34861  uint32x4_t __ret;
34862  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
34863  return __ret;
34864 }
34865 #else
34866 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
34867  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34868  uint32x4_t __ret;
34869  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
34870  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34871  return __ret;
34872 }
34873 #endif
34874 
34875 #ifdef __LITTLE_ENDIAN__
34876 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
34877  uint32x2_t __ret;
34878  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
34879  return __ret;
34880 }
34881 #else
34882 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
34883  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34884  uint32x2_t __ret;
34885  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
34886  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34887  return __ret;
34888 }
34889 #endif
34890 
34891 #endif
34892 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_CRYPTO)
34893 #ifdef __LITTLE_ENDIAN__
34894 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
34895  uint8x16_t __ret;
34896  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
34897  return __ret;
34898 }
34899 #else
34900 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
34901  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34902  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34903  uint8x16_t __ret;
34904  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
34905  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34906  return __ret;
34907 }
34908 #endif
34909 
34910 #ifdef __LITTLE_ENDIAN__
34911 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
34912  uint8x16_t __ret;
34913  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
34914  return __ret;
34915 }
34916 #else
34917 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
34918  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34919  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34920  uint8x16_t __ret;
34921  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
34922  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34923  return __ret;
34924 }
34925 #endif
34926 
34927 #ifdef __LITTLE_ENDIAN__
34928 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
34929  uint8x16_t __ret;
34930  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
34931  return __ret;
34932 }
34933 #else
34934 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
34935  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34936  uint8x16_t __ret;
34937  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
34938  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34939  return __ret;
34940 }
34941 #endif
34942 
34943 #ifdef __LITTLE_ENDIAN__
34944 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
34945  uint8x16_t __ret;
34946  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
34947  return __ret;
34948 }
34949 #else
34950 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
34951  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34952  uint8x16_t __ret;
34953  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
34954  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
34955  return __ret;
34956 }
34957 #endif
34958 
34959 #ifdef __LITTLE_ENDIAN__
34960 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
34961  uint32x4_t __ret;
34962  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2);
34963  return __ret;
34964 }
34965 #else
34966 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
34967  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34968  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
34969  uint32x4_t __ret;
34970  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2);
34971  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34972  return __ret;
34973 }
34974 #endif
34975 
34976 __ai uint32_t vsha1h_u32(uint32_t __p0) {
34977  uint32_t __ret;
34978  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
34979  return __ret;
34980 }
34981 #ifdef __LITTLE_ENDIAN__
34982 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
34983  uint32x4_t __ret;
34984  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2);
34985  return __ret;
34986 }
34987 #else
34988 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
34989  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34990  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
34991  uint32x4_t __ret;
34992  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2);
34993  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34994  return __ret;
34995 }
34996 #endif
34997 
34998 #ifdef __LITTLE_ENDIAN__
34999 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
35000  uint32x4_t __ret;
35001  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2);
35002  return __ret;
35003 }
35004 #else
35005 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
35006  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35007  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35008  uint32x4_t __ret;
35009  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2);
35010  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35011  return __ret;
35012 }
35013 #endif
35014 
35015 #ifdef __LITTLE_ENDIAN__
35016 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35017  uint32x4_t __ret;
35018  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
35019  return __ret;
35020 }
35021 #else
35022 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35023  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35024  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35025  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35026  uint32x4_t __ret;
35027  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
35028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35029  return __ret;
35030 }
35031 #endif
35032 
35033 #ifdef __LITTLE_ENDIAN__
35034 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
35035  uint32x4_t __ret;
35036  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
35037  return __ret;
35038 }
35039 #else
35040 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
35041  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35042  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35043  uint32x4_t __ret;
35044  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
35045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35046  return __ret;
35047 }
35048 #endif
35049 
35050 #ifdef __LITTLE_ENDIAN__
35051 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35052  uint32x4_t __ret;
35053  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
35054  return __ret;
35055 }
35056 #else
35057 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35058  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35059  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35060  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35061  uint32x4_t __ret;
35062  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
35063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35064  return __ret;
35065 }
35066 #endif
35067 
35068 #ifdef __LITTLE_ENDIAN__
35069 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35070  uint32x4_t __ret;
35071  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
35072  return __ret;
35073 }
35074 #else
35075 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35076  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35077  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35078  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35079  uint32x4_t __ret;
35080  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
35081  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35082  return __ret;
35083 }
35084 #endif
35085 
35086 #ifdef __LITTLE_ENDIAN__
35087 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
35088  uint32x4_t __ret;
35089  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
35090  return __ret;
35091 }
35092 #else
35093 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
35094  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35095  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35096  uint32x4_t __ret;
35097  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
35098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35099  return __ret;
35100 }
35101 #endif
35102 
35103 #ifdef __LITTLE_ENDIAN__
35104 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35105  uint32x4_t __ret;
35106  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
35107  return __ret;
35108 }
35109 #else
35110 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
35111  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35112  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35113  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
35114  uint32x4_t __ret;
35115  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
35116  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35117  return __ret;
35118 }
35119 #endif
35120 
35121 #endif
35122 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
35123 #ifdef __LITTLE_ENDIAN__
35124 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
35125  float32x4_t __ret;
35126  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
35127  return __ret;
35128 }
35129 #else
35130 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
35131  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35132  float32x4_t __ret;
35133  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
35134  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35135  return __ret;
35136 }
35137 #endif
35138 
35139 #ifdef __LITTLE_ENDIAN__
35140 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
35141  float32x2_t __ret;
35142  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
35143  return __ret;
35144 }
35145 #else
35146 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
35147  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35148  float32x2_t __ret;
35149  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
35150  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35151  return __ret;
35152 }
35153 #endif
35154 
35155 #ifdef __LITTLE_ENDIAN__
35156 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
35157  float32x4_t __ret;
35158  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
35159  return __ret;
35160 }
35161 #else
35162 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
35163  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35164  float32x4_t __ret;
35165  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
35166  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35167  return __ret;
35168 }
35169 #endif
35170 
35171 #ifdef __LITTLE_ENDIAN__
35172 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
35173  float32x2_t __ret;
35174  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
35175  return __ret;
35176 }
35177 #else
35178 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
35179  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35180  float32x2_t __ret;
35181  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
35182  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35183  return __ret;
35184 }
35185 #endif
35186 
35187 #ifdef __LITTLE_ENDIAN__
35188 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
35189  float32x4_t __ret;
35190  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
35191  return __ret;
35192 }
35193 #else
35194 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
35195  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35196  float32x4_t __ret;
35197  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
35198  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35199  return __ret;
35200 }
35201 #endif
35202 
35203 #ifdef __LITTLE_ENDIAN__
35204 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
35205  float32x2_t __ret;
35206  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
35207  return __ret;
35208 }
35209 #else
35210 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
35211  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35212  float32x2_t __ret;
35213  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
35214  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35215  return __ret;
35216 }
35217 #endif
35218 
35219 #ifdef __LITTLE_ENDIAN__
35220 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
35221  float32x4_t __ret;
35222  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
35223  return __ret;
35224 }
35225 #else
35226 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
35227  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35228  float32x4_t __ret;
35229  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
35230  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35231  return __ret;
35232 }
35233 #endif
35234 
35235 #ifdef __LITTLE_ENDIAN__
35236 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
35237  float32x2_t __ret;
35238  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
35239  return __ret;
35240 }
35241 #else
35242 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
35243  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35244  float32x2_t __ret;
35245  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
35246  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35247  return __ret;
35248 }
35249 #endif
35250 
35251 #ifdef __LITTLE_ENDIAN__
35252 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
35253  float32x4_t __ret;
35254  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
35255  return __ret;
35256 }
35257 #else
35258 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
35259  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35260  float32x4_t __ret;
35261  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
35262  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35263  return __ret;
35264 }
35265 #endif
35266 
35267 #ifdef __LITTLE_ENDIAN__
35268 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
35269  float32x2_t __ret;
35270  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
35271  return __ret;
35272 }
35273 #else
35274 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
35275  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35276  float32x2_t __ret;
35277  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
35278  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35279  return __ret;
35280 }
35281 #endif
35282 
35283 __ai float32_t vrndns_f32(float32_t __p0) {
35284  float32_t __ret;
35285  __ret = (float32_t) __builtin_neon_vrndns_f32(__p0);
35286  return __ret;
35287 }
35288 #ifdef __LITTLE_ENDIAN__
35289 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
35290  float32x4_t __ret;
35291  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
35292  return __ret;
35293 }
35294 #else
35295 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
35296  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35297  float32x4_t __ret;
35298  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
35299  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35300  return __ret;
35301 }
35302 #endif
35303 
35304 #ifdef __LITTLE_ENDIAN__
35305 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
35306  float32x2_t __ret;
35307  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
35308  return __ret;
35309 }
35310 #else
35311 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
35312  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35313  float32x2_t __ret;
35314  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
35315  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35316  return __ret;
35317 }
35318 #endif
35319 
35320 #ifdef __LITTLE_ENDIAN__
35321 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
35322  float32x4_t __ret;
35323  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
35324  return __ret;
35325 }
35326 #else
35327 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
35328  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35329  float32x4_t __ret;
35330  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
35331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35332  return __ret;
35333 }
35334 #endif
35335 
35336 #ifdef __LITTLE_ENDIAN__
35337 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
35338  float32x2_t __ret;
35339  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
35340  return __ret;
35341 }
35342 #else
35343 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
35344  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35345  float32x2_t __ret;
35346  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
35347  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35348  return __ret;
35349 }
35350 #endif
35351 
35352 #endif
35353 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING) && defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC)
35354 #ifdef __LITTLE_ENDIAN__
35355 __ai float16x8_t vrndq_f16(float16x8_t __p0) {
35356  float16x8_t __ret;
35357  __ret = (float16x8_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 40);
35358  return __ret;
35359 }
35360 #else
35361 __ai float16x8_t vrndq_f16(float16x8_t __p0) {
35362  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35363  float16x8_t __ret;
35364  __ret = (float16x8_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 40);
35365  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35366  return __ret;
35367 }
35368 #endif
35369 
35370 #ifdef __LITTLE_ENDIAN__
35371 __ai float16x4_t vrnd_f16(float16x4_t __p0) {
35372  float16x4_t __ret;
35373  __ret = (float16x4_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 8);
35374  return __ret;
35375 }
35376 #else
35377 __ai float16x4_t vrnd_f16(float16x4_t __p0) {
35378  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35379  float16x4_t __ret;
35380  __ret = (float16x4_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 8);
35381  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35382  return __ret;
35383 }
35384 #endif
35385 
35386 #ifdef __LITTLE_ENDIAN__
35387 __ai float16x8_t vrndaq_f16(float16x8_t __p0) {
35388  float16x8_t __ret;
35389  __ret = (float16x8_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 40);
35390  return __ret;
35391 }
35392 #else
35393 __ai float16x8_t vrndaq_f16(float16x8_t __p0) {
35394  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35395  float16x8_t __ret;
35396  __ret = (float16x8_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 40);
35397  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35398  return __ret;
35399 }
35400 #endif
35401 
35402 #ifdef __LITTLE_ENDIAN__
35403 __ai float16x4_t vrnda_f16(float16x4_t __p0) {
35404  float16x4_t __ret;
35405  __ret = (float16x4_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 8);
35406  return __ret;
35407 }
35408 #else
35409 __ai float16x4_t vrnda_f16(float16x4_t __p0) {
35410  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35411  float16x4_t __ret;
35412  __ret = (float16x4_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 8);
35413  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35414  return __ret;
35415 }
35416 #endif
35417 
35418 #ifdef __LITTLE_ENDIAN__
35419 __ai float16x8_t vrndmq_f16(float16x8_t __p0) {
35420  float16x8_t __ret;
35421  __ret = (float16x8_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 40);
35422  return __ret;
35423 }
35424 #else
35425 __ai float16x8_t vrndmq_f16(float16x8_t __p0) {
35426  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35427  float16x8_t __ret;
35428  __ret = (float16x8_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 40);
35429  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35430  return __ret;
35431 }
35432 #endif
35433 
35434 #ifdef __LITTLE_ENDIAN__
35435 __ai float16x4_t vrndm_f16(float16x4_t __p0) {
35436  float16x4_t __ret;
35437  __ret = (float16x4_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 8);
35438  return __ret;
35439 }
35440 #else
35441 __ai float16x4_t vrndm_f16(float16x4_t __p0) {
35442  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35443  float16x4_t __ret;
35444  __ret = (float16x4_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 8);
35445  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35446  return __ret;
35447 }
35448 #endif
35449 
35450 #ifdef __LITTLE_ENDIAN__
35451 __ai float16x8_t vrndnq_f16(float16x8_t __p0) {
35452  float16x8_t __ret;
35453  __ret = (float16x8_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 40);
35454  return __ret;
35455 }
35456 #else
35457 __ai float16x8_t vrndnq_f16(float16x8_t __p0) {
35458  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35459  float16x8_t __ret;
35460  __ret = (float16x8_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 40);
35461  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35462  return __ret;
35463 }
35464 #endif
35465 
35466 #ifdef __LITTLE_ENDIAN__
35467 __ai float16x4_t vrndn_f16(float16x4_t __p0) {
35468  float16x4_t __ret;
35469  __ret = (float16x4_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 8);
35470  return __ret;
35471 }
35472 #else
35473 __ai float16x4_t vrndn_f16(float16x4_t __p0) {
35474  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35475  float16x4_t __ret;
35476  __ret = (float16x4_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 8);
35477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35478  return __ret;
35479 }
35480 #endif
35481 
35482 #ifdef __LITTLE_ENDIAN__
35483 __ai float16x8_t vrndpq_f16(float16x8_t __p0) {
35484  float16x8_t __ret;
35485  __ret = (float16x8_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 40);
35486  return __ret;
35487 }
35488 #else
35489 __ai float16x8_t vrndpq_f16(float16x8_t __p0) {
35490  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35491  float16x8_t __ret;
35492  __ret = (float16x8_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 40);
35493  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35494  return __ret;
35495 }
35496 #endif
35497 
35498 #ifdef __LITTLE_ENDIAN__
35499 __ai float16x4_t vrndp_f16(float16x4_t __p0) {
35500  float16x4_t __ret;
35501  __ret = (float16x4_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 8);
35502  return __ret;
35503 }
35504 #else
35505 __ai float16x4_t vrndp_f16(float16x4_t __p0) {
35506  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35507  float16x4_t __ret;
35508  __ret = (float16x4_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 8);
35509  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35510  return __ret;
35511 }
35512 #endif
35513 
35514 #ifdef __LITTLE_ENDIAN__
35515 __ai float16x8_t vrndxq_f16(float16x8_t __p0) {
35516  float16x8_t __ret;
35517  __ret = (float16x8_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 40);
35518  return __ret;
35519 }
35520 #else
35521 __ai float16x8_t vrndxq_f16(float16x8_t __p0) {
35522  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35523  float16x8_t __ret;
35524  __ret = (float16x8_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 40);
35525  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35526  return __ret;
35527 }
35528 #endif
35529 
35530 #ifdef __LITTLE_ENDIAN__
35531 __ai float16x4_t vrndx_f16(float16x4_t __p0) {
35532  float16x4_t __ret;
35533  __ret = (float16x4_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 8);
35534  return __ret;
35535 }
35536 #else
35537 __ai float16x4_t vrndx_f16(float16x4_t __p0) {
35538  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35539  float16x4_t __ret;
35540  __ret = (float16x4_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 8);
35541  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35542  return __ret;
35543 }
35544 #endif
35545 
35546 #endif
35547 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
35548 #ifdef __LITTLE_ENDIAN__
35549 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
35550  float32x4_t __ret;
35551  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
35552  return __ret;
35553 }
35554 #else
35555 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
35556  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35557  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35558  float32x4_t __ret;
35559  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
35560  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35561  return __ret;
35562 }
35563 #endif
35564 
35565 #ifdef __LITTLE_ENDIAN__
35566 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
35567  float32x2_t __ret;
35568  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
35569  return __ret;
35570 }
35571 #else
35572 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
35573  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35574  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
35575  float32x2_t __ret;
35576  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
35577  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35578  return __ret;
35579 }
35580 #endif
35581 
35582 #ifdef __LITTLE_ENDIAN__
35583 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
35584  float32x4_t __ret;
35585  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
35586  return __ret;
35587 }
35588 #else
35589 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
35590  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35591  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35592  float32x4_t __ret;
35593  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
35594  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35595  return __ret;
35596 }
35597 #endif
35598 
35599 #ifdef __LITTLE_ENDIAN__
35600 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
35601  float32x2_t __ret;
35602  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
35603  return __ret;
35604 }
35605 #else
35606 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
35607  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35608  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
35609  float32x2_t __ret;
35610  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
35611  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35612  return __ret;
35613 }
35614 #endif
35615 
35616 #endif
35617 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN) && defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC)
35618 #ifdef __LITTLE_ENDIAN__
35619 __ai float16x8_t vmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
35620  float16x8_t __ret;
35621  __ret = (float16x8_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
35622  return __ret;
35623 }
35624 #else
35625 __ai float16x8_t vmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
35626  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35627  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35628  float16x8_t __ret;
35629  __ret = (float16x8_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35630  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35631  return __ret;
35632 }
35633 #endif
35634 
35635 #ifdef __LITTLE_ENDIAN__
35636 __ai float16x4_t vmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
35637  float16x4_t __ret;
35638  __ret = (float16x4_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
35639  return __ret;
35640 }
35641 #else
35642 __ai float16x4_t vmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
35643  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35644  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35645  float16x4_t __ret;
35646  __ret = (float16x4_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35648  return __ret;
35649 }
35650 #endif
35651 
35652 #ifdef __LITTLE_ENDIAN__
35653 __ai float16x8_t vminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
35654  float16x8_t __ret;
35655  __ret = (float16x8_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
35656  return __ret;
35657 }
35658 #else
35659 __ai float16x8_t vminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
35660  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
35661  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
35662  float16x8_t __ret;
35663  __ret = (float16x8_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
35664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
35665  return __ret;
35666 }
35667 #endif
35668 
35669 #ifdef __LITTLE_ENDIAN__
35670 __ai float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) {
35671  float16x4_t __ret;
35672  __ret = (float16x4_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
35673  return __ret;
35674 }
35675 #else
35676 __ai float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) {
35677  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
35678  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
35679  float16x4_t __ret;
35680  __ret = (float16x4_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
35681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
35682  return __ret;
35683 }
35684 #endif
35685 
35686 #endif
35687 #if __ARM_ARCH >= 8 && defined(__aarch64__)
35688 #ifdef __LITTLE_ENDIAN__
35689 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
35690  int64x2_t __ret;
35691  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
35692  return __ret;
35693 }
35694 #else
35695 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
35696  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35697  int64x2_t __ret;
35698  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
35699  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35700  return __ret;
35701 }
35702 #endif
35703 
35704 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
35705  int64x1_t __ret;
35706  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
35707  return __ret;
35708 }
35709 #ifdef __LITTLE_ENDIAN__
35710 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
35711  uint64x2_t __ret;
35712  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
35713  return __ret;
35714 }
35715 #else
35716 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
35717  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35718  uint64x2_t __ret;
35719  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
35720  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35721  return __ret;
35722 }
35723 #endif
35724 
35725 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
35726  uint64x1_t __ret;
35727  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
35728  return __ret;
35729 }
35730 #ifdef __LITTLE_ENDIAN__
35731 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
35732  int64x2_t __ret;
35733  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
35734  return __ret;
35735 }
35736 #else
35737 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
35738  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35739  int64x2_t __ret;
35740  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
35741  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35742  return __ret;
35743 }
35744 #endif
35745 
35746 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
35747  int64x1_t __ret;
35748  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
35749  return __ret;
35750 }
35751 #ifdef __LITTLE_ENDIAN__
35752 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
35753  uint64x2_t __ret;
35754  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
35755  return __ret;
35756 }
35757 #else
35758 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
35759  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35760  uint64x2_t __ret;
35761  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
35762  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35763  return __ret;
35764 }
35765 #endif
35766 
35767 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
35768  uint64x1_t __ret;
35769  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
35770  return __ret;
35771 }
35772 #ifdef __LITTLE_ENDIAN__
35773 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
35774  int64x2_t __ret;
35775  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
35776  return __ret;
35777 }
35778 #else
35779 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
35780  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35781  int64x2_t __ret;
35782  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
35783  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35784  return __ret;
35785 }
35786 #endif
35787 
35788 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
35789  int64x1_t __ret;
35790  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
35791  return __ret;
35792 }
35793 #ifdef __LITTLE_ENDIAN__
35794 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
35795  uint64x2_t __ret;
35796  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
35797  return __ret;
35798 }
35799 #else
35800 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
35801  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35802  uint64x2_t __ret;
35803  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
35804  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35805  return __ret;
35806 }
35807 #endif
35808 
35809 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
35810  uint64x1_t __ret;
35811  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
35812  return __ret;
35813 }
35814 #ifdef __LITTLE_ENDIAN__
35815 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
35816  int64x2_t __ret;
35817  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
35818  return __ret;
35819 }
35820 #else
35821 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
35822  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35823  int64x2_t __ret;
35824  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
35825  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35826  return __ret;
35827 }
35828 #endif
35829 
35830 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
35831  int64x1_t __ret;
35832  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
35833  return __ret;
35834 }
35835 #ifdef __LITTLE_ENDIAN__
35836 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
35837  uint64x2_t __ret;
35838  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
35839  return __ret;
35840 }
35841 #else
35842 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
35843  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
35844  uint64x2_t __ret;
35845  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
35846  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
35847  return __ret;
35848 }
35849 #endif
35850 
35851 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
35852  uint64x1_t __ret;
35853  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
35854  return __ret;
35855 }
35856 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
35857  poly8x8_t __ret;
35858  __ret = (poly8x8_t)(__p0);
35859  return __ret;
35860 }
35861 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
35862  poly8x8_t __ret;
35863  __ret = (poly8x8_t)(__p0);
35864  return __ret;
35865 }
35866 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
35867  poly8x8_t __ret;
35868  __ret = (poly8x8_t)(__p0);
35869  return __ret;
35870 }
35871 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
35872  poly8x8_t __ret;
35873  __ret = (poly8x8_t)(__p0);
35874  return __ret;
35875 }
35876 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
35877  poly8x8_t __ret;
35878  __ret = (poly8x8_t)(__p0);
35879  return __ret;
35880 }
35881 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
35882  poly8x8_t __ret;
35883  __ret = (poly8x8_t)(__p0);
35884  return __ret;
35885 }
35886 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
35887  poly8x8_t __ret;
35888  __ret = (poly8x8_t)(__p0);
35889  return __ret;
35890 }
35891 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
35892  poly8x8_t __ret;
35893  __ret = (poly8x8_t)(__p0);
35894  return __ret;
35895 }
35896 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
35897  poly8x8_t __ret;
35898  __ret = (poly8x8_t)(__p0);
35899  return __ret;
35900 }
35901 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
35902  poly8x8_t __ret;
35903  __ret = (poly8x8_t)(__p0);
35904  return __ret;
35905 }
35906 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
35907  poly8x8_t __ret;
35908  __ret = (poly8x8_t)(__p0);
35909  return __ret;
35910 }
35911 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
35912  poly8x8_t __ret;
35913  __ret = (poly8x8_t)(__p0);
35914  return __ret;
35915 }
35916 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
35917  poly8x8_t __ret;
35918  __ret = (poly8x8_t)(__p0);
35919  return __ret;
35920 }
35921 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
35922  poly64x1_t __ret;
35923  __ret = (poly64x1_t)(__p0);
35924  return __ret;
35925 }
35926 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
35927  poly64x1_t __ret;
35928  __ret = (poly64x1_t)(__p0);
35929  return __ret;
35930 }
35931 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
35932  poly64x1_t __ret;
35933  __ret = (poly64x1_t)(__p0);
35934  return __ret;
35935 }
35936 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
35937  poly64x1_t __ret;
35938  __ret = (poly64x1_t)(__p0);
35939  return __ret;
35940 }
35941 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
35942  poly64x1_t __ret;
35943  __ret = (poly64x1_t)(__p0);
35944  return __ret;
35945 }
35946 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
35947  poly64x1_t __ret;
35948  __ret = (poly64x1_t)(__p0);
35949  return __ret;
35950 }
35951 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
35952  poly64x1_t __ret;
35953  __ret = (poly64x1_t)(__p0);
35954  return __ret;
35955 }
35956 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
35957  poly64x1_t __ret;
35958  __ret = (poly64x1_t)(__p0);
35959  return __ret;
35960 }
35961 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
35962  poly64x1_t __ret;
35963  __ret = (poly64x1_t)(__p0);
35964  return __ret;
35965 }
35966 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
35967  poly64x1_t __ret;
35968  __ret = (poly64x1_t)(__p0);
35969  return __ret;
35970 }
35971 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
35972  poly64x1_t __ret;
35973  __ret = (poly64x1_t)(__p0);
35974  return __ret;
35975 }
35976 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
35977  poly64x1_t __ret;
35978  __ret = (poly64x1_t)(__p0);
35979  return __ret;
35980 }
35981 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
35982  poly64x1_t __ret;
35983  __ret = (poly64x1_t)(__p0);
35984  return __ret;
35985 }
35986 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
35987  poly16x4_t __ret;
35988  __ret = (poly16x4_t)(__p0);
35989  return __ret;
35990 }
35991 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
35992  poly16x4_t __ret;
35993  __ret = (poly16x4_t)(__p0);
35994  return __ret;
35995 }
35996 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
35997  poly16x4_t __ret;
35998  __ret = (poly16x4_t)(__p0);
35999  return __ret;
36000 }
36001 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
36002  poly16x4_t __ret;
36003  __ret = (poly16x4_t)(__p0);
36004  return __ret;
36005 }
36006 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
36007  poly16x4_t __ret;
36008  __ret = (poly16x4_t)(__p0);
36009  return __ret;
36010 }
36011 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
36012  poly16x4_t __ret;
36013  __ret = (poly16x4_t)(__p0);
36014  return __ret;
36015 }
36016 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
36017  poly16x4_t __ret;
36018  __ret = (poly16x4_t)(__p0);
36019  return __ret;
36020 }
36021 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
36022  poly16x4_t __ret;
36023  __ret = (poly16x4_t)(__p0);
36024  return __ret;
36025 }
36026 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
36027  poly16x4_t __ret;
36028  __ret = (poly16x4_t)(__p0);
36029  return __ret;
36030 }
36031 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
36032  poly16x4_t __ret;
36033  __ret = (poly16x4_t)(__p0);
36034  return __ret;
36035 }
36036 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
36037  poly16x4_t __ret;
36038  __ret = (poly16x4_t)(__p0);
36039  return __ret;
36040 }
36041 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
36042  poly16x4_t __ret;
36043  __ret = (poly16x4_t)(__p0);
36044  return __ret;
36045 }
36046 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
36047  poly16x4_t __ret;
36048  __ret = (poly16x4_t)(__p0);
36049  return __ret;
36050 }
36051 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
36052  poly8x16_t __ret;
36053  __ret = (poly8x16_t)(__p0);
36054  return __ret;
36055 }
36056 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
36057  poly8x16_t __ret;
36058  __ret = (poly8x16_t)(__p0);
36059  return __ret;
36060 }
36061 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
36062  poly8x16_t __ret;
36063  __ret = (poly8x16_t)(__p0);
36064  return __ret;
36065 }
36066 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
36067  poly8x16_t __ret;
36068  __ret = (poly8x16_t)(__p0);
36069  return __ret;
36070 }
36071 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
36072  poly8x16_t __ret;
36073  __ret = (poly8x16_t)(__p0);
36074  return __ret;
36075 }
36076 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
36077  poly8x16_t __ret;
36078  __ret = (poly8x16_t)(__p0);
36079  return __ret;
36080 }
36081 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
36082  poly8x16_t __ret;
36083  __ret = (poly8x16_t)(__p0);
36084  return __ret;
36085 }
36086 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
36087  poly8x16_t __ret;
36088  __ret = (poly8x16_t)(__p0);
36089  return __ret;
36090 }
36091 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
36092  poly8x16_t __ret;
36093  __ret = (poly8x16_t)(__p0);
36094  return __ret;
36095 }
36096 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
36097  poly8x16_t __ret;
36098  __ret = (poly8x16_t)(__p0);
36099  return __ret;
36100 }
36101 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
36102  poly8x16_t __ret;
36103  __ret = (poly8x16_t)(__p0);
36104  return __ret;
36105 }
36106 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
36107  poly8x16_t __ret;
36108  __ret = (poly8x16_t)(__p0);
36109  return __ret;
36110 }
36111 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
36112  poly8x16_t __ret;
36113  __ret = (poly8x16_t)(__p0);
36114  return __ret;
36115 }
36116 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
36117  poly8x16_t __ret;
36118  __ret = (poly8x16_t)(__p0);
36119  return __ret;
36120 }
36121 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
36122  poly128_t __ret;
36123  __ret = (poly128_t)(__p0);
36124  return __ret;
36125 }
36126 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
36127  poly128_t __ret;
36128  __ret = (poly128_t)(__p0);
36129  return __ret;
36130 }
36131 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
36132  poly128_t __ret;
36133  __ret = (poly128_t)(__p0);
36134  return __ret;
36135 }
36136 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
36137  poly128_t __ret;
36138  __ret = (poly128_t)(__p0);
36139  return __ret;
36140 }
36141 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
36142  poly128_t __ret;
36143  __ret = (poly128_t)(__p0);
36144  return __ret;
36145 }
36146 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
36147  poly128_t __ret;
36148  __ret = (poly128_t)(__p0);
36149  return __ret;
36150 }
36151 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
36152  poly128_t __ret;
36153  __ret = (poly128_t)(__p0);
36154  return __ret;
36155 }
36156 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
36157  poly128_t __ret;
36158  __ret = (poly128_t)(__p0);
36159  return __ret;
36160 }
36161 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
36162  poly128_t __ret;
36163  __ret = (poly128_t)(__p0);
36164  return __ret;
36165 }
36166 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
36167  poly128_t __ret;
36168  __ret = (poly128_t)(__p0);
36169  return __ret;
36170 }
36171 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
36172  poly128_t __ret;
36173  __ret = (poly128_t)(__p0);
36174  return __ret;
36175 }
36176 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
36177  poly128_t __ret;
36178  __ret = (poly128_t)(__p0);
36179  return __ret;
36180 }
36181 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
36182  poly128_t __ret;
36183  __ret = (poly128_t)(__p0);
36184  return __ret;
36185 }
36186 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
36187  poly128_t __ret;
36188  __ret = (poly128_t)(__p0);
36189  return __ret;
36190 }
36191 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
36192  poly64x2_t __ret;
36193  __ret = (poly64x2_t)(__p0);
36194  return __ret;
36195 }
36196 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
36197  poly64x2_t __ret;
36198  __ret = (poly64x2_t)(__p0);
36199  return __ret;
36200 }
36201 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
36202  poly64x2_t __ret;
36203  __ret = (poly64x2_t)(__p0);
36204  return __ret;
36205 }
36206 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
36207  poly64x2_t __ret;
36208  __ret = (poly64x2_t)(__p0);
36209  return __ret;
36210 }
36211 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
36212  poly64x2_t __ret;
36213  __ret = (poly64x2_t)(__p0);
36214  return __ret;
36215 }
36216 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
36217  poly64x2_t __ret;
36218  __ret = (poly64x2_t)(__p0);
36219  return __ret;
36220 }
36221 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
36222  poly64x2_t __ret;
36223  __ret = (poly64x2_t)(__p0);
36224  return __ret;
36225 }
36226 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
36227  poly64x2_t __ret;
36228  __ret = (poly64x2_t)(__p0);
36229  return __ret;
36230 }
36231 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
36232  poly64x2_t __ret;
36233  __ret = (poly64x2_t)(__p0);
36234  return __ret;
36235 }
36236 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
36237  poly64x2_t __ret;
36238  __ret = (poly64x2_t)(__p0);
36239  return __ret;
36240 }
36241 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
36242  poly64x2_t __ret;
36243  __ret = (poly64x2_t)(__p0);
36244  return __ret;
36245 }
36246 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
36247  poly64x2_t __ret;
36248  __ret = (poly64x2_t)(__p0);
36249  return __ret;
36250 }
36251 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
36252  poly64x2_t __ret;
36253  __ret = (poly64x2_t)(__p0);
36254  return __ret;
36255 }
36256 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
36257  poly64x2_t __ret;
36258  __ret = (poly64x2_t)(__p0);
36259  return __ret;
36260 }
36261 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
36262  poly16x8_t __ret;
36263  __ret = (poly16x8_t)(__p0);
36264  return __ret;
36265 }
36266 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
36267  poly16x8_t __ret;
36268  __ret = (poly16x8_t)(__p0);
36269  return __ret;
36270 }
36271 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
36272  poly16x8_t __ret;
36273  __ret = (poly16x8_t)(__p0);
36274  return __ret;
36275 }
36276 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
36277  poly16x8_t __ret;
36278  __ret = (poly16x8_t)(__p0);
36279  return __ret;
36280 }
36281 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
36282  poly16x8_t __ret;
36283  __ret = (poly16x8_t)(__p0);
36284  return __ret;
36285 }
36286 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
36287  poly16x8_t __ret;
36288  __ret = (poly16x8_t)(__p0);
36289  return __ret;
36290 }
36291 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
36292  poly16x8_t __ret;
36293  __ret = (poly16x8_t)(__p0);
36294  return __ret;
36295 }
36296 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
36297  poly16x8_t __ret;
36298  __ret = (poly16x8_t)(__p0);
36299  return __ret;
36300 }
36301 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
36302  poly16x8_t __ret;
36303  __ret = (poly16x8_t)(__p0);
36304  return __ret;
36305 }
36306 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
36307  poly16x8_t __ret;
36308  __ret = (poly16x8_t)(__p0);
36309  return __ret;
36310 }
36311 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
36312  poly16x8_t __ret;
36313  __ret = (poly16x8_t)(__p0);
36314  return __ret;
36315 }
36316 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
36317  poly16x8_t __ret;
36318  __ret = (poly16x8_t)(__p0);
36319  return __ret;
36320 }
36321 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
36322  poly16x8_t __ret;
36323  __ret = (poly16x8_t)(__p0);
36324  return __ret;
36325 }
36326 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
36327  poly16x8_t __ret;
36328  __ret = (poly16x8_t)(__p0);
36329  return __ret;
36330 }
36331 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
36332  uint8x16_t __ret;
36333  __ret = (uint8x16_t)(__p0);
36334  return __ret;
36335 }
36336 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
36337  uint8x16_t __ret;
36338  __ret = (uint8x16_t)(__p0);
36339  return __ret;
36340 }
36341 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
36342  uint8x16_t __ret;
36343  __ret = (uint8x16_t)(__p0);
36344  return __ret;
36345 }
36346 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
36347  uint8x16_t __ret;
36348  __ret = (uint8x16_t)(__p0);
36349  return __ret;
36350 }
36351 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
36352  uint8x16_t __ret;
36353  __ret = (uint8x16_t)(__p0);
36354  return __ret;
36355 }
36356 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
36357  uint8x16_t __ret;
36358  __ret = (uint8x16_t)(__p0);
36359  return __ret;
36360 }
36361 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
36362  uint8x16_t __ret;
36363  __ret = (uint8x16_t)(__p0);
36364  return __ret;
36365 }
36366 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
36367  uint8x16_t __ret;
36368  __ret = (uint8x16_t)(__p0);
36369  return __ret;
36370 }
36371 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
36372  uint8x16_t __ret;
36373  __ret = (uint8x16_t)(__p0);
36374  return __ret;
36375 }
36376 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
36377  uint8x16_t __ret;
36378  __ret = (uint8x16_t)(__p0);
36379  return __ret;
36380 }
36381 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
36382  uint8x16_t __ret;
36383  __ret = (uint8x16_t)(__p0);
36384  return __ret;
36385 }
36386 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
36387  uint8x16_t __ret;
36388  __ret = (uint8x16_t)(__p0);
36389  return __ret;
36390 }
36391 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
36392  uint8x16_t __ret;
36393  __ret = (uint8x16_t)(__p0);
36394  return __ret;
36395 }
36396 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
36397  uint8x16_t __ret;
36398  __ret = (uint8x16_t)(__p0);
36399  return __ret;
36400 }
36401 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
36402  uint32x4_t __ret;
36403  __ret = (uint32x4_t)(__p0);
36404  return __ret;
36405 }
36406 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
36407  uint32x4_t __ret;
36408  __ret = (uint32x4_t)(__p0);
36409  return __ret;
36410 }
36411 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
36412  uint32x4_t __ret;
36413  __ret = (uint32x4_t)(__p0);
36414  return __ret;
36415 }
36416 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
36417  uint32x4_t __ret;
36418  __ret = (uint32x4_t)(__p0);
36419  return __ret;
36420 }
36421 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
36422  uint32x4_t __ret;
36423  __ret = (uint32x4_t)(__p0);
36424  return __ret;
36425 }
36426 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
36427  uint32x4_t __ret;
36428  __ret = (uint32x4_t)(__p0);
36429  return __ret;
36430 }
36431 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
36432  uint32x4_t __ret;
36433  __ret = (uint32x4_t)(__p0);
36434  return __ret;
36435 }
36436 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
36437  uint32x4_t __ret;
36438  __ret = (uint32x4_t)(__p0);
36439  return __ret;
36440 }
36441 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
36442  uint32x4_t __ret;
36443  __ret = (uint32x4_t)(__p0);
36444  return __ret;
36445 }
36446 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36447  uint32x4_t __ret;
36448  __ret = (uint32x4_t)(__p0);
36449  return __ret;
36450 }
36451 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36452  uint32x4_t __ret;
36453  __ret = (uint32x4_t)(__p0);
36454  return __ret;
36455 }
36456 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36457  uint32x4_t __ret;
36458  __ret = (uint32x4_t)(__p0);
36459  return __ret;
36460 }
36461 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36462  uint32x4_t __ret;
36463  __ret = (uint32x4_t)(__p0);
36464  return __ret;
36465 }
36466 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36467  uint32x4_t __ret;
36468  __ret = (uint32x4_t)(__p0);
36469  return __ret;
36470 }
36471 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36472  uint64x2_t __ret;
36473  __ret = (uint64x2_t)(__p0);
36474  return __ret;
36475 }
36476 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36477  uint64x2_t __ret;
36478  __ret = (uint64x2_t)(__p0);
36479  return __ret;
36480 }
36481 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36482  uint64x2_t __ret;
36483  __ret = (uint64x2_t)(__p0);
36484  return __ret;
36485 }
36486 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36487  uint64x2_t __ret;
36488  __ret = (uint64x2_t)(__p0);
36489  return __ret;
36490 }
36491 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36492  uint64x2_t __ret;
36493  __ret = (uint64x2_t)(__p0);
36494  return __ret;
36495 }
36496 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36497  uint64x2_t __ret;
36498  __ret = (uint64x2_t)(__p0);
36499  return __ret;
36500 }
36501 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36502  uint64x2_t __ret;
36503  __ret = (uint64x2_t)(__p0);
36504  return __ret;
36505 }
36506 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36507  uint64x2_t __ret;
36508  __ret = (uint64x2_t)(__p0);
36509  return __ret;
36510 }
36511 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36512  uint64x2_t __ret;
36513  __ret = (uint64x2_t)(__p0);
36514  return __ret;
36515 }
36516 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36517  uint64x2_t __ret;
36518  __ret = (uint64x2_t)(__p0);
36519  return __ret;
36520 }
36521 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36522  uint64x2_t __ret;
36523  __ret = (uint64x2_t)(__p0);
36524  return __ret;
36525 }
36526 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36527  uint64x2_t __ret;
36528  __ret = (uint64x2_t)(__p0);
36529  return __ret;
36530 }
36531 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36532  uint64x2_t __ret;
36533  __ret = (uint64x2_t)(__p0);
36534  return __ret;
36535 }
36536 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36537  uint64x2_t __ret;
36538  __ret = (uint64x2_t)(__p0);
36539  return __ret;
36540 }
36541 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36542  uint16x8_t __ret;
36543  __ret = (uint16x8_t)(__p0);
36544  return __ret;
36545 }
36546 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36547  uint16x8_t __ret;
36548  __ret = (uint16x8_t)(__p0);
36549  return __ret;
36550 }
36551 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36552  uint16x8_t __ret;
36553  __ret = (uint16x8_t)(__p0);
36554  return __ret;
36555 }
36556 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36557  uint16x8_t __ret;
36558  __ret = (uint16x8_t)(__p0);
36559  return __ret;
36560 }
36561 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36562  uint16x8_t __ret;
36563  __ret = (uint16x8_t)(__p0);
36564  return __ret;
36565 }
36566 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36567  uint16x8_t __ret;
36568  __ret = (uint16x8_t)(__p0);
36569  return __ret;
36570 }
36571 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36572  uint16x8_t __ret;
36573  __ret = (uint16x8_t)(__p0);
36574  return __ret;
36575 }
36576 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36577  uint16x8_t __ret;
36578  __ret = (uint16x8_t)(__p0);
36579  return __ret;
36580 }
36581 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36582  uint16x8_t __ret;
36583  __ret = (uint16x8_t)(__p0);
36584  return __ret;
36585 }
36586 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36587  uint16x8_t __ret;
36588  __ret = (uint16x8_t)(__p0);
36589  return __ret;
36590 }
36591 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36592  uint16x8_t __ret;
36593  __ret = (uint16x8_t)(__p0);
36594  return __ret;
36595 }
36596 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36597  uint16x8_t __ret;
36598  __ret = (uint16x8_t)(__p0);
36599  return __ret;
36600 }
36601 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36602  uint16x8_t __ret;
36603  __ret = (uint16x8_t)(__p0);
36604  return __ret;
36605 }
36606 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36607  uint16x8_t __ret;
36608  __ret = (uint16x8_t)(__p0);
36609  return __ret;
36610 }
36611 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36612  int8x16_t __ret;
36613  __ret = (int8x16_t)(__p0);
36614  return __ret;
36615 }
36616 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36617  int8x16_t __ret;
36618  __ret = (int8x16_t)(__p0);
36619  return __ret;
36620 }
36621 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36622  int8x16_t __ret;
36623  __ret = (int8x16_t)(__p0);
36624  return __ret;
36625 }
36626 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36627  int8x16_t __ret;
36628  __ret = (int8x16_t)(__p0);
36629  return __ret;
36630 }
36631 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36632  int8x16_t __ret;
36633  __ret = (int8x16_t)(__p0);
36634  return __ret;
36635 }
36636 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36637  int8x16_t __ret;
36638  __ret = (int8x16_t)(__p0);
36639  return __ret;
36640 }
36641 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36642  int8x16_t __ret;
36643  __ret = (int8x16_t)(__p0);
36644  return __ret;
36645 }
36646 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36647  int8x16_t __ret;
36648  __ret = (int8x16_t)(__p0);
36649  return __ret;
36650 }
36651 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36652  int8x16_t __ret;
36653  __ret = (int8x16_t)(__p0);
36654  return __ret;
36655 }
36656 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36657  int8x16_t __ret;
36658  __ret = (int8x16_t)(__p0);
36659  return __ret;
36660 }
36661 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36662  int8x16_t __ret;
36663  __ret = (int8x16_t)(__p0);
36664  return __ret;
36665 }
36666 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36667  int8x16_t __ret;
36668  __ret = (int8x16_t)(__p0);
36669  return __ret;
36670 }
36671 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36672  int8x16_t __ret;
36673  __ret = (int8x16_t)(__p0);
36674  return __ret;
36675 }
36676 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36677  int8x16_t __ret;
36678  __ret = (int8x16_t)(__p0);
36679  return __ret;
36680 }
36681 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36682  float64x2_t __ret;
36683  __ret = (float64x2_t)(__p0);
36684  return __ret;
36685 }
36686 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36687  float64x2_t __ret;
36688  __ret = (float64x2_t)(__p0);
36689  return __ret;
36690 }
36691 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36692  float64x2_t __ret;
36693  __ret = (float64x2_t)(__p0);
36694  return __ret;
36695 }
36696 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36697  float64x2_t __ret;
36698  __ret = (float64x2_t)(__p0);
36699  return __ret;
36700 }
36701 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36702  float64x2_t __ret;
36703  __ret = (float64x2_t)(__p0);
36704  return __ret;
36705 }
36706 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36707  float64x2_t __ret;
36708  __ret = (float64x2_t)(__p0);
36709  return __ret;
36710 }
36711 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36712  float64x2_t __ret;
36713  __ret = (float64x2_t)(__p0);
36714  return __ret;
36715 }
36716 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36717  float64x2_t __ret;
36718  __ret = (float64x2_t)(__p0);
36719  return __ret;
36720 }
36721 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36722  float64x2_t __ret;
36723  __ret = (float64x2_t)(__p0);
36724  return __ret;
36725 }
36726 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36727  float64x2_t __ret;
36728  __ret = (float64x2_t)(__p0);
36729  return __ret;
36730 }
36731 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36732  float64x2_t __ret;
36733  __ret = (float64x2_t)(__p0);
36734  return __ret;
36735 }
36736 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36737  float64x2_t __ret;
36738  __ret = (float64x2_t)(__p0);
36739  return __ret;
36740 }
36741 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36742  float64x2_t __ret;
36743  __ret = (float64x2_t)(__p0);
36744  return __ret;
36745 }
36746 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36747  float64x2_t __ret;
36748  __ret = (float64x2_t)(__p0);
36749  return __ret;
36750 }
36751 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36752  float32x4_t __ret;
36753  __ret = (float32x4_t)(__p0);
36754  return __ret;
36755 }
36756 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36757  float32x4_t __ret;
36758  __ret = (float32x4_t)(__p0);
36759  return __ret;
36760 }
36761 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36762  float32x4_t __ret;
36763  __ret = (float32x4_t)(__p0);
36764  return __ret;
36765 }
36766 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36767  float32x4_t __ret;
36768  __ret = (float32x4_t)(__p0);
36769  return __ret;
36770 }
36771 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36772  float32x4_t __ret;
36773  __ret = (float32x4_t)(__p0);
36774  return __ret;
36775 }
36776 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36777  float32x4_t __ret;
36778  __ret = (float32x4_t)(__p0);
36779  return __ret;
36780 }
36781 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36782  float32x4_t __ret;
36783  __ret = (float32x4_t)(__p0);
36784  return __ret;
36785 }
36786 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36787  float32x4_t __ret;
36788  __ret = (float32x4_t)(__p0);
36789  return __ret;
36790 }
36791 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
36792  float32x4_t __ret;
36793  __ret = (float32x4_t)(__p0);
36794  return __ret;
36795 }
36796 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
36797  float32x4_t __ret;
36798  __ret = (float32x4_t)(__p0);
36799  return __ret;
36800 }
36801 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
36802  float32x4_t __ret;
36803  __ret = (float32x4_t)(__p0);
36804  return __ret;
36805 }
36806 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
36807  float32x4_t __ret;
36808  __ret = (float32x4_t)(__p0);
36809  return __ret;
36810 }
36811 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
36812  float32x4_t __ret;
36813  __ret = (float32x4_t)(__p0);
36814  return __ret;
36815 }
36816 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
36817  float32x4_t __ret;
36818  __ret = (float32x4_t)(__p0);
36819  return __ret;
36820 }
36821 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
36822  float16x8_t __ret;
36823  __ret = (float16x8_t)(__p0);
36824  return __ret;
36825 }
36826 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
36827  float16x8_t __ret;
36828  __ret = (float16x8_t)(__p0);
36829  return __ret;
36830 }
36831 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
36832  float16x8_t __ret;
36833  __ret = (float16x8_t)(__p0);
36834  return __ret;
36835 }
36836 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
36837  float16x8_t __ret;
36838  __ret = (float16x8_t)(__p0);
36839  return __ret;
36840 }
36841 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
36842  float16x8_t __ret;
36843  __ret = (float16x8_t)(__p0);
36844  return __ret;
36845 }
36846 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
36847  float16x8_t __ret;
36848  __ret = (float16x8_t)(__p0);
36849  return __ret;
36850 }
36851 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
36852  float16x8_t __ret;
36853  __ret = (float16x8_t)(__p0);
36854  return __ret;
36855 }
36856 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
36857  float16x8_t __ret;
36858  __ret = (float16x8_t)(__p0);
36859  return __ret;
36860 }
36861 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
36862  float16x8_t __ret;
36863  __ret = (float16x8_t)(__p0);
36864  return __ret;
36865 }
36866 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
36867  float16x8_t __ret;
36868  __ret = (float16x8_t)(__p0);
36869  return __ret;
36870 }
36871 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
36872  float16x8_t __ret;
36873  __ret = (float16x8_t)(__p0);
36874  return __ret;
36875 }
36876 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
36877  float16x8_t __ret;
36878  __ret = (float16x8_t)(__p0);
36879  return __ret;
36880 }
36881 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
36882  float16x8_t __ret;
36883  __ret = (float16x8_t)(__p0);
36884  return __ret;
36885 }
36886 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
36887  float16x8_t __ret;
36888  __ret = (float16x8_t)(__p0);
36889  return __ret;
36890 }
36891 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
36892  int32x4_t __ret;
36893  __ret = (int32x4_t)(__p0);
36894  return __ret;
36895 }
36896 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
36897  int32x4_t __ret;
36898  __ret = (int32x4_t)(__p0);
36899  return __ret;
36900 }
36901 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
36902  int32x4_t __ret;
36903  __ret = (int32x4_t)(__p0);
36904  return __ret;
36905 }
36906 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
36907  int32x4_t __ret;
36908  __ret = (int32x4_t)(__p0);
36909  return __ret;
36910 }
36911 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
36912  int32x4_t __ret;
36913  __ret = (int32x4_t)(__p0);
36914  return __ret;
36915 }
36916 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
36917  int32x4_t __ret;
36918  __ret = (int32x4_t)(__p0);
36919  return __ret;
36920 }
36921 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
36922  int32x4_t __ret;
36923  __ret = (int32x4_t)(__p0);
36924  return __ret;
36925 }
36926 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
36927  int32x4_t __ret;
36928  __ret = (int32x4_t)(__p0);
36929  return __ret;
36930 }
36931 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
36932  int32x4_t __ret;
36933  __ret = (int32x4_t)(__p0);
36934  return __ret;
36935 }
36936 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
36937  int32x4_t __ret;
36938  __ret = (int32x4_t)(__p0);
36939  return __ret;
36940 }
36941 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
36942  int32x4_t __ret;
36943  __ret = (int32x4_t)(__p0);
36944  return __ret;
36945 }
36946 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
36947  int32x4_t __ret;
36948  __ret = (int32x4_t)(__p0);
36949  return __ret;
36950 }
36951 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
36952  int32x4_t __ret;
36953  __ret = (int32x4_t)(__p0);
36954  return __ret;
36955 }
36956 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
36957  int32x4_t __ret;
36958  __ret = (int32x4_t)(__p0);
36959  return __ret;
36960 }
36961 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
36962  int64x2_t __ret;
36963  __ret = (int64x2_t)(__p0);
36964  return __ret;
36965 }
36966 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
36967  int64x2_t __ret;
36968  __ret = (int64x2_t)(__p0);
36969  return __ret;
36970 }
36971 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
36972  int64x2_t __ret;
36973  __ret = (int64x2_t)(__p0);
36974  return __ret;
36975 }
36976 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
36977  int64x2_t __ret;
36978  __ret = (int64x2_t)(__p0);
36979  return __ret;
36980 }
36981 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
36982  int64x2_t __ret;
36983  __ret = (int64x2_t)(__p0);
36984  return __ret;
36985 }
36986 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
36987  int64x2_t __ret;
36988  __ret = (int64x2_t)(__p0);
36989  return __ret;
36990 }
36991 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
36992  int64x2_t __ret;
36993  __ret = (int64x2_t)(__p0);
36994  return __ret;
36995 }
36996 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
36997  int64x2_t __ret;
36998  __ret = (int64x2_t)(__p0);
36999  return __ret;
37000 }
37001 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37002  int64x2_t __ret;
37003  __ret = (int64x2_t)(__p0);
37004  return __ret;
37005 }
37006 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37007  int64x2_t __ret;
37008  __ret = (int64x2_t)(__p0);
37009  return __ret;
37010 }
37011 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37012  int64x2_t __ret;
37013  __ret = (int64x2_t)(__p0);
37014  return __ret;
37015 }
37016 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37017  int64x2_t __ret;
37018  __ret = (int64x2_t)(__p0);
37019  return __ret;
37020 }
37021 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37022  int64x2_t __ret;
37023  __ret = (int64x2_t)(__p0);
37024  return __ret;
37025 }
37026 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37027  int64x2_t __ret;
37028  __ret = (int64x2_t)(__p0);
37029  return __ret;
37030 }
37031 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37032  int16x8_t __ret;
37033  __ret = (int16x8_t)(__p0);
37034  return __ret;
37035 }
37036 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37037  int16x8_t __ret;
37038  __ret = (int16x8_t)(__p0);
37039  return __ret;
37040 }
37041 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37042  int16x8_t __ret;
37043  __ret = (int16x8_t)(__p0);
37044  return __ret;
37045 }
37046 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37047  int16x8_t __ret;
37048  __ret = (int16x8_t)(__p0);
37049  return __ret;
37050 }
37051 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37052  int16x8_t __ret;
37053  __ret = (int16x8_t)(__p0);
37054  return __ret;
37055 }
37056 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37057  int16x8_t __ret;
37058  __ret = (int16x8_t)(__p0);
37059  return __ret;
37060 }
37061 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37062  int16x8_t __ret;
37063  __ret = (int16x8_t)(__p0);
37064  return __ret;
37065 }
37066 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37067  int16x8_t __ret;
37068  __ret = (int16x8_t)(__p0);
37069  return __ret;
37070 }
37071 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37072  int16x8_t __ret;
37073  __ret = (int16x8_t)(__p0);
37074  return __ret;
37075 }
37076 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37077  int16x8_t __ret;
37078  __ret = (int16x8_t)(__p0);
37079  return __ret;
37080 }
37081 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37082  int16x8_t __ret;
37083  __ret = (int16x8_t)(__p0);
37084  return __ret;
37085 }
37086 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37087  int16x8_t __ret;
37088  __ret = (int16x8_t)(__p0);
37089  return __ret;
37090 }
37091 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37092  int16x8_t __ret;
37093  __ret = (int16x8_t)(__p0);
37094  return __ret;
37095 }
37096 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37097  int16x8_t __ret;
37098  __ret = (int16x8_t)(__p0);
37099  return __ret;
37100 }
37101 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37102  uint8x8_t __ret;
37103  __ret = (uint8x8_t)(__p0);
37104  return __ret;
37105 }
37106 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37107  uint8x8_t __ret;
37108  __ret = (uint8x8_t)(__p0);
37109  return __ret;
37110 }
37111 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37112  uint8x8_t __ret;
37113  __ret = (uint8x8_t)(__p0);
37114  return __ret;
37115 }
37116 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37117  uint8x8_t __ret;
37118  __ret = (uint8x8_t)(__p0);
37119  return __ret;
37120 }
37121 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37122  uint8x8_t __ret;
37123  __ret = (uint8x8_t)(__p0);
37124  return __ret;
37125 }
37126 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37127  uint8x8_t __ret;
37128  __ret = (uint8x8_t)(__p0);
37129  return __ret;
37130 }
37131 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37132  uint8x8_t __ret;
37133  __ret = (uint8x8_t)(__p0);
37134  return __ret;
37135 }
37136 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37137  uint8x8_t __ret;
37138  __ret = (uint8x8_t)(__p0);
37139  return __ret;
37140 }
37141 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37142  uint8x8_t __ret;
37143  __ret = (uint8x8_t)(__p0);
37144  return __ret;
37145 }
37146 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
37147  uint8x8_t __ret;
37148  __ret = (uint8x8_t)(__p0);
37149  return __ret;
37150 }
37151 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
37152  uint8x8_t __ret;
37153  __ret = (uint8x8_t)(__p0);
37154  return __ret;
37155 }
37156 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
37157  uint8x8_t __ret;
37158  __ret = (uint8x8_t)(__p0);
37159  return __ret;
37160 }
37161 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
37162  uint8x8_t __ret;
37163  __ret = (uint8x8_t)(__p0);
37164  return __ret;
37165 }
37166 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
37167  uint32x2_t __ret;
37168  __ret = (uint32x2_t)(__p0);
37169  return __ret;
37170 }
37171 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
37172  uint32x2_t __ret;
37173  __ret = (uint32x2_t)(__p0);
37174  return __ret;
37175 }
37176 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
37177  uint32x2_t __ret;
37178  __ret = (uint32x2_t)(__p0);
37179  return __ret;
37180 }
37181 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
37182  uint32x2_t __ret;
37183  __ret = (uint32x2_t)(__p0);
37184  return __ret;
37185 }
37186 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
37187  uint32x2_t __ret;
37188  __ret = (uint32x2_t)(__p0);
37189  return __ret;
37190 }
37191 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
37192  uint32x2_t __ret;
37193  __ret = (uint32x2_t)(__p0);
37194  return __ret;
37195 }
37196 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
37197  uint32x2_t __ret;
37198  __ret = (uint32x2_t)(__p0);
37199  return __ret;
37200 }
37201 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
37202  uint32x2_t __ret;
37203  __ret = (uint32x2_t)(__p0);
37204  return __ret;
37205 }
37206 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
37207  uint32x2_t __ret;
37208  __ret = (uint32x2_t)(__p0);
37209  return __ret;
37210 }
37211 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
37212  uint32x2_t __ret;
37213  __ret = (uint32x2_t)(__p0);
37214  return __ret;
37215 }
37216 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
37217  uint32x2_t __ret;
37218  __ret = (uint32x2_t)(__p0);
37219  return __ret;
37220 }
37221 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
37222  uint32x2_t __ret;
37223  __ret = (uint32x2_t)(__p0);
37224  return __ret;
37225 }
37226 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
37227  uint32x2_t __ret;
37228  __ret = (uint32x2_t)(__p0);
37229  return __ret;
37230 }
37231 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
37232  uint64x1_t __ret;
37233  __ret = (uint64x1_t)(__p0);
37234  return __ret;
37235 }
37236 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
37237  uint64x1_t __ret;
37238  __ret = (uint64x1_t)(__p0);
37239  return __ret;
37240 }
37241 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
37242  uint64x1_t __ret;
37243  __ret = (uint64x1_t)(__p0);
37244  return __ret;
37245 }
37246 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
37247  uint64x1_t __ret;
37248  __ret = (uint64x1_t)(__p0);
37249  return __ret;
37250 }
37251 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
37252  uint64x1_t __ret;
37253  __ret = (uint64x1_t)(__p0);
37254  return __ret;
37255 }
37256 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
37257  uint64x1_t __ret;
37258  __ret = (uint64x1_t)(__p0);
37259  return __ret;
37260 }
37261 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
37262  uint64x1_t __ret;
37263  __ret = (uint64x1_t)(__p0);
37264  return __ret;
37265 }
37266 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
37267  uint64x1_t __ret;
37268  __ret = (uint64x1_t)(__p0);
37269  return __ret;
37270 }
37271 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
37272  uint64x1_t __ret;
37273  __ret = (uint64x1_t)(__p0);
37274  return __ret;
37275 }
37276 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
37277  uint64x1_t __ret;
37278  __ret = (uint64x1_t)(__p0);
37279  return __ret;
37280 }
37281 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
37282  uint64x1_t __ret;
37283  __ret = (uint64x1_t)(__p0);
37284  return __ret;
37285 }
37286 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
37287  uint64x1_t __ret;
37288  __ret = (uint64x1_t)(__p0);
37289  return __ret;
37290 }
37291 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
37292  uint64x1_t __ret;
37293  __ret = (uint64x1_t)(__p0);
37294  return __ret;
37295 }
37296 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
37297  uint16x4_t __ret;
37298  __ret = (uint16x4_t)(__p0);
37299  return __ret;
37300 }
37301 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
37302  uint16x4_t __ret;
37303  __ret = (uint16x4_t)(__p0);
37304  return __ret;
37305 }
37306 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
37307  uint16x4_t __ret;
37308  __ret = (uint16x4_t)(__p0);
37309  return __ret;
37310 }
37311 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
37312  uint16x4_t __ret;
37313  __ret = (uint16x4_t)(__p0);
37314  return __ret;
37315 }
37316 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
37317  uint16x4_t __ret;
37318  __ret = (uint16x4_t)(__p0);
37319  return __ret;
37320 }
37321 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
37322  uint16x4_t __ret;
37323  __ret = (uint16x4_t)(__p0);
37324  return __ret;
37325 }
37326 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
37327  uint16x4_t __ret;
37328  __ret = (uint16x4_t)(__p0);
37329  return __ret;
37330 }
37331 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
37332  uint16x4_t __ret;
37333  __ret = (uint16x4_t)(__p0);
37334  return __ret;
37335 }
37336 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
37337  uint16x4_t __ret;
37338  __ret = (uint16x4_t)(__p0);
37339  return __ret;
37340 }
37341 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
37342  uint16x4_t __ret;
37343  __ret = (uint16x4_t)(__p0);
37344  return __ret;
37345 }
37346 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
37347  uint16x4_t __ret;
37348  __ret = (uint16x4_t)(__p0);
37349  return __ret;
37350 }
37351 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
37352  uint16x4_t __ret;
37353  __ret = (uint16x4_t)(__p0);
37354  return __ret;
37355 }
37356 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
37357  uint16x4_t __ret;
37358  __ret = (uint16x4_t)(__p0);
37359  return __ret;
37360 }
37361 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
37362  int8x8_t __ret;
37363  __ret = (int8x8_t)(__p0);
37364  return __ret;
37365 }
37366 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
37367  int8x8_t __ret;
37368  __ret = (int8x8_t)(__p0);
37369  return __ret;
37370 }
37371 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
37372  int8x8_t __ret;
37373  __ret = (int8x8_t)(__p0);
37374  return __ret;
37375 }
37376 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
37377  int8x8_t __ret;
37378  __ret = (int8x8_t)(__p0);
37379  return __ret;
37380 }
37381 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
37382  int8x8_t __ret;
37383  __ret = (int8x8_t)(__p0);
37384  return __ret;
37385 }
37386 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
37387  int8x8_t __ret;
37388  __ret = (int8x8_t)(__p0);
37389  return __ret;
37390 }
37391 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
37392  int8x8_t __ret;
37393  __ret = (int8x8_t)(__p0);
37394  return __ret;
37395 }
37396 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
37397  int8x8_t __ret;
37398  __ret = (int8x8_t)(__p0);
37399  return __ret;
37400 }
37401 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
37402  int8x8_t __ret;
37403  __ret = (int8x8_t)(__p0);
37404  return __ret;
37405 }
37406 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
37407  int8x8_t __ret;
37408  __ret = (int8x8_t)(__p0);
37409  return __ret;
37410 }
37411 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
37412  int8x8_t __ret;
37413  __ret = (int8x8_t)(__p0);
37414  return __ret;
37415 }
37416 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
37417  int8x8_t __ret;
37418  __ret = (int8x8_t)(__p0);
37419  return __ret;
37420 }
37421 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
37422  int8x8_t __ret;
37423  __ret = (int8x8_t)(__p0);
37424  return __ret;
37425 }
37426 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
37427  float64x1_t __ret;
37428  __ret = (float64x1_t)(__p0);
37429  return __ret;
37430 }
37431 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
37432  float64x1_t __ret;
37433  __ret = (float64x1_t)(__p0);
37434  return __ret;
37435 }
37436 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
37437  float64x1_t __ret;
37438  __ret = (float64x1_t)(__p0);
37439  return __ret;
37440 }
37441 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
37442  float64x1_t __ret;
37443  __ret = (float64x1_t)(__p0);
37444  return __ret;
37445 }
37446 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
37447  float64x1_t __ret;
37448  __ret = (float64x1_t)(__p0);
37449  return __ret;
37450 }
37451 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
37452  float64x1_t __ret;
37453  __ret = (float64x1_t)(__p0);
37454  return __ret;
37455 }
37456 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
37457  float64x1_t __ret;
37458  __ret = (float64x1_t)(__p0);
37459  return __ret;
37460 }
37461 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
37462  float64x1_t __ret;
37463  __ret = (float64x1_t)(__p0);
37464  return __ret;
37465 }
37466 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
37467  float64x1_t __ret;
37468  __ret = (float64x1_t)(__p0);
37469  return __ret;
37470 }
37471 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
37472  float64x1_t __ret;
37473  __ret = (float64x1_t)(__p0);
37474  return __ret;
37475 }
37476 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
37477  float64x1_t __ret;
37478  __ret = (float64x1_t)(__p0);
37479  return __ret;
37480 }
37481 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
37482  float64x1_t __ret;
37483  __ret = (float64x1_t)(__p0);
37484  return __ret;
37485 }
37486 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
37487  float64x1_t __ret;
37488  __ret = (float64x1_t)(__p0);
37489  return __ret;
37490 }
37491 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
37492  float32x2_t __ret;
37493  __ret = (float32x2_t)(__p0);
37494  return __ret;
37495 }
37496 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
37497  float32x2_t __ret;
37498  __ret = (float32x2_t)(__p0);
37499  return __ret;
37500 }
37501 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
37502  float32x2_t __ret;
37503  __ret = (float32x2_t)(__p0);
37504  return __ret;
37505 }
37506 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
37507  float32x2_t __ret;
37508  __ret = (float32x2_t)(__p0);
37509  return __ret;
37510 }
37511 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
37512  float32x2_t __ret;
37513  __ret = (float32x2_t)(__p0);
37514  return __ret;
37515 }
37516 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
37517  float32x2_t __ret;
37518  __ret = (float32x2_t)(__p0);
37519  return __ret;
37520 }
37521 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
37522  float32x2_t __ret;
37523  __ret = (float32x2_t)(__p0);
37524  return __ret;
37525 }
37526 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
37527  float32x2_t __ret;
37528  __ret = (float32x2_t)(__p0);
37529  return __ret;
37530 }
37531 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
37532  float32x2_t __ret;
37533  __ret = (float32x2_t)(__p0);
37534  return __ret;
37535 }
37536 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
37537  float32x2_t __ret;
37538  __ret = (float32x2_t)(__p0);
37539  return __ret;
37540 }
37541 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
37542  float32x2_t __ret;
37543  __ret = (float32x2_t)(__p0);
37544  return __ret;
37545 }
37546 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
37547  float32x2_t __ret;
37548  __ret = (float32x2_t)(__p0);
37549  return __ret;
37550 }
37551 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
37552  float32x2_t __ret;
37553  __ret = (float32x2_t)(__p0);
37554  return __ret;
37555 }
37556 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
37557  float16x4_t __ret;
37558  __ret = (float16x4_t)(__p0);
37559  return __ret;
37560 }
37561 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
37562  float16x4_t __ret;
37563  __ret = (float16x4_t)(__p0);
37564  return __ret;
37565 }
37566 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
37567  float16x4_t __ret;
37568  __ret = (float16x4_t)(__p0);
37569  return __ret;
37570 }
37571 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
37572  float16x4_t __ret;
37573  __ret = (float16x4_t)(__p0);
37574  return __ret;
37575 }
37576 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
37577  float16x4_t __ret;
37578  __ret = (float16x4_t)(__p0);
37579  return __ret;
37580 }
37581 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
37582  float16x4_t __ret;
37583  __ret = (float16x4_t)(__p0);
37584  return __ret;
37585 }
37586 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
37587  float16x4_t __ret;
37588  __ret = (float16x4_t)(__p0);
37589  return __ret;
37590 }
37591 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
37592  float16x4_t __ret;
37593  __ret = (float16x4_t)(__p0);
37594  return __ret;
37595 }
37596 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
37597  float16x4_t __ret;
37598  __ret = (float16x4_t)(__p0);
37599  return __ret;
37600 }
37601 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
37602  float16x4_t __ret;
37603  __ret = (float16x4_t)(__p0);
37604  return __ret;
37605 }
37606 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
37607  float16x4_t __ret;
37608  __ret = (float16x4_t)(__p0);
37609  return __ret;
37610 }
37611 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
37612  float16x4_t __ret;
37613  __ret = (float16x4_t)(__p0);
37614  return __ret;
37615 }
37616 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
37617  float16x4_t __ret;
37618  __ret = (float16x4_t)(__p0);
37619  return __ret;
37620 }
37621 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
37622  int32x2_t __ret;
37623  __ret = (int32x2_t)(__p0);
37624  return __ret;
37625 }
37626 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
37627  int32x2_t __ret;
37628  __ret = (int32x2_t)(__p0);
37629  return __ret;
37630 }
37631 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
37632  int32x2_t __ret;
37633  __ret = (int32x2_t)(__p0);
37634  return __ret;
37635 }
37636 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
37637  int32x2_t __ret;
37638  __ret = (int32x2_t)(__p0);
37639  return __ret;
37640 }
37641 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
37642  int32x2_t __ret;
37643  __ret = (int32x2_t)(__p0);
37644  return __ret;
37645 }
37646 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
37647  int32x2_t __ret;
37648  __ret = (int32x2_t)(__p0);
37649  return __ret;
37650 }
37651 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
37652  int32x2_t __ret;
37653  __ret = (int32x2_t)(__p0);
37654  return __ret;
37655 }
37656 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
37657  int32x2_t __ret;
37658  __ret = (int32x2_t)(__p0);
37659  return __ret;
37660 }
37661 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
37662  int32x2_t __ret;
37663  __ret = (int32x2_t)(__p0);
37664  return __ret;
37665 }
37666 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
37667  int32x2_t __ret;
37668  __ret = (int32x2_t)(__p0);
37669  return __ret;
37670 }
37671 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
37672  int32x2_t __ret;
37673  __ret = (int32x2_t)(__p0);
37674  return __ret;
37675 }
37676 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
37677  int32x2_t __ret;
37678  __ret = (int32x2_t)(__p0);
37679  return __ret;
37680 }
37681 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
37682  int32x2_t __ret;
37683  __ret = (int32x2_t)(__p0);
37684  return __ret;
37685 }
37686 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
37687  int64x1_t __ret;
37688  __ret = (int64x1_t)(__p0);
37689  return __ret;
37690 }
37691 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
37692  int64x1_t __ret;
37693  __ret = (int64x1_t)(__p0);
37694  return __ret;
37695 }
37696 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
37697  int64x1_t __ret;
37698  __ret = (int64x1_t)(__p0);
37699  return __ret;
37700 }
37701 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
37702  int64x1_t __ret;
37703  __ret = (int64x1_t)(__p0);
37704  return __ret;
37705 }
37706 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
37707  int64x1_t __ret;
37708  __ret = (int64x1_t)(__p0);
37709  return __ret;
37710 }
37711 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
37712  int64x1_t __ret;
37713  __ret = (int64x1_t)(__p0);
37714  return __ret;
37715 }
37716 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
37717  int64x1_t __ret;
37718  __ret = (int64x1_t)(__p0);
37719  return __ret;
37720 }
37721 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
37722  int64x1_t __ret;
37723  __ret = (int64x1_t)(__p0);
37724  return __ret;
37725 }
37726 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
37727  int64x1_t __ret;
37728  __ret = (int64x1_t)(__p0);
37729  return __ret;
37730 }
37731 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
37732  int64x1_t __ret;
37733  __ret = (int64x1_t)(__p0);
37734  return __ret;
37735 }
37736 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
37737  int64x1_t __ret;
37738  __ret = (int64x1_t)(__p0);
37739  return __ret;
37740 }
37741 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
37742  int64x1_t __ret;
37743  __ret = (int64x1_t)(__p0);
37744  return __ret;
37745 }
37746 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
37747  int64x1_t __ret;
37748  __ret = (int64x1_t)(__p0);
37749  return __ret;
37750 }
37751 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
37752  int16x4_t __ret;
37753  __ret = (int16x4_t)(__p0);
37754  return __ret;
37755 }
37756 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
37757  int16x4_t __ret;
37758  __ret = (int16x4_t)(__p0);
37759  return __ret;
37760 }
37761 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
37762  int16x4_t __ret;
37763  __ret = (int16x4_t)(__p0);
37764  return __ret;
37765 }
37766 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
37767  int16x4_t __ret;
37768  __ret = (int16x4_t)(__p0);
37769  return __ret;
37770 }
37771 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
37772  int16x4_t __ret;
37773  __ret = (int16x4_t)(__p0);
37774  return __ret;
37775 }
37776 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
37777  int16x4_t __ret;
37778  __ret = (int16x4_t)(__p0);
37779  return __ret;
37780 }
37781 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
37782  int16x4_t __ret;
37783  __ret = (int16x4_t)(__p0);
37784  return __ret;
37785 }
37786 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
37787  int16x4_t __ret;
37788  __ret = (int16x4_t)(__p0);
37789  return __ret;
37790 }
37791 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
37792  int16x4_t __ret;
37793  __ret = (int16x4_t)(__p0);
37794  return __ret;
37795 }
37796 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
37797  int16x4_t __ret;
37798  __ret = (int16x4_t)(__p0);
37799  return __ret;
37800 }
37801 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
37802  int16x4_t __ret;
37803  __ret = (int16x4_t)(__p0);
37804  return __ret;
37805 }
37806 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
37807  int16x4_t __ret;
37808  __ret = (int16x4_t)(__p0);
37809  return __ret;
37810 }
37811 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
37812  int16x4_t __ret;
37813  __ret = (int16x4_t)(__p0);
37814  return __ret;
37815 }
37816 #endif
37817 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
37818 #ifdef __LITTLE_ENDIAN__
37819 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
37820  float64x2_t __ret;
37821  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
37822  return __ret;
37823 }
37824 #else
37825 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
37826  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37827  float64x2_t __ret;
37828  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
37829  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37830  return __ret;
37831 }
37832 #endif
37833 
37834 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
37835  float64x1_t __ret;
37836  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
37837  return __ret;
37838 }
37839 #ifdef __LITTLE_ENDIAN__
37840 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
37841  float64x2_t __ret;
37842  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
37843  return __ret;
37844 }
37845 #else
37846 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
37847  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37848  float64x2_t __ret;
37849  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
37850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37851  return __ret;
37852 }
37853 #endif
37854 
37855 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
37856  float64x1_t __ret;
37857  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
37858  return __ret;
37859 }
37860 #ifdef __LITTLE_ENDIAN__
37861 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
37862  float64x2_t __ret;
37863  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
37864  return __ret;
37865 }
37866 #else
37867 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
37868  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37869  float64x2_t __ret;
37870  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
37871  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37872  return __ret;
37873 }
37874 #endif
37875 
37876 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
37877  float64x1_t __ret;
37878  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
37879  return __ret;
37880 }
37881 #ifdef __LITTLE_ENDIAN__
37882 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
37883  float64x2_t __ret;
37884  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
37885  return __ret;
37886 }
37887 #else
37888 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
37889  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37890  float64x2_t __ret;
37891  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
37892  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37893  return __ret;
37894 }
37895 #endif
37896 
37897 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
37898  float64x1_t __ret;
37899  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
37900  return __ret;
37901 }
37902 #ifdef __LITTLE_ENDIAN__
37903 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
37904  float64x2_t __ret;
37905  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
37906  return __ret;
37907 }
37908 #else
37909 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
37910  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37911  float64x2_t __ret;
37912  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
37913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37914  return __ret;
37915 }
37916 #endif
37917 
37918 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
37919  float64x1_t __ret;
37920  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
37921  return __ret;
37922 }
37923 #ifdef __LITTLE_ENDIAN__
37924 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
37925  float64x2_t __ret;
37926  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
37927  return __ret;
37928 }
37929 #else
37930 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
37931  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37932  float64x2_t __ret;
37933  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
37934  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37935  return __ret;
37936 }
37937 #endif
37938 
37939 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
37940  float64x1_t __ret;
37941  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
37942  return __ret;
37943 }
37944 #ifdef __LITTLE_ENDIAN__
37945 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
37946  float64x2_t __ret;
37947  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
37948  return __ret;
37949 }
37950 #else
37951 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
37952  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37953  float64x2_t __ret;
37954  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
37955  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37956  return __ret;
37957 }
37958 #endif
37959 
37960 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
37961  float64x1_t __ret;
37962  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
37963  return __ret;
37964 }
37965 #endif
37966 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
37967 #ifdef __LITTLE_ENDIAN__
37968 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
37969  float64x2_t __ret;
37970  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
37971  return __ret;
37972 }
37973 #else
37974 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
37975  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37976  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
37977  float64x2_t __ret;
37978  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
37979  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
37980  return __ret;
37981 }
37982 #endif
37983 
37984 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
37985  float64x1_t __ret;
37986  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
37987  return __ret;
37988 }
37989 #ifdef __LITTLE_ENDIAN__
37990 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
37991  float64x2_t __ret;
37992  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
37993  return __ret;
37994 }
37995 #else
37996 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
37997  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
37998  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
37999  float64x2_t __ret;
38000  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
38001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
38002  return __ret;
38003 }
38004 #endif
38005 
38006 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
38007  float64x1_t __ret;
38008  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
38009  return __ret;
38010 }
38011 #endif
38012 #if defined(__ARM_FEATURE_BF16) && !defined(__aarch64__)
38013 __ai poly8x8_t vreinterpret_p8_bf16(bfloat16x4_t __p0) {
38014  poly8x8_t __ret;
38015  __ret = (poly8x8_t)(__p0);
38016  return __ret;
38017 }
38018 __ai poly64x1_t vreinterpret_p64_bf16(bfloat16x4_t __p0) {
38019  poly64x1_t __ret;
38020  __ret = (poly64x1_t)(__p0);
38021  return __ret;
38022 }
38023 __ai poly16x4_t vreinterpret_p16_bf16(bfloat16x4_t __p0) {
38024  poly16x4_t __ret;
38025  __ret = (poly16x4_t)(__p0);
38026  return __ret;
38027 }
38028 __ai poly8x16_t vreinterpretq_p8_bf16(bfloat16x8_t __p0) {
38029  poly8x16_t __ret;
38030  __ret = (poly8x16_t)(__p0);
38031  return __ret;
38032 }
38033 __ai poly64x2_t vreinterpretq_p64_bf16(bfloat16x8_t __p0) {
38034  poly64x2_t __ret;
38035  __ret = (poly64x2_t)(__p0);
38036  return __ret;
38037 }
38038 __ai poly16x8_t vreinterpretq_p16_bf16(bfloat16x8_t __p0) {
38039  poly16x8_t __ret;
38040  __ret = (poly16x8_t)(__p0);
38041  return __ret;
38042 }
38043 __ai uint8x16_t vreinterpretq_u8_bf16(bfloat16x8_t __p0) {
38044  uint8x16_t __ret;
38045  __ret = (uint8x16_t)(__p0);
38046  return __ret;
38047 }
38048 __ai uint32x4_t vreinterpretq_u32_bf16(bfloat16x8_t __p0) {
38049  uint32x4_t __ret;
38050  __ret = (uint32x4_t)(__p0);
38051  return __ret;
38052 }
38053 __ai uint64x2_t vreinterpretq_u64_bf16(bfloat16x8_t __p0) {
38054  uint64x2_t __ret;
38055  __ret = (uint64x2_t)(__p0);
38056  return __ret;
38057 }
38058 __ai uint16x8_t vreinterpretq_u16_bf16(bfloat16x8_t __p0) {
38059  uint16x8_t __ret;
38060  __ret = (uint16x8_t)(__p0);
38061  return __ret;
38062 }
38063 __ai int8x16_t vreinterpretq_s8_bf16(bfloat16x8_t __p0) {
38064  int8x16_t __ret;
38065  __ret = (int8x16_t)(__p0);
38066  return __ret;
38067 }
38068 __ai float32x4_t vreinterpretq_f32_bf16(bfloat16x8_t __p0) {
38069  float32x4_t __ret;
38070  __ret = (float32x4_t)(__p0);
38071  return __ret;
38072 }
38073 __ai float16x8_t vreinterpretq_f16_bf16(bfloat16x8_t __p0) {
38074  float16x8_t __ret;
38075  __ret = (float16x8_t)(__p0);
38076  return __ret;
38077 }
38078 __ai int32x4_t vreinterpretq_s32_bf16(bfloat16x8_t __p0) {
38079  int32x4_t __ret;
38080  __ret = (int32x4_t)(__p0);
38081  return __ret;
38082 }
38083 __ai int64x2_t vreinterpretq_s64_bf16(bfloat16x8_t __p0) {
38084  int64x2_t __ret;
38085  __ret = (int64x2_t)(__p0);
38086  return __ret;
38087 }
38088 __ai int16x8_t vreinterpretq_s16_bf16(bfloat16x8_t __p0) {
38089  int16x8_t __ret;
38090  __ret = (int16x8_t)(__p0);
38091  return __ret;
38092 }
38093 __ai uint8x8_t vreinterpret_u8_bf16(bfloat16x4_t __p0) {
38094  uint8x8_t __ret;
38095  __ret = (uint8x8_t)(__p0);
38096  return __ret;
38097 }
38098 __ai uint32x2_t vreinterpret_u32_bf16(bfloat16x4_t __p0) {
38099  uint32x2_t __ret;
38100  __ret = (uint32x2_t)(__p0);
38101  return __ret;
38102 }
38103 __ai uint64x1_t vreinterpret_u64_bf16(bfloat16x4_t __p0) {
38104  uint64x1_t __ret;
38105  __ret = (uint64x1_t)(__p0);
38106  return __ret;
38107 }
38108 __ai uint16x4_t vreinterpret_u16_bf16(bfloat16x4_t __p0) {
38109  uint16x4_t __ret;
38110  __ret = (uint16x4_t)(__p0);
38111  return __ret;
38112 }
38113 __ai int8x8_t vreinterpret_s8_bf16(bfloat16x4_t __p0) {
38114  int8x8_t __ret;
38115  __ret = (int8x8_t)(__p0);
38116  return __ret;
38117 }
38118 __ai float32x2_t vreinterpret_f32_bf16(bfloat16x4_t __p0) {
38119  float32x2_t __ret;
38120  __ret = (float32x2_t)(__p0);
38121  return __ret;
38122 }
38123 __ai float16x4_t vreinterpret_f16_bf16(bfloat16x4_t __p0) {
38124  float16x4_t __ret;
38125  __ret = (float16x4_t)(__p0);
38126  return __ret;
38127 }
38128 __ai int32x2_t vreinterpret_s32_bf16(bfloat16x4_t __p0) {
38129  int32x2_t __ret;
38130  __ret = (int32x2_t)(__p0);
38131  return __ret;
38132 }
38133 __ai int64x1_t vreinterpret_s64_bf16(bfloat16x4_t __p0) {
38134  int64x1_t __ret;
38135  __ret = (int64x1_t)(__p0);
38136  return __ret;
38137 }
38138 __ai int16x4_t vreinterpret_s16_bf16(bfloat16x4_t __p0) {
38139  int16x4_t __ret;
38140  __ret = (int16x4_t)(__p0);
38141  return __ret;
38142 }
38143 __ai bfloat16x8_t vreinterpretq_bf16_p8(poly8x16_t __p0) {
38144  bfloat16x8_t __ret;
38145  __ret = (bfloat16x8_t)(__p0);
38146  return __ret;
38147 }
38148 __ai bfloat16x8_t vreinterpretq_bf16_p64(poly64x2_t __p0) {
38149  bfloat16x8_t __ret;
38150  __ret = (bfloat16x8_t)(__p0);
38151  return __ret;
38152 }
38153 __ai bfloat16x8_t vreinterpretq_bf16_p16(poly16x8_t __p0) {
38154  bfloat16x8_t __ret;
38155  __ret = (bfloat16x8_t)(__p0);
38156  return __ret;
38157 }
38158 __ai bfloat16x8_t vreinterpretq_bf16_u8(uint8x16_t __p0) {
38159  bfloat16x8_t __ret;
38160  __ret = (bfloat16x8_t)(__p0);
38161  return __ret;
38162 }
38163 __ai bfloat16x8_t vreinterpretq_bf16_u32(uint32x4_t __p0) {
38164  bfloat16x8_t __ret;
38165  __ret = (bfloat16x8_t)(__p0);
38166  return __ret;
38167 }
38168 __ai bfloat16x8_t vreinterpretq_bf16_u64(uint64x2_t __p0) {
38169  bfloat16x8_t __ret;
38170  __ret = (bfloat16x8_t)(__p0);
38171  return __ret;
38172 }
38173 __ai bfloat16x8_t vreinterpretq_bf16_u16(uint16x8_t __p0) {
38174  bfloat16x8_t __ret;
38175  __ret = (bfloat16x8_t)(__p0);
38176  return __ret;
38177 }
38178 __ai bfloat16x8_t vreinterpretq_bf16_s8(int8x16_t __p0) {
38179  bfloat16x8_t __ret;
38180  __ret = (bfloat16x8_t)(__p0);
38181  return __ret;
38182 }
38183 __ai bfloat16x8_t vreinterpretq_bf16_f32(float32x4_t __p0) {
38184  bfloat16x8_t __ret;
38185  __ret = (bfloat16x8_t)(__p0);
38186  return __ret;
38187 }
38188 __ai bfloat16x8_t vreinterpretq_bf16_f16(float16x8_t __p0) {
38189  bfloat16x8_t __ret;
38190  __ret = (bfloat16x8_t)(__p0);
38191  return __ret;
38192 }
38193 __ai bfloat16x8_t vreinterpretq_bf16_s32(int32x4_t __p0) {
38194  bfloat16x8_t __ret;
38195  __ret = (bfloat16x8_t)(__p0);
38196  return __ret;
38197 }
38198 __ai bfloat16x8_t vreinterpretq_bf16_s64(int64x2_t __p0) {
38199  bfloat16x8_t __ret;
38200  __ret = (bfloat16x8_t)(__p0);
38201  return __ret;
38202 }
38203 __ai bfloat16x8_t vreinterpretq_bf16_s16(int16x8_t __p0) {
38204  bfloat16x8_t __ret;
38205  __ret = (bfloat16x8_t)(__p0);
38206  return __ret;
38207 }
38208 __ai bfloat16x4_t vreinterpret_bf16_p8(poly8x8_t __p0) {
38209  bfloat16x4_t __ret;
38210  __ret = (bfloat16x4_t)(__p0);
38211  return __ret;
38212 }
38213 __ai bfloat16x4_t vreinterpret_bf16_p64(poly64x1_t __p0) {
38214  bfloat16x4_t __ret;
38215  __ret = (bfloat16x4_t)(__p0);
38216  return __ret;
38217 }
38218 __ai bfloat16x4_t vreinterpret_bf16_p16(poly16x4_t __p0) {
38219  bfloat16x4_t __ret;
38220  __ret = (bfloat16x4_t)(__p0);
38221  return __ret;
38222 }
38223 __ai bfloat16x4_t vreinterpret_bf16_u8(uint8x8_t __p0) {
38224  bfloat16x4_t __ret;
38225  __ret = (bfloat16x4_t)(__p0);
38226  return __ret;
38227 }
38228 __ai bfloat16x4_t vreinterpret_bf16_u32(uint32x2_t __p0) {
38229  bfloat16x4_t __ret;
38230  __ret = (bfloat16x4_t)(__p0);
38231  return __ret;
38232 }
38233 __ai bfloat16x4_t vreinterpret_bf16_u64(uint64x1_t __p0) {
38234  bfloat16x4_t __ret;
38235  __ret = (bfloat16x4_t)(__p0);
38236  return __ret;
38237 }
38238 __ai bfloat16x4_t vreinterpret_bf16_u16(uint16x4_t __p0) {
38239  bfloat16x4_t __ret;
38240  __ret = (bfloat16x4_t)(__p0);
38241  return __ret;
38242 }
38243 __ai bfloat16x4_t vreinterpret_bf16_s8(int8x8_t __p0) {
38244  bfloat16x4_t __ret;
38245  __ret = (bfloat16x4_t)(__p0);
38246  return __ret;
38247 }
38248 __ai bfloat16x4_t vreinterpret_bf16_f32(float32x2_t __p0) {
38249  bfloat16x4_t __ret;
38250  __ret = (bfloat16x4_t)(__p0);
38251  return __ret;
38252 }
38253 __ai bfloat16x4_t vreinterpret_bf16_f16(float16x4_t __p0) {
38254  bfloat16x4_t __ret;
38255  __ret = (bfloat16x4_t)(__p0);
38256  return __ret;
38257 }
38258 __ai bfloat16x4_t vreinterpret_bf16_s32(int32x2_t __p0) {
38259  bfloat16x4_t __ret;
38260  __ret = (bfloat16x4_t)(__p0);
38261  return __ret;
38262 }
38263 __ai bfloat16x4_t vreinterpret_bf16_s64(int64x1_t __p0) {
38264  bfloat16x4_t __ret;
38265  __ret = (bfloat16x4_t)(__p0);
38266  return __ret;
38267 }
38268 __ai bfloat16x4_t vreinterpret_bf16_s16(int16x4_t __p0) {
38269  bfloat16x4_t __ret;
38270  __ret = (bfloat16x4_t)(__p0);
38271  return __ret;
38272 }
38273 #endif
38274 #if defined(__ARM_FEATURE_BF16) && defined(__aarch64__)
38275 __ai poly8x8_t vreinterpret_p8_bf16(bfloat16x4_t __p0) {
38276  poly8x8_t __ret;
38277  __ret = (poly8x8_t)(__p0);
38278  return __ret;
38279 }
38280 __ai poly64x1_t vreinterpret_p64_bf16(bfloat16x4_t __p0) {
38281  poly64x1_t __ret;
38282  __ret = (poly64x1_t)(__p0);
38283  return __ret;
38284 }
38285 __ai poly16x4_t vreinterpret_p16_bf16(bfloat16x4_t __p0) {
38286  poly16x4_t __ret;
38287  __ret = (poly16x4_t)(__p0);
38288  return __ret;
38289 }
38290 __ai poly8x16_t vreinterpretq_p8_bf16(bfloat16x8_t __p0) {
38291  poly8x16_t __ret;
38292  __ret = (poly8x16_t)(__p0);
38293  return __ret;
38294 }
38295 __ai poly128_t vreinterpretq_p128_bf16(bfloat16x8_t __p0) {
38296  poly128_t __ret;
38297  __ret = (poly128_t)(__p0);
38298  return __ret;
38299 }
38300 __ai poly64x2_t vreinterpretq_p64_bf16(bfloat16x8_t __p0) {
38301  poly64x2_t __ret;
38302  __ret = (poly64x2_t)(__p0);
38303  return __ret;
38304 }
38305 __ai poly16x8_t vreinterpretq_p16_bf16(bfloat16x8_t __p0) {
38306  poly16x8_t __ret;
38307  __ret = (poly16x8_t)(__p0);
38308  return __ret;
38309 }
38310 __ai uint8x16_t vreinterpretq_u8_bf16(bfloat16x8_t __p0) {
38311  uint8x16_t __ret;
38312  __ret = (uint8x16_t)(__p0);
38313  return __ret;
38314 }
38315 __ai uint32x4_t vreinterpretq_u32_bf16(bfloat16x8_t __p0) {
38316  uint32x4_t __ret;
38317  __ret = (uint32x4_t)(__p0);
38318  return __ret;
38319 }
38320 __ai uint64x2_t vreinterpretq_u64_bf16(bfloat16x8_t __p0) {
38321  uint64x2_t __ret;
38322  __ret = (uint64x2_t)(__p0);
38323  return __ret;
38324 }
38325 __ai uint16x8_t vreinterpretq_u16_bf16(bfloat16x8_t __p0) {
38326  uint16x8_t __ret;
38327  __ret = (uint16x8_t)(__p0);
38328  return __ret;
38329 }
38330 __ai int8x16_t vreinterpretq_s8_bf16(bfloat16x8_t __p0) {
38331  int8x16_t __ret;
38332  __ret = (int8x16_t)(__p0);
38333  return __ret;
38334 }
38335 __ai float64x2_t vreinterpretq_f64_bf16(bfloat16x8_t __p0) {
38336  float64x2_t __ret;
38337  __ret = (float64x2_t)(__p0);
38338  return __ret;
38339 }
38340 __ai float32x4_t vreinterpretq_f32_bf16(bfloat16x8_t __p0) {
38341  float32x4_t __ret;
38342  __ret = (float32x4_t)(__p0);
38343  return __ret;
38344 }
38345 __ai float16x8_t vreinterpretq_f16_bf16(bfloat16x8_t __p0) {
38346  float16x8_t __ret;
38347  __ret = (float16x8_t)(__p0);
38348  return __ret;
38349 }
38350 __ai int32x4_t vreinterpretq_s32_bf16(bfloat16x8_t __p0) {
38351  int32x4_t __ret;
38352  __ret = (int32x4_t)(__p0);
38353  return __ret;
38354 }
38355 __ai int64x2_t vreinterpretq_s64_bf16(bfloat16x8_t __p0) {
38356  int64x2_t __ret;
38357  __ret = (int64x2_t)(__p0);
38358  return __ret;
38359 }
38360 __ai int16x8_t vreinterpretq_s16_bf16(bfloat16x8_t __p0) {
38361  int16x8_t __ret;
38362  __ret = (int16x8_t)(__p0);
38363  return __ret;
38364 }
38365 __ai uint8x8_t vreinterpret_u8_bf16(bfloat16x4_t __p0) {
38366  uint8x8_t __ret;
38367  __ret = (uint8x8_t)(__p0);
38368  return __ret;
38369 }
38370 __ai uint32x2_t vreinterpret_u32_bf16(bfloat16x4_t __p0) {
38371  uint32x2_t __ret;
38372  __ret = (uint32x2_t)(__p0);
38373  return __ret;
38374 }
38375 __ai uint64x1_t vreinterpret_u64_bf16(bfloat16x4_t __p0) {
38376  uint64x1_t __ret;
38377  __ret = (uint64x1_t)(__p0);
38378  return __ret;
38379 }
38380 __ai uint16x4_t vreinterpret_u16_bf16(bfloat16x4_t __p0) {
38381  uint16x4_t __ret;
38382  __ret = (uint16x4_t)(__p0);
38383  return __ret;
38384 }
38385 __ai int8x8_t vreinterpret_s8_bf16(bfloat16x4_t __p0) {
38386  int8x8_t __ret;
38387  __ret = (int8x8_t)(__p0);
38388  return __ret;
38389 }
38390 __ai float64x1_t vreinterpret_f64_bf16(bfloat16x4_t __p0) {
38391  float64x1_t __ret;
38392  __ret = (float64x1_t)(__p0);
38393  return __ret;
38394 }
38395 __ai float32x2_t vreinterpret_f32_bf16(bfloat16x4_t __p0) {
38396  float32x2_t __ret;
38397  __ret = (float32x2_t)(__p0);
38398  return __ret;
38399 }
38400 __ai float16x4_t vreinterpret_f16_bf16(bfloat16x4_t __p0) {
38401  float16x4_t __ret;
38402  __ret = (float16x4_t)(__p0);
38403  return __ret;
38404 }
38405 __ai int32x2_t vreinterpret_s32_bf16(bfloat16x4_t __p0) {
38406  int32x2_t __ret;
38407  __ret = (int32x2_t)(__p0);
38408  return __ret;
38409 }
38410 __ai int64x1_t vreinterpret_s64_bf16(bfloat16x4_t __p0) {
38411  int64x1_t __ret;
38412  __ret = (int64x1_t)(__p0);
38413  return __ret;
38414 }
38415 __ai int16x4_t vreinterpret_s16_bf16(bfloat16x4_t __p0) {
38416  int16x4_t __ret;
38417  __ret = (int16x4_t)(__p0);
38418  return __ret;
38419 }
38420 __ai bfloat16x8_t vreinterpretq_bf16_p8(poly8x16_t __p0) {
38421  bfloat16x8_t __ret;
38422  __ret = (bfloat16x8_t)(__p0);
38423  return __ret;
38424 }
38425 __ai bfloat16x8_t vreinterpretq_bf16_p128(poly128_t __p0) {
38426  bfloat16x8_t __ret;
38427  __ret = (bfloat16x8_t)(__p0);
38428  return __ret;
38429 }
38430 __ai bfloat16x8_t vreinterpretq_bf16_p64(poly64x2_t __p0) {
38431  bfloat16x8_t __ret;
38432  __ret = (bfloat16x8_t)(__p0);
38433  return __ret;
38434 }
38435 __ai bfloat16x8_t vreinterpretq_bf16_p16(poly16x8_t __p0) {
38436  bfloat16x8_t __ret;
38437  __ret = (bfloat16x8_t)(__p0);
38438  return __ret;
38439 }
38440 __ai bfloat16x8_t vreinterpretq_bf16_u8(uint8x16_t __p0) {
38441  bfloat16x8_t __ret;
38442  __ret = (bfloat16x8_t)(__p0);
38443  return __ret;
38444 }
38445 __ai bfloat16x8_t vreinterpretq_bf16_u32(uint32x4_t __p0) {
38446  bfloat16x8_t __ret;
38447  __ret = (bfloat16x8_t)(__p0);
38448  return __ret;
38449 }
38450 __ai bfloat16x8_t vreinterpretq_bf16_u64(uint64x2_t __p0) {
38451  bfloat16x8_t __ret;
38452  __ret = (bfloat16x8_t)(__p0);
38453  return __ret;
38454 }
38455 __ai bfloat16x8_t vreinterpretq_bf16_u16(uint16x8_t __p0) {
38456  bfloat16x8_t __ret;
38457  __ret = (bfloat16x8_t)(__p0);
38458  return __ret;
38459 }
38460 __ai bfloat16x8_t vreinterpretq_bf16_s8(int8x16_t __p0) {
38461  bfloat16x8_t __ret;
38462  __ret = (bfloat16x8_t)(__p0);
38463  return __ret;
38464 }
38465 __ai bfloat16x8_t vreinterpretq_bf16_f64(float64x2_t __p0) {
38466  bfloat16x8_t __ret;
38467  __ret = (bfloat16x8_t)(__p0);
38468  return __ret;
38469 }
38470 __ai bfloat16x8_t vreinterpretq_bf16_f32(float32x4_t __p0) {
38471  bfloat16x8_t __ret;
38472  __ret = (bfloat16x8_t)(__p0);
38473  return __ret;
38474 }
38475 __ai bfloat16x8_t vreinterpretq_bf16_f16(float16x8_t __p0) {
38476  bfloat16x8_t __ret;
38477  __ret = (bfloat16x8_t)(__p0);
38478  return __ret;
38479 }
38480 __ai bfloat16x8_t vreinterpretq_bf16_s32(int32x4_t __p0) {
38481  bfloat16x8_t __ret;
38482  __ret = (bfloat16x8_t)(__p0);
38483  return __ret;
38484 }
38485 __ai bfloat16x8_t vreinterpretq_bf16_s64(int64x2_t __p0) {
38486  bfloat16x8_t __ret;
38487  __ret = (bfloat16x8_t)(__p0);
38488  return __ret;
38489 }
38490 __ai bfloat16x8_t vreinterpretq_bf16_s16(int16x8_t __p0) {
38491  bfloat16x8_t __ret;
38492  __ret = (bfloat16x8_t)(__p0);
38493  return __ret;
38494 }
38495 __ai bfloat16x4_t vreinterpret_bf16_p8(poly8x8_t __p0) {
38496  bfloat16x4_t __ret;
38497  __ret = (bfloat16x4_t)(__p0);
38498  return __ret;
38499 }
38500 __ai bfloat16x4_t vreinterpret_bf16_p64(poly64x1_t __p0) {
38501  bfloat16x4_t __ret;
38502  __ret = (bfloat16x4_t)(__p0);
38503  return __ret;
38504 }
38505 __ai bfloat16x4_t vreinterpret_bf16_p16(poly16x4_t __p0) {
38506  bfloat16x4_t __ret;
38507  __ret = (bfloat16x4_t)(__p0);
38508  return __ret;
38509 }
38510 __ai bfloat16x4_t vreinterpret_bf16_u8(uint8x8_t __p0) {
38511  bfloat16x4_t __ret;
38512  __ret = (bfloat16x4_t)(__p0);
38513  return __ret;
38514 }
38515 __ai bfloat16x4_t vreinterpret_bf16_u32(uint32x2_t __p0) {
38516  bfloat16x4_t __ret;
38517  __ret = (bfloat16x4_t)(__p0);
38518  return __ret;
38519 }
38520 __ai bfloat16x4_t vreinterpret_bf16_u64(uint64x1_t __p0) {
38521  bfloat16x4_t __ret;
38522  __ret = (bfloat16x4_t)(__p0);
38523  return __ret;
38524 }
38525 __ai bfloat16x4_t vreinterpret_bf16_u16(uint16x4_t __p0) {
38526  bfloat16x4_t __ret;
38527  __ret = (bfloat16x4_t)(__p0);
38528  return __ret;
38529 }
38530 __ai bfloat16x4_t vreinterpret_bf16_s8(int8x8_t __p0) {
38531  bfloat16x4_t __ret;
38532  __ret = (bfloat16x4_t)(__p0);
38533  return __ret;
38534 }
38535 __ai bfloat16x4_t vreinterpret_bf16_f64(float64x1_t __p0) {
38536  bfloat16x4_t __ret;
38537  __ret = (bfloat16x4_t)(__p0);
38538  return __ret;
38539 }
38540 __ai bfloat16x4_t vreinterpret_bf16_f32(float32x2_t __p0) {
38541  bfloat16x4_t __ret;
38542  __ret = (bfloat16x4_t)(__p0);
38543  return __ret;
38544 }
38545 __ai bfloat16x4_t vreinterpret_bf16_f16(float16x4_t __p0) {
38546  bfloat16x4_t __ret;
38547  __ret = (bfloat16x4_t)(__p0);
38548  return __ret;
38549 }
38550 __ai bfloat16x4_t vreinterpret_bf16_s32(int32x2_t __p0) {
38551  bfloat16x4_t __ret;
38552  __ret = (bfloat16x4_t)(__p0);
38553  return __ret;
38554 }
38555 __ai bfloat16x4_t vreinterpret_bf16_s64(int64x1_t __p0) {
38556  bfloat16x4_t __ret;
38557  __ret = (bfloat16x4_t)(__p0);
38558  return __ret;
38559 }
38560 __ai bfloat16x4_t vreinterpret_bf16_s16(int16x4_t __p0) {
38561  bfloat16x4_t __ret;
38562  __ret = (bfloat16x4_t)(__p0);
38563  return __ret;
38564 }
38565 #endif
38566 #if defined(__ARM_FEATURE_BF16_VECTOR_ARITHMETIC)
38567 #ifdef __LITTLE_ENDIAN__
38568 #define splatq_lane_bf16(__p0, __p1) __extension__ ({ \
38569  bfloat16x4_t __s0 = __p0; \
38570  bfloat16x8_t __ret; \
38571  __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 11); \
38572  __ret; \
38573 })
38574 #else
38575 #define splatq_lane_bf16(__p0, __p1) __extension__ ({ \
38576  bfloat16x4_t __s0 = __p0; \
38577  bfloat16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
38578  bfloat16x8_t __ret; \
38579  __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__rev0, __p1, 11); \
38580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
38581  __ret; \
38582 })
38583 #define __noswap_splatq_lane_bf16(__p0, __p1) __extension__ ({ \
38584  bfloat16x4_t __s0 = __p0; \
38585  bfloat16x8_t __ret; \
38586  __ret = (bfloat16x8_t) __builtin_neon_splatq_lane_v((int8x8_t)__s0, __p1, 11); \
38587  __ret; \
38588 })
38589 #endif
38590 
38591 #ifdef __LITTLE_ENDIAN__
38592 #define splat_lane_bf16(__p0, __p1) __extension__ ({ \
38593  bfloat16x4_t __s0 = __p0; \
38594  bfloat16x4_t __ret; \
38595  __ret = (bfloat16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 11); \
38596  __ret; \
38597 })
38598 #else
38599 #define splat_lane_bf16(__p0, __p1) __extension__ ({ \
38600  bfloat16x4_t __s0 = __p0; \
38601  bfloat16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
38602  bfloat16x4_t __ret; \
38603  __ret = (bfloat16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__rev0, __p1, 11); \
38604  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
38605  __ret; \
38606 })
38607 #define __noswap_splat_lane_bf16(__p0, __p1) __extension__ ({ \
38608  bfloat16x4_t __s0 = __p0; \
38609  bfloat16x4_t __ret; \
38610  __ret = (bfloat16x4_t) __builtin_neon_splat_lane_v((int8x8_t)__s0, __p1, 11); \
38611  __ret; \
38612 })
38613 #endif
38614 
38615 #ifdef __LITTLE_ENDIAN__
38616 #define splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
38617  bfloat16x8_t __s0 = __p0; \
38618  bfloat16x8_t __ret; \
38619  __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 43); \
38620  __ret; \
38621 })
38622 #else
38623 #define splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
38624  bfloat16x8_t __s0 = __p0; \
38625  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
38626  bfloat16x8_t __ret; \
38627  __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__rev0, __p1, 43); \
38628  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
38629  __ret; \
38630 })
38631 #define __noswap_splatq_laneq_bf16(__p0, __p1) __extension__ ({ \
38632  bfloat16x8_t __s0 = __p0; \
38633  bfloat16x8_t __ret; \
38634  __ret = (bfloat16x8_t) __builtin_neon_splatq_laneq_v((int8x16_t)__s0, __p1, 43); \
38635  __ret; \
38636 })
38637 #endif
38638 
38639 #ifdef __LITTLE_ENDIAN__
38640 #define splat_laneq_bf16(__p0, __p1) __extension__ ({ \
38641  bfloat16x8_t __s0 = __p0; \
38642  bfloat16x4_t __ret; \
38643  __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 43); \
38644  __ret; \
38645 })
38646 #else
38647 #define splat_laneq_bf16(__p0, __p1) __extension__ ({ \
38648  bfloat16x8_t __s0 = __p0; \
38649  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
38650  bfloat16x4_t __ret; \
38651  __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__rev0, __p1, 43); \
38652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
38653  __ret; \
38654 })
38655 #define __noswap_splat_laneq_bf16(__p0, __p1) __extension__ ({ \
38656  bfloat16x8_t __s0 = __p0; \
38657  bfloat16x4_t __ret; \
38658  __ret = (bfloat16x4_t) __builtin_neon_splat_laneq_v((int8x16_t)__s0, __p1, 43); \
38659  __ret; \
38660 })
38661 #endif
38662 
38663 #ifdef __LITTLE_ENDIAN__
38664 __ai float32x4_t vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38665  float32x4_t __ret;
38666  __ret = (float32x4_t) __builtin_neon_vbfdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38667  return __ret;
38668 }
38669 #else
38670 __ai float32x4_t vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38671  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38672  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
38673  bfloat16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
38674  float32x4_t __ret;
38675  __ret = (float32x4_t) __builtin_neon_vbfdotq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
38676  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38677  return __ret;
38678 }
38679 __ai float32x4_t __noswap_vbfdotq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38680  float32x4_t __ret;
38681  __ret = (float32x4_t) __builtin_neon_vbfdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38682  return __ret;
38683 }
38684 #endif
38685 
38686 #ifdef __LITTLE_ENDIAN__
38687 __ai float32x2_t vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
38688  float32x2_t __ret;
38689  __ret = (float32x2_t) __builtin_neon_vbfdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
38690  return __ret;
38691 }
38692 #else
38693 __ai float32x2_t vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
38694  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
38695  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
38696  bfloat16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
38697  float32x2_t __ret;
38698  __ret = (float32x2_t) __builtin_neon_vbfdot_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
38699  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
38700  return __ret;
38701 }
38702 __ai float32x2_t __noswap_vbfdot_f32(float32x2_t __p0, bfloat16x4_t __p1, bfloat16x4_t __p2) {
38703  float32x2_t __ret;
38704  __ret = (float32x2_t) __builtin_neon_vbfdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
38705  return __ret;
38706 }
38707 #endif
38708 
38709 #ifdef __LITTLE_ENDIAN__
38710 #define vbfdotq_lane_f32(__p0_142, __p1_142, __p2_142, __p3_142) __extension__ ({ \
38711  float32x4_t __s0_142 = __p0_142; \
38712  bfloat16x8_t __s1_142 = __p1_142; \
38713  bfloat16x4_t __s2_142 = __p2_142; \
38714  float32x4_t __ret_142; \
38715 bfloat16x4_t __reint_142 = __s2_142; \
38716 float32x4_t __reint1_142 = splatq_lane_f32(*(float32x2_t *) &__reint_142, __p3_142); \
38717  __ret_142 = vbfdotq_f32(__s0_142, __s1_142, *(bfloat16x8_t *) &__reint1_142); \
38718  __ret_142; \
38719 })
38720 #else
38721 #define vbfdotq_lane_f32(__p0_143, __p1_143, __p2_143, __p3_143) __extension__ ({ \
38722  float32x4_t __s0_143 = __p0_143; \
38723  bfloat16x8_t __s1_143 = __p1_143; \
38724  bfloat16x4_t __s2_143 = __p2_143; \
38725  float32x4_t __rev0_143; __rev0_143 = __builtin_shufflevector(__s0_143, __s0_143, 3, 2, 1, 0); \
38726  bfloat16x8_t __rev1_143; __rev1_143 = __builtin_shufflevector(__s1_143, __s1_143, 7, 6, 5, 4, 3, 2, 1, 0); \
38727  bfloat16x4_t __rev2_143; __rev2_143 = __builtin_shufflevector(__s2_143, __s2_143, 3, 2, 1, 0); \
38728  float32x4_t __ret_143; \
38729 bfloat16x4_t __reint_143 = __rev2_143; \
38730 float32x4_t __reint1_143 = __noswap_splatq_lane_f32(*(float32x2_t *) &__reint_143, __p3_143); \
38731  __ret_143 = __noswap_vbfdotq_f32(__rev0_143, __rev1_143, *(bfloat16x8_t *) &__reint1_143); \
38732  __ret_143 = __builtin_shufflevector(__ret_143, __ret_143, 3, 2, 1, 0); \
38733  __ret_143; \
38734 })
38735 #endif
38736 
38737 #ifdef __LITTLE_ENDIAN__
38738 #define vbfdot_lane_f32(__p0_144, __p1_144, __p2_144, __p3_144) __extension__ ({ \
38739  float32x2_t __s0_144 = __p0_144; \
38740  bfloat16x4_t __s1_144 = __p1_144; \
38741  bfloat16x4_t __s2_144 = __p2_144; \
38742  float32x2_t __ret_144; \
38743 bfloat16x4_t __reint_144 = __s2_144; \
38744 float32x2_t __reint1_144 = splat_lane_f32(*(float32x2_t *) &__reint_144, __p3_144); \
38745  __ret_144 = vbfdot_f32(__s0_144, __s1_144, *(bfloat16x4_t *) &__reint1_144); \
38746  __ret_144; \
38747 })
38748 #else
38749 #define vbfdot_lane_f32(__p0_145, __p1_145, __p2_145, __p3_145) __extension__ ({ \
38750  float32x2_t __s0_145 = __p0_145; \
38751  bfloat16x4_t __s1_145 = __p1_145; \
38752  bfloat16x4_t __s2_145 = __p2_145; \
38753  float32x2_t __rev0_145; __rev0_145 = __builtin_shufflevector(__s0_145, __s0_145, 1, 0); \
38754  bfloat16x4_t __rev1_145; __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 3, 2, 1, 0); \
38755  bfloat16x4_t __rev2_145; __rev2_145 = __builtin_shufflevector(__s2_145, __s2_145, 3, 2, 1, 0); \
38756  float32x2_t __ret_145; \
38757 bfloat16x4_t __reint_145 = __rev2_145; \
38758 float32x2_t __reint1_145 = __noswap_splat_lane_f32(*(float32x2_t *) &__reint_145, __p3_145); \
38759  __ret_145 = __noswap_vbfdot_f32(__rev0_145, __rev1_145, *(bfloat16x4_t *) &__reint1_145); \
38760  __ret_145 = __builtin_shufflevector(__ret_145, __ret_145, 1, 0); \
38761  __ret_145; \
38762 })
38763 #endif
38764 
38765 #ifdef __LITTLE_ENDIAN__
38766 #define vbfdotq_laneq_f32(__p0_146, __p1_146, __p2_146, __p3_146) __extension__ ({ \
38767  float32x4_t __s0_146 = __p0_146; \
38768  bfloat16x8_t __s1_146 = __p1_146; \
38769  bfloat16x8_t __s2_146 = __p2_146; \
38770  float32x4_t __ret_146; \
38771 bfloat16x8_t __reint_146 = __s2_146; \
38772 float32x4_t __reint1_146 = splatq_laneq_f32(*(float32x4_t *) &__reint_146, __p3_146); \
38773  __ret_146 = vbfdotq_f32(__s0_146, __s1_146, *(bfloat16x8_t *) &__reint1_146); \
38774  __ret_146; \
38775 })
38776 #else
38777 #define vbfdotq_laneq_f32(__p0_147, __p1_147, __p2_147, __p3_147) __extension__ ({ \
38778  float32x4_t __s0_147 = __p0_147; \
38779  bfloat16x8_t __s1_147 = __p1_147; \
38780  bfloat16x8_t __s2_147 = __p2_147; \
38781  float32x4_t __rev0_147; __rev0_147 = __builtin_shufflevector(__s0_147, __s0_147, 3, 2, 1, 0); \
38782  bfloat16x8_t __rev1_147; __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 7, 6, 5, 4, 3, 2, 1, 0); \
38783  bfloat16x8_t __rev2_147; __rev2_147 = __builtin_shufflevector(__s2_147, __s2_147, 7, 6, 5, 4, 3, 2, 1, 0); \
38784  float32x4_t __ret_147; \
38785 bfloat16x8_t __reint_147 = __rev2_147; \
38786 float32x4_t __reint1_147 = __noswap_splatq_laneq_f32(*(float32x4_t *) &__reint_147, __p3_147); \
38787  __ret_147 = __noswap_vbfdotq_f32(__rev0_147, __rev1_147, *(bfloat16x8_t *) &__reint1_147); \
38788  __ret_147 = __builtin_shufflevector(__ret_147, __ret_147, 3, 2, 1, 0); \
38789  __ret_147; \
38790 })
38791 #endif
38792 
38793 #ifdef __LITTLE_ENDIAN__
38794 #define vbfdot_laneq_f32(__p0_148, __p1_148, __p2_148, __p3_148) __extension__ ({ \
38795  float32x2_t __s0_148 = __p0_148; \
38796  bfloat16x4_t __s1_148 = __p1_148; \
38797  bfloat16x8_t __s2_148 = __p2_148; \
38798  float32x2_t __ret_148; \
38799 bfloat16x8_t __reint_148 = __s2_148; \
38800 float32x2_t __reint1_148 = splat_laneq_f32(*(float32x4_t *) &__reint_148, __p3_148); \
38801  __ret_148 = vbfdot_f32(__s0_148, __s1_148, *(bfloat16x4_t *) &__reint1_148); \
38802  __ret_148; \
38803 })
38804 #else
38805 #define vbfdot_laneq_f32(__p0_149, __p1_149, __p2_149, __p3_149) __extension__ ({ \
38806  float32x2_t __s0_149 = __p0_149; \
38807  bfloat16x4_t __s1_149 = __p1_149; \
38808  bfloat16x8_t __s2_149 = __p2_149; \
38809  float32x2_t __rev0_149; __rev0_149 = __builtin_shufflevector(__s0_149, __s0_149, 1, 0); \
38810  bfloat16x4_t __rev1_149; __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
38811  bfloat16x8_t __rev2_149; __rev2_149 = __builtin_shufflevector(__s2_149, __s2_149, 7, 6, 5, 4, 3, 2, 1, 0); \
38812  float32x2_t __ret_149; \
38813 bfloat16x8_t __reint_149 = __rev2_149; \
38814 float32x2_t __reint1_149 = __noswap_splat_laneq_f32(*(float32x4_t *) &__reint_149, __p3_149); \
38815  __ret_149 = __noswap_vbfdot_f32(__rev0_149, __rev1_149, *(bfloat16x4_t *) &__reint1_149); \
38816  __ret_149 = __builtin_shufflevector(__ret_149, __ret_149, 1, 0); \
38817  __ret_149; \
38818 })
38819 #endif
38820 
38821 #ifdef __LITTLE_ENDIAN__
38822 __ai float32x4_t vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38823  float32x4_t __ret;
38824  __ret = (float32x4_t) __builtin_neon_vbfmlalbq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38825  return __ret;
38826 }
38827 #else
38828 __ai float32x4_t vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38829  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38830  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
38831  bfloat16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
38832  float32x4_t __ret;
38833  __ret = (float32x4_t) __builtin_neon_vbfmlalbq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
38834  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38835  return __ret;
38836 }
38837 __ai float32x4_t __noswap_vbfmlalbq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38838  float32x4_t __ret;
38839  __ret = (float32x4_t) __builtin_neon_vbfmlalbq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38840  return __ret;
38841 }
38842 #endif
38843 
38844 #ifdef __LITTLE_ENDIAN__
38845 __ai float32x4_t vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38846  float32x4_t __ret;
38847  __ret = (float32x4_t) __builtin_neon_vbfmlaltq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38848  return __ret;
38849 }
38850 #else
38851 __ai float32x4_t vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38852  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38853  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
38854  bfloat16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
38855  float32x4_t __ret;
38856  __ret = (float32x4_t) __builtin_neon_vbfmlaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
38857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38858  return __ret;
38859 }
38860 __ai float32x4_t __noswap_vbfmlaltq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38861  float32x4_t __ret;
38862  __ret = (float32x4_t) __builtin_neon_vbfmlaltq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38863  return __ret;
38864 }
38865 #endif
38866 
38867 #ifdef __LITTLE_ENDIAN__
38868 __ai float32x4_t vbfmmlaq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38869  float32x4_t __ret;
38870  __ret = (float32x4_t) __builtin_neon_vbfmmlaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
38871  return __ret;
38872 }
38873 #else
38874 __ai float32x4_t vbfmmlaq_f32(float32x4_t __p0, bfloat16x8_t __p1, bfloat16x8_t __p2) {
38875  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38876  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
38877  bfloat16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
38878  float32x4_t __ret;
38879  __ret = (float32x4_t) __builtin_neon_vbfmmlaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
38880  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
38881  return __ret;
38882 }
38883 #endif
38884 
38885 #ifdef __LITTLE_ENDIAN__
38886 __ai bfloat16x8_t vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
38887  bfloat16x8_t __ret;
38888  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
38889  return __ret;
38890 }
38891 #else
38892 __ai bfloat16x8_t vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
38893  bfloat16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
38894  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
38895  bfloat16x8_t __ret;
38896  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
38897  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
38898  return __ret;
38899 }
38900 __ai bfloat16x8_t __noswap_vcombine_bf16(bfloat16x4_t __p0, bfloat16x4_t __p1) {
38901  bfloat16x8_t __ret;
38902  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
38903  return __ret;
38904 }
38905 #endif
38906 
38907 #define vcreate_bf16(__p0) __extension__ ({ \
38908  bfloat16x4_t __ret; \
38909  uint64_t __promote = __p0; \
38910  __ret = (bfloat16x4_t)(__promote); \
38911  __ret; \
38912 })
38913 #ifdef __LITTLE_ENDIAN__
38914 __ai float32x4_t vcvt_f32_bf16(bfloat16x4_t __p0_150) {
38915  float32x4_t __ret_150;
38916 bfloat16x4_t __reint_150 = __p0_150;
38917 int32x4_t __reint1_150 = vshll_n_s16(*(int16x4_t *) &__reint_150, 16);
38918  __ret_150 = *(float32x4_t *) &__reint1_150;
38919  return __ret_150;
38920 }
38921 #else
38922 __ai float32x4_t vcvt_f32_bf16(bfloat16x4_t __p0_151) {
38923  bfloat16x4_t __rev0_151; __rev0_151 = __builtin_shufflevector(__p0_151, __p0_151, 3, 2, 1, 0);
38924  float32x4_t __ret_151;
38925 bfloat16x4_t __reint_151 = __rev0_151;
38926 int32x4_t __reint1_151 = __noswap_vshll_n_s16(*(int16x4_t *) &__reint_151, 16);
38927  __ret_151 = *(float32x4_t *) &__reint1_151;
38928  __ret_151 = __builtin_shufflevector(__ret_151, __ret_151, 3, 2, 1, 0);
38929  return __ret_151;
38930 }
38931 __ai float32x4_t __noswap_vcvt_f32_bf16(bfloat16x4_t __p0_152) {
38932  float32x4_t __ret_152;
38933 bfloat16x4_t __reint_152 = __p0_152;
38934 int32x4_t __reint1_152 = __noswap_vshll_n_s16(*(int16x4_t *) &__reint_152, 16);
38935  __ret_152 = *(float32x4_t *) &__reint1_152;
38936  return __ret_152;
38937 }
38938 #endif
38939 
38940 __ai float32_t vcvtah_f32_bf16(bfloat16_t __p0) {
38941  float32_t __ret;
38942 bfloat16_t __reint = __p0;
38943 int32_t __reint1 = *(int32_t *) &__reint << 16;
38944  __ret = *(float32_t *) &__reint1;
38945  return __ret;
38946 }
38947 __ai bfloat16_t vcvth_bf16_f32(float32_t __p0) {
38948  bfloat16_t __ret;
38949  __ret = (bfloat16_t) __builtin_neon_vcvth_bf16_f32(__p0);
38950  return __ret;
38951 }
38952 #ifdef __LITTLE_ENDIAN__
38953 #define vduph_lane_bf16(__p0, __p1) __extension__ ({ \
38954  bfloat16x4_t __s0 = __p0; \
38955  bfloat16_t __ret; \
38956  __ret = (bfloat16_t) __builtin_neon_vduph_lane_bf16((bfloat16x4_t)__s0, __p1); \
38957  __ret; \
38958 })
38959 #else
38960 #define vduph_lane_bf16(__p0, __p1) __extension__ ({ \
38961  bfloat16x4_t __s0 = __p0; \
38962  bfloat16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
38963  bfloat16_t __ret; \
38964  __ret = (bfloat16_t) __builtin_neon_vduph_lane_bf16((bfloat16x4_t)__rev0, __p1); \
38965  __ret; \
38966 })
38967 #endif
38968 
38969 #ifdef __LITTLE_ENDIAN__
38970 #define vdupq_lane_bf16(__p0_153, __p1_153) __extension__ ({ \
38971  bfloat16x4_t __s0_153 = __p0_153; \
38972  bfloat16x8_t __ret_153; \
38973  __ret_153 = splatq_lane_bf16(__s0_153, __p1_153); \
38974  __ret_153; \
38975 })
38976 #else
38977 #define vdupq_lane_bf16(__p0_154, __p1_154) __extension__ ({ \
38978  bfloat16x4_t __s0_154 = __p0_154; \
38979  bfloat16x4_t __rev0_154; __rev0_154 = __builtin_shufflevector(__s0_154, __s0_154, 3, 2, 1, 0); \
38980  bfloat16x8_t __ret_154; \
38981  __ret_154 = __noswap_splatq_lane_bf16(__rev0_154, __p1_154); \
38982  __ret_154 = __builtin_shufflevector(__ret_154, __ret_154, 7, 6, 5, 4, 3, 2, 1, 0); \
38983  __ret_154; \
38984 })
38985 #endif
38986 
38987 #ifdef __LITTLE_ENDIAN__
38988 #define vdup_lane_bf16(__p0_155, __p1_155) __extension__ ({ \
38989  bfloat16x4_t __s0_155 = __p0_155; \
38990  bfloat16x4_t __ret_155; \
38991  __ret_155 = splat_lane_bf16(__s0_155, __p1_155); \
38992  __ret_155; \
38993 })
38994 #else
38995 #define vdup_lane_bf16(__p0_156, __p1_156) __extension__ ({ \
38996  bfloat16x4_t __s0_156 = __p0_156; \
38997  bfloat16x4_t __rev0_156; __rev0_156 = __builtin_shufflevector(__s0_156, __s0_156, 3, 2, 1, 0); \
38998  bfloat16x4_t __ret_156; \
38999  __ret_156 = __noswap_splat_lane_bf16(__rev0_156, __p1_156); \
39000  __ret_156 = __builtin_shufflevector(__ret_156, __ret_156, 3, 2, 1, 0); \
39001  __ret_156; \
39002 })
39003 #endif
39004 
39005 #ifdef __LITTLE_ENDIAN__
39006 #define vduph_laneq_bf16(__p0, __p1) __extension__ ({ \
39007  bfloat16x8_t __s0 = __p0; \
39008  bfloat16_t __ret; \
39009  __ret = (bfloat16_t) __builtin_neon_vduph_laneq_bf16((bfloat16x8_t)__s0, __p1); \
39010  __ret; \
39011 })
39012 #else
39013 #define vduph_laneq_bf16(__p0, __p1) __extension__ ({ \
39014  bfloat16x8_t __s0 = __p0; \
39015  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
39016  bfloat16_t __ret; \
39017  __ret = (bfloat16_t) __builtin_neon_vduph_laneq_bf16((bfloat16x8_t)__rev0, __p1); \
39018  __ret; \
39019 })
39020 #endif
39021 
39022 #ifdef __LITTLE_ENDIAN__
39023 #define vdupq_laneq_bf16(__p0_157, __p1_157) __extension__ ({ \
39024  bfloat16x8_t __s0_157 = __p0_157; \
39025  bfloat16x8_t __ret_157; \
39026  __ret_157 = splatq_laneq_bf16(__s0_157, __p1_157); \
39027  __ret_157; \
39028 })
39029 #else
39030 #define vdupq_laneq_bf16(__p0_158, __p1_158) __extension__ ({ \
39031  bfloat16x8_t __s0_158 = __p0_158; \
39032  bfloat16x8_t __rev0_158; __rev0_158 = __builtin_shufflevector(__s0_158, __s0_158, 7, 6, 5, 4, 3, 2, 1, 0); \
39033  bfloat16x8_t __ret_158; \
39034  __ret_158 = __noswap_splatq_laneq_bf16(__rev0_158, __p1_158); \
39035  __ret_158 = __builtin_shufflevector(__ret_158, __ret_158, 7, 6, 5, 4, 3, 2, 1, 0); \
39036  __ret_158; \
39037 })
39038 #endif
39039 
39040 #ifdef __LITTLE_ENDIAN__
39041 #define vdup_laneq_bf16(__p0_159, __p1_159) __extension__ ({ \
39042  bfloat16x8_t __s0_159 = __p0_159; \
39043  bfloat16x4_t __ret_159; \
39044  __ret_159 = splat_laneq_bf16(__s0_159, __p1_159); \
39045  __ret_159; \
39046 })
39047 #else
39048 #define vdup_laneq_bf16(__p0_160, __p1_160) __extension__ ({ \
39049  bfloat16x8_t __s0_160 = __p0_160; \
39050  bfloat16x8_t __rev0_160; __rev0_160 = __builtin_shufflevector(__s0_160, __s0_160, 7, 6, 5, 4, 3, 2, 1, 0); \
39051  bfloat16x4_t __ret_160; \
39052  __ret_160 = __noswap_splat_laneq_bf16(__rev0_160, __p1_160); \
39053  __ret_160 = __builtin_shufflevector(__ret_160, __ret_160, 3, 2, 1, 0); \
39054  __ret_160; \
39055 })
39056 #endif
39057 
39058 #ifdef __LITTLE_ENDIAN__
39059 __ai bfloat16x8_t vdupq_n_bf16(bfloat16_t __p0) {
39060  bfloat16x8_t __ret;
39061  __ret = (bfloat16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
39062  return __ret;
39063 }
39064 #else
39065 __ai bfloat16x8_t vdupq_n_bf16(bfloat16_t __p0) {
39066  bfloat16x8_t __ret;
39067  __ret = (bfloat16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
39068  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
39069  return __ret;
39070 }
39071 #endif
39072 
39073 #ifdef __LITTLE_ENDIAN__
39074 __ai bfloat16x4_t vdup_n_bf16(bfloat16_t __p0) {
39075  bfloat16x4_t __ret;
39076  __ret = (bfloat16x4_t) {__p0, __p0, __p0, __p0};
39077  return __ret;
39078 }
39079 #else
39080 __ai bfloat16x4_t vdup_n_bf16(bfloat16_t __p0) {
39081  bfloat16x4_t __ret;
39082  __ret = (bfloat16x4_t) {__p0, __p0, __p0, __p0};
39083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39084  return __ret;
39085 }
39086 #endif
39087 
39088 #ifdef __LITTLE_ENDIAN__
39089 __ai bfloat16x4_t vget_high_bf16(bfloat16x8_t __p0) {
39090  bfloat16x4_t __ret;
39091  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
39092  return __ret;
39093 }
39094 #else
39095 __ai bfloat16x4_t vget_high_bf16(bfloat16x8_t __p0) {
39096  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
39097  bfloat16x4_t __ret;
39098  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
39099  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39100  return __ret;
39101 }
39102 __ai bfloat16x4_t __noswap_vget_high_bf16(bfloat16x8_t __p0) {
39103  bfloat16x4_t __ret;
39104  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
39105  return __ret;
39106 }
39107 #endif
39108 
39109 #ifdef __LITTLE_ENDIAN__
39110 #define vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
39111  bfloat16x8_t __s0 = __p0; \
39112  bfloat16_t __ret; \
39113  __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__s0, __p1); \
39114  __ret; \
39115 })
39116 #else
39117 #define vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
39118  bfloat16x8_t __s0 = __p0; \
39119  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
39120  bfloat16_t __ret; \
39121  __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__rev0, __p1); \
39122  __ret; \
39123 })
39124 #define __noswap_vgetq_lane_bf16(__p0, __p1) __extension__ ({ \
39125  bfloat16x8_t __s0 = __p0; \
39126  bfloat16_t __ret; \
39127  __ret = (bfloat16_t) __builtin_neon_vgetq_lane_bf16((bfloat16x8_t)__s0, __p1); \
39128  __ret; \
39129 })
39130 #endif
39131 
39132 #ifdef __LITTLE_ENDIAN__
39133 #define vget_lane_bf16(__p0, __p1) __extension__ ({ \
39134  bfloat16x4_t __s0 = __p0; \
39135  bfloat16_t __ret; \
39136  __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__s0, __p1); \
39137  __ret; \
39138 })
39139 #else
39140 #define vget_lane_bf16(__p0, __p1) __extension__ ({ \
39141  bfloat16x4_t __s0 = __p0; \
39142  bfloat16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
39143  bfloat16_t __ret; \
39144  __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__rev0, __p1); \
39145  __ret; \
39146 })
39147 #define __noswap_vget_lane_bf16(__p0, __p1) __extension__ ({ \
39148  bfloat16x4_t __s0 = __p0; \
39149  bfloat16_t __ret; \
39150  __ret = (bfloat16_t) __builtin_neon_vget_lane_bf16((bfloat16x4_t)__s0, __p1); \
39151  __ret; \
39152 })
39153 #endif
39154 
39155 #ifdef __LITTLE_ENDIAN__
39156 __ai bfloat16x4_t vget_low_bf16(bfloat16x8_t __p0) {
39157  bfloat16x4_t __ret;
39158  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
39159  return __ret;
39160 }
39161 #else
39162 __ai bfloat16x4_t vget_low_bf16(bfloat16x8_t __p0) {
39163  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
39164  bfloat16x4_t __ret;
39165  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
39166  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39167  return __ret;
39168 }
39169 __ai bfloat16x4_t __noswap_vget_low_bf16(bfloat16x8_t __p0) {
39170  bfloat16x4_t __ret;
39171  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
39172  return __ret;
39173 }
39174 #endif
39175 
39176 #ifdef __LITTLE_ENDIAN__
39177 #define vld1q_bf16(__p0) __extension__ ({ \
39178  bfloat16x8_t __ret; \
39179  __ret = (bfloat16x8_t) __builtin_neon_vld1q_v(__p0, 43); \
39180  __ret; \
39181 })
39182 #else
39183 #define vld1q_bf16(__p0) __extension__ ({ \
39184  bfloat16x8_t __ret; \
39185  __ret = (bfloat16x8_t) __builtin_neon_vld1q_v(__p0, 43); \
39186  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39187  __ret; \
39188 })
39189 #endif
39190 
39191 #ifdef __LITTLE_ENDIAN__
39192 #define vld1_bf16(__p0) __extension__ ({ \
39193  bfloat16x4_t __ret; \
39194  __ret = (bfloat16x4_t) __builtin_neon_vld1_v(__p0, 11); \
39195  __ret; \
39196 })
39197 #else
39198 #define vld1_bf16(__p0) __extension__ ({ \
39199  bfloat16x4_t __ret; \
39200  __ret = (bfloat16x4_t) __builtin_neon_vld1_v(__p0, 11); \
39201  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39202  __ret; \
39203 })
39204 #endif
39205 
39206 #ifdef __LITTLE_ENDIAN__
39207 #define vld1q_dup_bf16(__p0) __extension__ ({ \
39208  bfloat16x8_t __ret; \
39209  __ret = (bfloat16x8_t) __builtin_neon_vld1q_dup_v(__p0, 43); \
39210  __ret; \
39211 })
39212 #else
39213 #define vld1q_dup_bf16(__p0) __extension__ ({ \
39214  bfloat16x8_t __ret; \
39215  __ret = (bfloat16x8_t) __builtin_neon_vld1q_dup_v(__p0, 43); \
39216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39217  __ret; \
39218 })
39219 #endif
39220 
39221 #ifdef __LITTLE_ENDIAN__
39222 #define vld1_dup_bf16(__p0) __extension__ ({ \
39223  bfloat16x4_t __ret; \
39224  __ret = (bfloat16x4_t) __builtin_neon_vld1_dup_v(__p0, 11); \
39225  __ret; \
39226 })
39227 #else
39228 #define vld1_dup_bf16(__p0) __extension__ ({ \
39229  bfloat16x4_t __ret; \
39230  __ret = (bfloat16x4_t) __builtin_neon_vld1_dup_v(__p0, 11); \
39231  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39232  __ret; \
39233 })
39234 #endif
39235 
39236 #ifdef __LITTLE_ENDIAN__
39237 #define vld1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39238  bfloat16x8_t __s1 = __p1; \
39239  bfloat16x8_t __ret; \
39240  __ret = (bfloat16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 43); \
39241  __ret; \
39242 })
39243 #else
39244 #define vld1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39245  bfloat16x8_t __s1 = __p1; \
39246  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39247  bfloat16x8_t __ret; \
39248  __ret = (bfloat16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 43); \
39249  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39250  __ret; \
39251 })
39252 #endif
39253 
39254 #ifdef __LITTLE_ENDIAN__
39255 #define vld1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39256  bfloat16x4_t __s1 = __p1; \
39257  bfloat16x4_t __ret; \
39258  __ret = (bfloat16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 11); \
39259  __ret; \
39260 })
39261 #else
39262 #define vld1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39263  bfloat16x4_t __s1 = __p1; \
39264  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39265  bfloat16x4_t __ret; \
39266  __ret = (bfloat16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 11); \
39267  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39268  __ret; \
39269 })
39270 #endif
39271 
39272 #ifdef __LITTLE_ENDIAN__
39273 #define vld1q_bf16_x2(__p0) __extension__ ({ \
39274  bfloat16x8x2_t __ret; \
39275  __builtin_neon_vld1q_x2_v(&__ret, __p0, 43); \
39276  __ret; \
39277 })
39278 #else
39279 #define vld1q_bf16_x2(__p0) __extension__ ({ \
39280  bfloat16x8x2_t __ret; \
39281  __builtin_neon_vld1q_x2_v(&__ret, __p0, 43); \
39282  \
39283  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39284  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39285  __ret; \
39286 })
39287 #endif
39288 
39289 #ifdef __LITTLE_ENDIAN__
39290 #define vld1_bf16_x2(__p0) __extension__ ({ \
39291  bfloat16x4x2_t __ret; \
39292  __builtin_neon_vld1_x2_v(&__ret, __p0, 11); \
39293  __ret; \
39294 })
39295 #else
39296 #define vld1_bf16_x2(__p0) __extension__ ({ \
39297  bfloat16x4x2_t __ret; \
39298  __builtin_neon_vld1_x2_v(&__ret, __p0, 11); \
39299  \
39300  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39301  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39302  __ret; \
39303 })
39304 #endif
39305 
39306 #ifdef __LITTLE_ENDIAN__
39307 #define vld1q_bf16_x3(__p0) __extension__ ({ \
39308  bfloat16x8x3_t __ret; \
39309  __builtin_neon_vld1q_x3_v(&__ret, __p0, 43); \
39310  __ret; \
39311 })
39312 #else
39313 #define vld1q_bf16_x3(__p0) __extension__ ({ \
39314  bfloat16x8x3_t __ret; \
39315  __builtin_neon_vld1q_x3_v(&__ret, __p0, 43); \
39316  \
39317  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39318  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39319  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39320  __ret; \
39321 })
39322 #endif
39323 
39324 #ifdef __LITTLE_ENDIAN__
39325 #define vld1_bf16_x3(__p0) __extension__ ({ \
39326  bfloat16x4x3_t __ret; \
39327  __builtin_neon_vld1_x3_v(&__ret, __p0, 11); \
39328  __ret; \
39329 })
39330 #else
39331 #define vld1_bf16_x3(__p0) __extension__ ({ \
39332  bfloat16x4x3_t __ret; \
39333  __builtin_neon_vld1_x3_v(&__ret, __p0, 11); \
39334  \
39335  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39336  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39337  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39338  __ret; \
39339 })
39340 #endif
39341 
39342 #ifdef __LITTLE_ENDIAN__
39343 #define vld1q_bf16_x4(__p0) __extension__ ({ \
39344  bfloat16x8x4_t __ret; \
39345  __builtin_neon_vld1q_x4_v(&__ret, __p0, 43); \
39346  __ret; \
39347 })
39348 #else
39349 #define vld1q_bf16_x4(__p0) __extension__ ({ \
39350  bfloat16x8x4_t __ret; \
39351  __builtin_neon_vld1q_x4_v(&__ret, __p0, 43); \
39352  \
39353  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39354  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39355  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39356  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39357  __ret; \
39358 })
39359 #endif
39360 
39361 #ifdef __LITTLE_ENDIAN__
39362 #define vld1_bf16_x4(__p0) __extension__ ({ \
39363  bfloat16x4x4_t __ret; \
39364  __builtin_neon_vld1_x4_v(&__ret, __p0, 11); \
39365  __ret; \
39366 })
39367 #else
39368 #define vld1_bf16_x4(__p0) __extension__ ({ \
39369  bfloat16x4x4_t __ret; \
39370  __builtin_neon_vld1_x4_v(&__ret, __p0, 11); \
39371  \
39372  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39373  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39374  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39375  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39376  __ret; \
39377 })
39378 #endif
39379 
39380 #ifdef __LITTLE_ENDIAN__
39381 #define vld2q_bf16(__p0) __extension__ ({ \
39382  bfloat16x8x2_t __ret; \
39383  __builtin_neon_vld2q_v(&__ret, __p0, 43); \
39384  __ret; \
39385 })
39386 #else
39387 #define vld2q_bf16(__p0) __extension__ ({ \
39388  bfloat16x8x2_t __ret; \
39389  __builtin_neon_vld2q_v(&__ret, __p0, 43); \
39390  \
39391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39393  __ret; \
39394 })
39395 #endif
39396 
39397 #ifdef __LITTLE_ENDIAN__
39398 #define vld2_bf16(__p0) __extension__ ({ \
39399  bfloat16x4x2_t __ret; \
39400  __builtin_neon_vld2_v(&__ret, __p0, 11); \
39401  __ret; \
39402 })
39403 #else
39404 #define vld2_bf16(__p0) __extension__ ({ \
39405  bfloat16x4x2_t __ret; \
39406  __builtin_neon_vld2_v(&__ret, __p0, 11); \
39407  \
39408  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39409  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39410  __ret; \
39411 })
39412 #endif
39413 
39414 #ifdef __LITTLE_ENDIAN__
39415 #define vld2q_dup_bf16(__p0) __extension__ ({ \
39416  bfloat16x8x2_t __ret; \
39417  __builtin_neon_vld2q_dup_v(&__ret, __p0, 43); \
39418  __ret; \
39419 })
39420 #else
39421 #define vld2q_dup_bf16(__p0) __extension__ ({ \
39422  bfloat16x8x2_t __ret; \
39423  __builtin_neon_vld2q_dup_v(&__ret, __p0, 43); \
39424  \
39425  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39426  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39427  __ret; \
39428 })
39429 #endif
39430 
39431 #ifdef __LITTLE_ENDIAN__
39432 #define vld2_dup_bf16(__p0) __extension__ ({ \
39433  bfloat16x4x2_t __ret; \
39434  __builtin_neon_vld2_dup_v(&__ret, __p0, 11); \
39435  __ret; \
39436 })
39437 #else
39438 #define vld2_dup_bf16(__p0) __extension__ ({ \
39439  bfloat16x4x2_t __ret; \
39440  __builtin_neon_vld2_dup_v(&__ret, __p0, 11); \
39441  \
39442  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39443  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39444  __ret; \
39445 })
39446 #endif
39447 
39448 #ifdef __LITTLE_ENDIAN__
39449 #define vld2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39450  bfloat16x8x2_t __s1 = __p1; \
39451  bfloat16x8x2_t __ret; \
39452  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 43); \
39453  __ret; \
39454 })
39455 #else
39456 #define vld2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39457  bfloat16x8x2_t __s1 = __p1; \
39458  bfloat16x8x2_t __rev1; \
39459  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39460  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39461  bfloat16x8x2_t __ret; \
39462  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 43); \
39463  \
39464  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39465  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39466  __ret; \
39467 })
39468 #endif
39469 
39470 #ifdef __LITTLE_ENDIAN__
39471 #define vld2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39472  bfloat16x4x2_t __s1 = __p1; \
39473  bfloat16x4x2_t __ret; \
39474  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 11); \
39475  __ret; \
39476 })
39477 #else
39478 #define vld2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39479  bfloat16x4x2_t __s1 = __p1; \
39480  bfloat16x4x2_t __rev1; \
39481  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39482  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39483  bfloat16x4x2_t __ret; \
39484  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 11); \
39485  \
39486  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39487  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39488  __ret; \
39489 })
39490 #endif
39491 
39492 #ifdef __LITTLE_ENDIAN__
39493 #define vld3q_bf16(__p0) __extension__ ({ \
39494  bfloat16x8x3_t __ret; \
39495  __builtin_neon_vld3q_v(&__ret, __p0, 43); \
39496  __ret; \
39497 })
39498 #else
39499 #define vld3q_bf16(__p0) __extension__ ({ \
39500  bfloat16x8x3_t __ret; \
39501  __builtin_neon_vld3q_v(&__ret, __p0, 43); \
39502  \
39503  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39504  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39505  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39506  __ret; \
39507 })
39508 #endif
39509 
39510 #ifdef __LITTLE_ENDIAN__
39511 #define vld3_bf16(__p0) __extension__ ({ \
39512  bfloat16x4x3_t __ret; \
39513  __builtin_neon_vld3_v(&__ret, __p0, 11); \
39514  __ret; \
39515 })
39516 #else
39517 #define vld3_bf16(__p0) __extension__ ({ \
39518  bfloat16x4x3_t __ret; \
39519  __builtin_neon_vld3_v(&__ret, __p0, 11); \
39520  \
39521  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39522  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39523  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39524  __ret; \
39525 })
39526 #endif
39527 
39528 #ifdef __LITTLE_ENDIAN__
39529 #define vld3q_dup_bf16(__p0) __extension__ ({ \
39530  bfloat16x8x3_t __ret; \
39531  __builtin_neon_vld3q_dup_v(&__ret, __p0, 43); \
39532  __ret; \
39533 })
39534 #else
39535 #define vld3q_dup_bf16(__p0) __extension__ ({ \
39536  bfloat16x8x3_t __ret; \
39537  __builtin_neon_vld3q_dup_v(&__ret, __p0, 43); \
39538  \
39539  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39540  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39541  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39542  __ret; \
39543 })
39544 #endif
39545 
39546 #ifdef __LITTLE_ENDIAN__
39547 #define vld3_dup_bf16(__p0) __extension__ ({ \
39548  bfloat16x4x3_t __ret; \
39549  __builtin_neon_vld3_dup_v(&__ret, __p0, 11); \
39550  __ret; \
39551 })
39552 #else
39553 #define vld3_dup_bf16(__p0) __extension__ ({ \
39554  bfloat16x4x3_t __ret; \
39555  __builtin_neon_vld3_dup_v(&__ret, __p0, 11); \
39556  \
39557  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39558  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39559  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39560  __ret; \
39561 })
39562 #endif
39563 
39564 #ifdef __LITTLE_ENDIAN__
39565 #define vld3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39566  bfloat16x8x3_t __s1 = __p1; \
39567  bfloat16x8x3_t __ret; \
39568  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 43); \
39569  __ret; \
39570 })
39571 #else
39572 #define vld3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39573  bfloat16x8x3_t __s1 = __p1; \
39574  bfloat16x8x3_t __rev1; \
39575  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39576  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39577  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39578  bfloat16x8x3_t __ret; \
39579  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 43); \
39580  \
39581  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39582  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39583  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39584  __ret; \
39585 })
39586 #endif
39587 
39588 #ifdef __LITTLE_ENDIAN__
39589 #define vld3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39590  bfloat16x4x3_t __s1 = __p1; \
39591  bfloat16x4x3_t __ret; \
39592  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 11); \
39593  __ret; \
39594 })
39595 #else
39596 #define vld3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39597  bfloat16x4x3_t __s1 = __p1; \
39598  bfloat16x4x3_t __rev1; \
39599  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39600  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39601  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39602  bfloat16x4x3_t __ret; \
39603  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 11); \
39604  \
39605  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39606  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39607  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39608  __ret; \
39609 })
39610 #endif
39611 
39612 #ifdef __LITTLE_ENDIAN__
39613 #define vld4q_bf16(__p0) __extension__ ({ \
39614  bfloat16x8x4_t __ret; \
39615  __builtin_neon_vld4q_v(&__ret, __p0, 43); \
39616  __ret; \
39617 })
39618 #else
39619 #define vld4q_bf16(__p0) __extension__ ({ \
39620  bfloat16x8x4_t __ret; \
39621  __builtin_neon_vld4q_v(&__ret, __p0, 43); \
39622  \
39623  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39624  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39625  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39626  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39627  __ret; \
39628 })
39629 #endif
39630 
39631 #ifdef __LITTLE_ENDIAN__
39632 #define vld4_bf16(__p0) __extension__ ({ \
39633  bfloat16x4x4_t __ret; \
39634  __builtin_neon_vld4_v(&__ret, __p0, 11); \
39635  __ret; \
39636 })
39637 #else
39638 #define vld4_bf16(__p0) __extension__ ({ \
39639  bfloat16x4x4_t __ret; \
39640  __builtin_neon_vld4_v(&__ret, __p0, 11); \
39641  \
39642  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39643  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39644  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39645  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39646  __ret; \
39647 })
39648 #endif
39649 
39650 #ifdef __LITTLE_ENDIAN__
39651 #define vld4q_dup_bf16(__p0) __extension__ ({ \
39652  bfloat16x8x4_t __ret; \
39653  __builtin_neon_vld4q_dup_v(&__ret, __p0, 43); \
39654  __ret; \
39655 })
39656 #else
39657 #define vld4q_dup_bf16(__p0) __extension__ ({ \
39658  bfloat16x8x4_t __ret; \
39659  __builtin_neon_vld4q_dup_v(&__ret, __p0, 43); \
39660  \
39661  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39662  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39663  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39664  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39665  __ret; \
39666 })
39667 #endif
39668 
39669 #ifdef __LITTLE_ENDIAN__
39670 #define vld4_dup_bf16(__p0) __extension__ ({ \
39671  bfloat16x4x4_t __ret; \
39672  __builtin_neon_vld4_dup_v(&__ret, __p0, 11); \
39673  __ret; \
39674 })
39675 #else
39676 #define vld4_dup_bf16(__p0) __extension__ ({ \
39677  bfloat16x4x4_t __ret; \
39678  __builtin_neon_vld4_dup_v(&__ret, __p0, 11); \
39679  \
39680  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39681  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39682  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39683  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39684  __ret; \
39685 })
39686 #endif
39687 
39688 #ifdef __LITTLE_ENDIAN__
39689 #define vld4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39690  bfloat16x8x4_t __s1 = __p1; \
39691  bfloat16x8x4_t __ret; \
39692  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 43); \
39693  __ret; \
39694 })
39695 #else
39696 #define vld4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39697  bfloat16x8x4_t __s1 = __p1; \
39698  bfloat16x8x4_t __rev1; \
39699  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39700  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39701  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39702  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39703  bfloat16x8x4_t __ret; \
39704  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 43); \
39705  \
39706  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39707  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39708  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39709  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39710  __ret; \
39711 })
39712 #endif
39713 
39714 #ifdef __LITTLE_ENDIAN__
39715 #define vld4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39716  bfloat16x4x4_t __s1 = __p1; \
39717  bfloat16x4x4_t __ret; \
39718  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 11); \
39719  __ret; \
39720 })
39721 #else
39722 #define vld4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39723  bfloat16x4x4_t __s1 = __p1; \
39724  bfloat16x4x4_t __rev1; \
39725  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39726  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39727  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39728  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
39729  bfloat16x4x4_t __ret; \
39730  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 11); \
39731  \
39732  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
39733  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
39734  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
39735  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
39736  __ret; \
39737 })
39738 #endif
39739 
39740 #ifdef __LITTLE_ENDIAN__
39741 #define vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39742  bfloat16_t __s0 = __p0; \
39743  bfloat16x8_t __s1 = __p1; \
39744  bfloat16x8_t __ret; \
39745  __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__s1, __p2); \
39746  __ret; \
39747 })
39748 #else
39749 #define vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39750  bfloat16_t __s0 = __p0; \
39751  bfloat16x8_t __s1 = __p1; \
39752  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39753  bfloat16x8_t __ret; \
39754  __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__rev1, __p2); \
39755  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
39756  __ret; \
39757 })
39758 #define __noswap_vsetq_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39759  bfloat16_t __s0 = __p0; \
39760  bfloat16x8_t __s1 = __p1; \
39761  bfloat16x8_t __ret; \
39762  __ret = (bfloat16x8_t) __builtin_neon_vsetq_lane_bf16(__s0, (bfloat16x8_t)__s1, __p2); \
39763  __ret; \
39764 })
39765 #endif
39766 
39767 #ifdef __LITTLE_ENDIAN__
39768 #define vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39769  bfloat16_t __s0 = __p0; \
39770  bfloat16x4_t __s1 = __p1; \
39771  bfloat16x4_t __ret; \
39772  __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__s1, __p2); \
39773  __ret; \
39774 })
39775 #else
39776 #define vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39777  bfloat16_t __s0 = __p0; \
39778  bfloat16x4_t __s1 = __p1; \
39779  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39780  bfloat16x4_t __ret; \
39781  __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__rev1, __p2); \
39782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
39783  __ret; \
39784 })
39785 #define __noswap_vset_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39786  bfloat16_t __s0 = __p0; \
39787  bfloat16x4_t __s1 = __p1; \
39788  bfloat16x4_t __ret; \
39789  __ret = (bfloat16x4_t) __builtin_neon_vset_lane_bf16(__s0, (bfloat16x4_t)__s1, __p2); \
39790  __ret; \
39791 })
39792 #endif
39793 
39794 #ifdef __LITTLE_ENDIAN__
39795 #define vst1q_bf16(__p0, __p1) __extension__ ({ \
39796  bfloat16x8_t __s1 = __p1; \
39797  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 43); \
39798 })
39799 #else
39800 #define vst1q_bf16(__p0, __p1) __extension__ ({ \
39801  bfloat16x8_t __s1 = __p1; \
39802  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39803  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 43); \
39804 })
39805 #endif
39806 
39807 #ifdef __LITTLE_ENDIAN__
39808 #define vst1_bf16(__p0, __p1) __extension__ ({ \
39809  bfloat16x4_t __s1 = __p1; \
39810  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 11); \
39811 })
39812 #else
39813 #define vst1_bf16(__p0, __p1) __extension__ ({ \
39814  bfloat16x4_t __s1 = __p1; \
39815  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39816  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 11); \
39817 })
39818 #endif
39819 
39820 #ifdef __LITTLE_ENDIAN__
39821 #define vst1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39822  bfloat16x8_t __s1 = __p1; \
39823  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 43); \
39824 })
39825 #else
39826 #define vst1q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39827  bfloat16x8_t __s1 = __p1; \
39828  bfloat16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
39829  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 43); \
39830 })
39831 #endif
39832 
39833 #ifdef __LITTLE_ENDIAN__
39834 #define vst1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39835  bfloat16x4_t __s1 = __p1; \
39836  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 11); \
39837 })
39838 #else
39839 #define vst1_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39840  bfloat16x4_t __s1 = __p1; \
39841  bfloat16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
39842  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 11); \
39843 })
39844 #endif
39845 
39846 #ifdef __LITTLE_ENDIAN__
39847 #define vst1q_bf16_x2(__p0, __p1) __extension__ ({ \
39848  bfloat16x8x2_t __s1 = __p1; \
39849  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 43); \
39850 })
39851 #else
39852 #define vst1q_bf16_x2(__p0, __p1) __extension__ ({ \
39853  bfloat16x8x2_t __s1 = __p1; \
39854  bfloat16x8x2_t __rev1; \
39855  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39856  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39857  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 43); \
39858 })
39859 #endif
39860 
39861 #ifdef __LITTLE_ENDIAN__
39862 #define vst1_bf16_x2(__p0, __p1) __extension__ ({ \
39863  bfloat16x4x2_t __s1 = __p1; \
39864  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 11); \
39865 })
39866 #else
39867 #define vst1_bf16_x2(__p0, __p1) __extension__ ({ \
39868  bfloat16x4x2_t __s1 = __p1; \
39869  bfloat16x4x2_t __rev1; \
39870  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39871  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39872  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 11); \
39873 })
39874 #endif
39875 
39876 #ifdef __LITTLE_ENDIAN__
39877 #define vst1q_bf16_x3(__p0, __p1) __extension__ ({ \
39878  bfloat16x8x3_t __s1 = __p1; \
39879  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 43); \
39880 })
39881 #else
39882 #define vst1q_bf16_x3(__p0, __p1) __extension__ ({ \
39883  bfloat16x8x3_t __s1 = __p1; \
39884  bfloat16x8x3_t __rev1; \
39885  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39886  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39887  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39888  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 43); \
39889 })
39890 #endif
39891 
39892 #ifdef __LITTLE_ENDIAN__
39893 #define vst1_bf16_x3(__p0, __p1) __extension__ ({ \
39894  bfloat16x4x3_t __s1 = __p1; \
39895  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 11); \
39896 })
39897 #else
39898 #define vst1_bf16_x3(__p0, __p1) __extension__ ({ \
39899  bfloat16x4x3_t __s1 = __p1; \
39900  bfloat16x4x3_t __rev1; \
39901  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39902  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39903  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39904  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 11); \
39905 })
39906 #endif
39907 
39908 #ifdef __LITTLE_ENDIAN__
39909 #define vst1q_bf16_x4(__p0, __p1) __extension__ ({ \
39910  bfloat16x8x4_t __s1 = __p1; \
39911  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 43); \
39912 })
39913 #else
39914 #define vst1q_bf16_x4(__p0, __p1) __extension__ ({ \
39915  bfloat16x8x4_t __s1 = __p1; \
39916  bfloat16x8x4_t __rev1; \
39917  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39918  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39919  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
39920  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
39921  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 43); \
39922 })
39923 #endif
39924 
39925 #ifdef __LITTLE_ENDIAN__
39926 #define vst1_bf16_x4(__p0, __p1) __extension__ ({ \
39927  bfloat16x4x4_t __s1 = __p1; \
39928  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 11); \
39929 })
39930 #else
39931 #define vst1_bf16_x4(__p0, __p1) __extension__ ({ \
39932  bfloat16x4x4_t __s1 = __p1; \
39933  bfloat16x4x4_t __rev1; \
39934  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39935  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39936  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
39937  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
39938  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 11); \
39939 })
39940 #endif
39941 
39942 #ifdef __LITTLE_ENDIAN__
39943 #define vst2q_bf16(__p0, __p1) __extension__ ({ \
39944  bfloat16x8x2_t __s1 = __p1; \
39945  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 43); \
39946 })
39947 #else
39948 #define vst2q_bf16(__p0, __p1) __extension__ ({ \
39949  bfloat16x8x2_t __s1 = __p1; \
39950  bfloat16x8x2_t __rev1; \
39951  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39952  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39953  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 43); \
39954 })
39955 #endif
39956 
39957 #ifdef __LITTLE_ENDIAN__
39958 #define vst2_bf16(__p0, __p1) __extension__ ({ \
39959  bfloat16x4x2_t __s1 = __p1; \
39960  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 11); \
39961 })
39962 #else
39963 #define vst2_bf16(__p0, __p1) __extension__ ({ \
39964  bfloat16x4x2_t __s1 = __p1; \
39965  bfloat16x4x2_t __rev1; \
39966  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39967  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39968  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 11); \
39969 })
39970 #endif
39971 
39972 #ifdef __LITTLE_ENDIAN__
39973 #define vst2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39974  bfloat16x8x2_t __s1 = __p1; \
39975  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 43); \
39976 })
39977 #else
39978 #define vst2q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39979  bfloat16x8x2_t __s1 = __p1; \
39980  bfloat16x8x2_t __rev1; \
39981  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
39982  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
39983  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 43); \
39984 })
39985 #endif
39986 
39987 #ifdef __LITTLE_ENDIAN__
39988 #define vst2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39989  bfloat16x4x2_t __s1 = __p1; \
39990  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 11); \
39991 })
39992 #else
39993 #define vst2_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
39994  bfloat16x4x2_t __s1 = __p1; \
39995  bfloat16x4x2_t __rev1; \
39996  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
39997  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
39998  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 11); \
39999 })
40000 #endif
40001 
40002 #ifdef __LITTLE_ENDIAN__
40003 #define vst3q_bf16(__p0, __p1) __extension__ ({ \
40004  bfloat16x8x3_t __s1 = __p1; \
40005  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 43); \
40006 })
40007 #else
40008 #define vst3q_bf16(__p0, __p1) __extension__ ({ \
40009  bfloat16x8x3_t __s1 = __p1; \
40010  bfloat16x8x3_t __rev1; \
40011  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40012  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40013  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40014  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 43); \
40015 })
40016 #endif
40017 
40018 #ifdef __LITTLE_ENDIAN__
40019 #define vst3_bf16(__p0, __p1) __extension__ ({ \
40020  bfloat16x4x3_t __s1 = __p1; \
40021  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 11); \
40022 })
40023 #else
40024 #define vst3_bf16(__p0, __p1) __extension__ ({ \
40025  bfloat16x4x3_t __s1 = __p1; \
40026  bfloat16x4x3_t __rev1; \
40027  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40028  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40029  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40030  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 11); \
40031 })
40032 #endif
40033 
40034 #ifdef __LITTLE_ENDIAN__
40035 #define vst3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40036  bfloat16x8x3_t __s1 = __p1; \
40037  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 43); \
40038 })
40039 #else
40040 #define vst3q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40041  bfloat16x8x3_t __s1 = __p1; \
40042  bfloat16x8x3_t __rev1; \
40043  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40044  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40045  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40046  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 43); \
40047 })
40048 #endif
40049 
40050 #ifdef __LITTLE_ENDIAN__
40051 #define vst3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40052  bfloat16x4x3_t __s1 = __p1; \
40053  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 11); \
40054 })
40055 #else
40056 #define vst3_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40057  bfloat16x4x3_t __s1 = __p1; \
40058  bfloat16x4x3_t __rev1; \
40059  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40060  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40061  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40062  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 11); \
40063 })
40064 #endif
40065 
40066 #ifdef __LITTLE_ENDIAN__
40067 #define vst4q_bf16(__p0, __p1) __extension__ ({ \
40068  bfloat16x8x4_t __s1 = __p1; \
40069  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 43); \
40070 })
40071 #else
40072 #define vst4q_bf16(__p0, __p1) __extension__ ({ \
40073  bfloat16x8x4_t __s1 = __p1; \
40074  bfloat16x8x4_t __rev1; \
40075  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40076  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40077  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40078  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
40079  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 43); \
40080 })
40081 #endif
40082 
40083 #ifdef __LITTLE_ENDIAN__
40084 #define vst4_bf16(__p0, __p1) __extension__ ({ \
40085  bfloat16x4x4_t __s1 = __p1; \
40086  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 11); \
40087 })
40088 #else
40089 #define vst4_bf16(__p0, __p1) __extension__ ({ \
40090  bfloat16x4x4_t __s1 = __p1; \
40091  bfloat16x4x4_t __rev1; \
40092  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40093  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40094  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40095  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
40096  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 11); \
40097 })
40098 #endif
40099 
40100 #ifdef __LITTLE_ENDIAN__
40101 #define vst4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40102  bfloat16x8x4_t __s1 = __p1; \
40103  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 43); \
40104 })
40105 #else
40106 #define vst4q_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40107  bfloat16x8x4_t __s1 = __p1; \
40108  bfloat16x8x4_t __rev1; \
40109  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
40110  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
40111  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
40112  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
40113  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 43); \
40114 })
40115 #endif
40116 
40117 #ifdef __LITTLE_ENDIAN__
40118 #define vst4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40119  bfloat16x4x4_t __s1 = __p1; \
40120  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 11); \
40121 })
40122 #else
40123 #define vst4_lane_bf16(__p0, __p1, __p2) __extension__ ({ \
40124  bfloat16x4x4_t __s1 = __p1; \
40125  bfloat16x4x4_t __rev1; \
40126  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
40127  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
40128  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
40129  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
40130  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 11); \
40131 })
40132 #endif
40133 
40134 #endif
40135 #if defined(__ARM_FEATURE_BF16_VECTOR_ARITHMETIC) && !defined(__aarch64__)
40136 #ifdef __LITTLE_ENDIAN__
40137 __ai bfloat16x4_t __a32_vcvt_bf16_f32(float32x4_t __p0) {
40138  bfloat16x4_t __ret;
40139  __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_v((int8x16_t)__p0, 11);
40140  return __ret;
40141 }
40142 #else
40143 __ai bfloat16x4_t __a32_vcvt_bf16_f32(float32x4_t __p0) {
40144  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40145  bfloat16x4_t __ret;
40146  __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_v((int8x16_t)__rev0, 11);
40147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40148  return __ret;
40149 }
40150 __ai bfloat16x4_t __noswap___a32_vcvt_bf16_f32(float32x4_t __p0) {
40151  bfloat16x4_t __ret;
40152  __ret = (bfloat16x4_t) __builtin_neon___a32_vcvt_bf16_v((int8x16_t)__p0, 11);
40153  return __ret;
40154 }
40155 #endif
40156 
40157 #ifdef __LITTLE_ENDIAN__
40158 __ai bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
40159  bfloat16x4_t __ret;
40160  __ret = __a32_vcvt_bf16_f32(__p0);
40161  return __ret;
40162 }
40163 #else
40164 __ai bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
40165  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40166  bfloat16x4_t __ret;
40167  __ret = __noswap___a32_vcvt_bf16_f32(__rev0);
40168  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40169  return __ret;
40170 }
40171 #endif
40172 
40173 #ifdef __LITTLE_ENDIAN__
40174 __ai bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
40175  bfloat16x8_t __ret;
40176  __ret = vcombine_bf16(__a32_vcvt_bf16_f32(__p1), vget_low_bf16(__p0));
40177  return __ret;
40178 }
40179 #else
40180 __ai bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
40181  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40182  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40183  bfloat16x8_t __ret;
40184  __ret = __noswap_vcombine_bf16(__noswap___a32_vcvt_bf16_f32(__rev1), __noswap_vget_low_bf16(__rev0));
40185  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40186  return __ret;
40187 }
40188 #endif
40189 
40190 #ifdef __LITTLE_ENDIAN__
40191 __ai bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
40192  bfloat16x8_t __ret;
40193  __ret = vcombine_bf16((bfloat16x4_t)(0ULL), __a32_vcvt_bf16_f32(__p0));
40194  return __ret;
40195 }
40196 #else
40197 __ai bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
40198  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40199  bfloat16x8_t __ret;
40200  __ret = __noswap_vcombine_bf16((bfloat16x4_t)(0ULL), __noswap___a32_vcvt_bf16_f32(__rev0));
40201  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40202  return __ret;
40203 }
40204 #endif
40205 
40206 #endif
40207 #if defined(__ARM_FEATURE_BF16_VECTOR_ARITHMETIC) && defined(__aarch64__)
40208 #ifdef __LITTLE_ENDIAN__
40209 __ai bfloat16x8_t __a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
40210  bfloat16x8_t __ret;
40211  __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_v((int8x16_t)__p0, 43);
40212  return __ret;
40213 }
40214 #else
40215 __ai bfloat16x8_t __a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
40216  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40217  bfloat16x8_t __ret;
40218  __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_v((int8x16_t)__rev0, 43);
40219  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40220  return __ret;
40221 }
40222 __ai bfloat16x8_t __noswap___a64_vcvtq_low_bf16_f32(float32x4_t __p0) {
40223  bfloat16x8_t __ret;
40224  __ret = (bfloat16x8_t) __builtin_neon___a64_vcvtq_low_bf16_v((int8x16_t)__p0, 43);
40225  return __ret;
40226 }
40227 #endif
40228 
40229 #ifdef __LITTLE_ENDIAN__
40230 #define vcopyq_lane_bf16(__p0_161, __p1_161, __p2_161, __p3_161) __extension__ ({ \
40231  bfloat16x8_t __s0_161 = __p0_161; \
40232  bfloat16x4_t __s2_161 = __p2_161; \
40233  bfloat16x8_t __ret_161; \
40234  __ret_161 = vsetq_lane_bf16(vget_lane_bf16(__s2_161, __p3_161), __s0_161, __p1_161); \
40235  __ret_161; \
40236 })
40237 #else
40238 #define vcopyq_lane_bf16(__p0_162, __p1_162, __p2_162, __p3_162) __extension__ ({ \
40239  bfloat16x8_t __s0_162 = __p0_162; \
40240  bfloat16x4_t __s2_162 = __p2_162; \
40241  bfloat16x8_t __rev0_162; __rev0_162 = __builtin_shufflevector(__s0_162, __s0_162, 7, 6, 5, 4, 3, 2, 1, 0); \
40242  bfloat16x4_t __rev2_162; __rev2_162 = __builtin_shufflevector(__s2_162, __s2_162, 3, 2, 1, 0); \
40243  bfloat16x8_t __ret_162; \
40244  __ret_162 = __noswap_vsetq_lane_bf16(__noswap_vget_lane_bf16(__rev2_162, __p3_162), __rev0_162, __p1_162); \
40245  __ret_162 = __builtin_shufflevector(__ret_162, __ret_162, 7, 6, 5, 4, 3, 2, 1, 0); \
40246  __ret_162; \
40247 })
40248 #endif
40249 
40250 #ifdef __LITTLE_ENDIAN__
40251 #define vcopy_lane_bf16(__p0_163, __p1_163, __p2_163, __p3_163) __extension__ ({ \
40252  bfloat16x4_t __s0_163 = __p0_163; \
40253  bfloat16x4_t __s2_163 = __p2_163; \
40254  bfloat16x4_t __ret_163; \
40255  __ret_163 = vset_lane_bf16(vget_lane_bf16(__s2_163, __p3_163), __s0_163, __p1_163); \
40256  __ret_163; \
40257 })
40258 #else
40259 #define vcopy_lane_bf16(__p0_164, __p1_164, __p2_164, __p3_164) __extension__ ({ \
40260  bfloat16x4_t __s0_164 = __p0_164; \
40261  bfloat16x4_t __s2_164 = __p2_164; \
40262  bfloat16x4_t __rev0_164; __rev0_164 = __builtin_shufflevector(__s0_164, __s0_164, 3, 2, 1, 0); \
40263  bfloat16x4_t __rev2_164; __rev2_164 = __builtin_shufflevector(__s2_164, __s2_164, 3, 2, 1, 0); \
40264  bfloat16x4_t __ret_164; \
40265  __ret_164 = __noswap_vset_lane_bf16(__noswap_vget_lane_bf16(__rev2_164, __p3_164), __rev0_164, __p1_164); \
40266  __ret_164 = __builtin_shufflevector(__ret_164, __ret_164, 3, 2, 1, 0); \
40267  __ret_164; \
40268 })
40269 #endif
40270 
40271 #ifdef __LITTLE_ENDIAN__
40272 #define vcopyq_laneq_bf16(__p0_165, __p1_165, __p2_165, __p3_165) __extension__ ({ \
40273  bfloat16x8_t __s0_165 = __p0_165; \
40274  bfloat16x8_t __s2_165 = __p2_165; \
40275  bfloat16x8_t __ret_165; \
40276  __ret_165 = vsetq_lane_bf16(vgetq_lane_bf16(__s2_165, __p3_165), __s0_165, __p1_165); \
40277  __ret_165; \
40278 })
40279 #else
40280 #define vcopyq_laneq_bf16(__p0_166, __p1_166, __p2_166, __p3_166) __extension__ ({ \
40281  bfloat16x8_t __s0_166 = __p0_166; \
40282  bfloat16x8_t __s2_166 = __p2_166; \
40283  bfloat16x8_t __rev0_166; __rev0_166 = __builtin_shufflevector(__s0_166, __s0_166, 7, 6, 5, 4, 3, 2, 1, 0); \
40284  bfloat16x8_t __rev2_166; __rev2_166 = __builtin_shufflevector(__s2_166, __s2_166, 7, 6, 5, 4, 3, 2, 1, 0); \
40285  bfloat16x8_t __ret_166; \
40286  __ret_166 = __noswap_vsetq_lane_bf16(__noswap_vgetq_lane_bf16(__rev2_166, __p3_166), __rev0_166, __p1_166); \
40287  __ret_166 = __builtin_shufflevector(__ret_166, __ret_166, 7, 6, 5, 4, 3, 2, 1, 0); \
40288  __ret_166; \
40289 })
40290 #endif
40291 
40292 #ifdef __LITTLE_ENDIAN__
40293 #define vcopy_laneq_bf16(__p0_167, __p1_167, __p2_167, __p3_167) __extension__ ({ \
40294  bfloat16x4_t __s0_167 = __p0_167; \
40295  bfloat16x8_t __s2_167 = __p2_167; \
40296  bfloat16x4_t __ret_167; \
40297  __ret_167 = vset_lane_bf16(vgetq_lane_bf16(__s2_167, __p3_167), __s0_167, __p1_167); \
40298  __ret_167; \
40299 })
40300 #else
40301 #define vcopy_laneq_bf16(__p0_168, __p1_168, __p2_168, __p3_168) __extension__ ({ \
40302  bfloat16x4_t __s0_168 = __p0_168; \
40303  bfloat16x8_t __s2_168 = __p2_168; \
40304  bfloat16x4_t __rev0_168; __rev0_168 = __builtin_shufflevector(__s0_168, __s0_168, 3, 2, 1, 0); \
40305  bfloat16x8_t __rev2_168; __rev2_168 = __builtin_shufflevector(__s2_168, __s2_168, 7, 6, 5, 4, 3, 2, 1, 0); \
40306  bfloat16x4_t __ret_168; \
40307  __ret_168 = __noswap_vset_lane_bf16(__noswap_vgetq_lane_bf16(__rev2_168, __p3_168), __rev0_168, __p1_168); \
40308  __ret_168 = __builtin_shufflevector(__ret_168, __ret_168, 3, 2, 1, 0); \
40309  __ret_168; \
40310 })
40311 #endif
40312 
40313 #ifdef __LITTLE_ENDIAN__
40314 __ai bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
40315  bfloat16x4_t __ret;
40316  __ret = vget_low_bf16(__a64_vcvtq_low_bf16_f32(__p0));
40317  return __ret;
40318 }
40319 #else
40320 __ai bfloat16x4_t vcvt_bf16_f32(float32x4_t __p0) {
40321  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40322  bfloat16x4_t __ret;
40323  __ret = __noswap_vget_low_bf16(__noswap___a64_vcvtq_low_bf16_f32(__rev0));
40324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40325  return __ret;
40326 }
40327 #endif
40328 
40329 #ifdef __LITTLE_ENDIAN__
40330 __ai bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
40331  bfloat16x8_t __ret;
40332  __ret = (bfloat16x8_t) __builtin_neon_vcvtq_high_bf16_v((int8x16_t)__p0, (int8x16_t)__p1, 43);
40333  return __ret;
40334 }
40335 #else
40336 __ai bfloat16x8_t vcvtq_high_bf16_f32(bfloat16x8_t __p0, float32x4_t __p1) {
40337  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40338  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40339  bfloat16x8_t __ret;
40340  __ret = (bfloat16x8_t) __builtin_neon_vcvtq_high_bf16_v((int8x16_t)__rev0, (int8x16_t)__rev1, 43);
40341  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40342  return __ret;
40343 }
40344 #endif
40345 
40346 #ifdef __LITTLE_ENDIAN__
40347 __ai bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
40348  bfloat16x8_t __ret;
40349  __ret = __a64_vcvtq_low_bf16_f32(__p0);
40350  return __ret;
40351 }
40352 #else
40353 __ai bfloat16x8_t vcvtq_low_bf16_f32(float32x4_t __p0) {
40354  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40355  bfloat16x8_t __ret;
40356  __ret = __noswap___a64_vcvtq_low_bf16_f32(__rev0);
40357  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40358  return __ret;
40359 }
40360 #endif
40361 
40362 #endif
40363 #if defined(__ARM_FEATURE_COMPLEX)
40364 #ifdef __LITTLE_ENDIAN__
40365 __ai float32x2_t vcadd_rot270_f32(float32x2_t __p0, float32x2_t __p1) {
40366  float32x2_t __ret;
40367  __ret = (float32x2_t) __builtin_neon_vcadd_rot270_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
40368  return __ret;
40369 }
40370 #else
40371 __ai float32x2_t vcadd_rot270_f32(float32x2_t __p0, float32x2_t __p1) {
40372  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40373  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40374  float32x2_t __ret;
40375  __ret = (float32x2_t) __builtin_neon_vcadd_rot270_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
40376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40377  return __ret;
40378 }
40379 #endif
40380 
40381 #ifdef __LITTLE_ENDIAN__
40382 __ai float32x2_t vcadd_rot90_f32(float32x2_t __p0, float32x2_t __p1) {
40383  float32x2_t __ret;
40384  __ret = (float32x2_t) __builtin_neon_vcadd_rot90_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
40385  return __ret;
40386 }
40387 #else
40388 __ai float32x2_t vcadd_rot90_f32(float32x2_t __p0, float32x2_t __p1) {
40389  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40390  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40391  float32x2_t __ret;
40392  __ret = (float32x2_t) __builtin_neon_vcadd_rot90_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
40393  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40394  return __ret;
40395 }
40396 #endif
40397 
40398 #ifdef __LITTLE_ENDIAN__
40399 __ai float32x4_t vcaddq_rot270_f32(float32x4_t __p0, float32x4_t __p1) {
40400  float32x4_t __ret;
40401  __ret = (float32x4_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
40402  return __ret;
40403 }
40404 #else
40405 __ai float32x4_t vcaddq_rot270_f32(float32x4_t __p0, float32x4_t __p1) {
40406  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40407  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40408  float32x4_t __ret;
40409  __ret = (float32x4_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
40410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40411  return __ret;
40412 }
40413 #endif
40414 
40415 #ifdef __LITTLE_ENDIAN__
40416 __ai float32x4_t vcaddq_rot90_f32(float32x4_t __p0, float32x4_t __p1) {
40417  float32x4_t __ret;
40418  __ret = (float32x4_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
40419  return __ret;
40420 }
40421 #else
40422 __ai float32x4_t vcaddq_rot90_f32(float32x4_t __p0, float32x4_t __p1) {
40423  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40424  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40425  float32x4_t __ret;
40426  __ret = (float32x4_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
40427  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40428  return __ret;
40429 }
40430 #endif
40431 
40432 #endif
40433 #if defined(__ARM_FEATURE_COMPLEX) && defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC)
40434 #ifdef __LITTLE_ENDIAN__
40435 __ai float16x4_t vcadd_rot270_f16(float16x4_t __p0, float16x4_t __p1) {
40436  float16x4_t __ret;
40437  __ret = (float16x4_t) __builtin_neon_vcadd_rot270_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
40438  return __ret;
40439 }
40440 #else
40441 __ai float16x4_t vcadd_rot270_f16(float16x4_t __p0, float16x4_t __p1) {
40442  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40443  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40444  float16x4_t __ret;
40445  __ret = (float16x4_t) __builtin_neon_vcadd_rot270_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
40446  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40447  return __ret;
40448 }
40449 #endif
40450 
40451 #ifdef __LITTLE_ENDIAN__
40452 __ai float16x4_t vcadd_rot90_f16(float16x4_t __p0, float16x4_t __p1) {
40453  float16x4_t __ret;
40454  __ret = (float16x4_t) __builtin_neon_vcadd_rot90_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
40455  return __ret;
40456 }
40457 #else
40458 __ai float16x4_t vcadd_rot90_f16(float16x4_t __p0, float16x4_t __p1) {
40459  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40460  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40461  float16x4_t __ret;
40462  __ret = (float16x4_t) __builtin_neon_vcadd_rot90_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
40463  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40464  return __ret;
40465 }
40466 #endif
40467 
40468 #ifdef __LITTLE_ENDIAN__
40469 __ai float16x8_t vcaddq_rot270_f16(float16x8_t __p0, float16x8_t __p1) {
40470  float16x8_t __ret;
40471  __ret = (float16x8_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
40472  return __ret;
40473 }
40474 #else
40475 __ai float16x8_t vcaddq_rot270_f16(float16x8_t __p0, float16x8_t __p1) {
40476  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40477  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40478  float16x8_t __ret;
40479  __ret = (float16x8_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
40480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40481  return __ret;
40482 }
40483 #endif
40484 
40485 #ifdef __LITTLE_ENDIAN__
40486 __ai float16x8_t vcaddq_rot90_f16(float16x8_t __p0, float16x8_t __p1) {
40487  float16x8_t __ret;
40488  __ret = (float16x8_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
40489  return __ret;
40490 }
40491 #else
40492 __ai float16x8_t vcaddq_rot90_f16(float16x8_t __p0, float16x8_t __p1) {
40493  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40494  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40495  float16x8_t __ret;
40496  __ret = (float16x8_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
40497  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40498  return __ret;
40499 }
40500 #endif
40501 
40502 #endif
40503 #if defined(__ARM_FEATURE_COMPLEX) && defined(__aarch64__)
40504 #ifdef __LITTLE_ENDIAN__
40505 __ai float64x2_t vcaddq_rot270_f64(float64x2_t __p0, float64x2_t __p1) {
40506  float64x2_t __ret;
40507  __ret = (float64x2_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40508  return __ret;
40509 }
40510 #else
40511 __ai float64x2_t vcaddq_rot270_f64(float64x2_t __p0, float64x2_t __p1) {
40512  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40513  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40514  float64x2_t __ret;
40515  __ret = (float64x2_t) __builtin_neon_vcaddq_rot270_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40517  return __ret;
40518 }
40519 #endif
40520 
40521 #ifdef __LITTLE_ENDIAN__
40522 __ai float64x2_t vcaddq_rot90_f64(float64x2_t __p0, float64x2_t __p1) {
40523  float64x2_t __ret;
40524  __ret = (float64x2_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40525  return __ret;
40526 }
40527 #else
40528 __ai float64x2_t vcaddq_rot90_f64(float64x2_t __p0, float64x2_t __p1) {
40529  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40530  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40531  float64x2_t __ret;
40532  __ret = (float64x2_t) __builtin_neon_vcaddq_rot90_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40534  return __ret;
40535 }
40536 #endif
40537 
40538 #endif
40539 #if defined(__ARM_FEATURE_DOTPROD)
40540 #ifdef __LITTLE_ENDIAN__
40541 __ai uint32x4_t vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
40542  uint32x4_t __ret;
40543  __ret = (uint32x4_t) __builtin_neon_vdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40544  return __ret;
40545 }
40546 #else
40547 __ai uint32x4_t vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
40548  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40549  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40550  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40551  uint32x4_t __ret;
40552  __ret = (uint32x4_t) __builtin_neon_vdotq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40554  return __ret;
40555 }
40556 __ai uint32x4_t __noswap_vdotq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
40557  uint32x4_t __ret;
40558  __ret = (uint32x4_t) __builtin_neon_vdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40559  return __ret;
40560 }
40561 #endif
40562 
40563 #ifdef __LITTLE_ENDIAN__
40564 __ai int32x4_t vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
40565  int32x4_t __ret;
40566  __ret = (int32x4_t) __builtin_neon_vdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
40567  return __ret;
40568 }
40569 #else
40570 __ai int32x4_t vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
40571  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40572  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40573  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40574  int32x4_t __ret;
40575  __ret = (int32x4_t) __builtin_neon_vdotq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
40576  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40577  return __ret;
40578 }
40579 __ai int32x4_t __noswap_vdotq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
40580  int32x4_t __ret;
40581  __ret = (int32x4_t) __builtin_neon_vdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
40582  return __ret;
40583 }
40584 #endif
40585 
40586 #ifdef __LITTLE_ENDIAN__
40587 __ai uint32x2_t vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
40588  uint32x2_t __ret;
40589  __ret = (uint32x2_t) __builtin_neon_vdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
40590  return __ret;
40591 }
40592 #else
40593 __ai uint32x2_t vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
40594  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40595  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40596  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40597  uint32x2_t __ret;
40598  __ret = (uint32x2_t) __builtin_neon_vdot_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
40599  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40600  return __ret;
40601 }
40602 __ai uint32x2_t __noswap_vdot_u32(uint32x2_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
40603  uint32x2_t __ret;
40604  __ret = (uint32x2_t) __builtin_neon_vdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
40605  return __ret;
40606 }
40607 #endif
40608 
40609 #ifdef __LITTLE_ENDIAN__
40610 __ai int32x2_t vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
40611  int32x2_t __ret;
40612  __ret = (int32x2_t) __builtin_neon_vdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
40613  return __ret;
40614 }
40615 #else
40616 __ai int32x2_t vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
40617  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40618  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40619  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40620  int32x2_t __ret;
40621  __ret = (int32x2_t) __builtin_neon_vdot_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
40622  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40623  return __ret;
40624 }
40625 __ai int32x2_t __noswap_vdot_s32(int32x2_t __p0, int8x8_t __p1, int8x8_t __p2) {
40626  int32x2_t __ret;
40627  __ret = (int32x2_t) __builtin_neon_vdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
40628  return __ret;
40629 }
40630 #endif
40631 
40632 #ifdef __LITTLE_ENDIAN__
40633 #define vdotq_lane_u32(__p0_169, __p1_169, __p2_169, __p3_169) __extension__ ({ \
40634  uint32x4_t __s0_169 = __p0_169; \
40635  uint8x16_t __s1_169 = __p1_169; \
40636  uint8x8_t __s2_169 = __p2_169; \
40637  uint32x4_t __ret_169; \
40638 uint8x8_t __reint_169 = __s2_169; \
40639 uint32x4_t __reint1_169 = splatq_lane_u32(*(uint32x2_t *) &__reint_169, __p3_169); \
40640  __ret_169 = vdotq_u32(__s0_169, __s1_169, *(uint8x16_t *) &__reint1_169); \
40641  __ret_169; \
40642 })
40643 #else
40644 #define vdotq_lane_u32(__p0_170, __p1_170, __p2_170, __p3_170) __extension__ ({ \
40645  uint32x4_t __s0_170 = __p0_170; \
40646  uint8x16_t __s1_170 = __p1_170; \
40647  uint8x8_t __s2_170 = __p2_170; \
40648  uint32x4_t __rev0_170; __rev0_170 = __builtin_shufflevector(__s0_170, __s0_170, 3, 2, 1, 0); \
40649  uint8x16_t __rev1_170; __rev1_170 = __builtin_shufflevector(__s1_170, __s1_170, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40650  uint8x8_t __rev2_170; __rev2_170 = __builtin_shufflevector(__s2_170, __s2_170, 7, 6, 5, 4, 3, 2, 1, 0); \
40651  uint32x4_t __ret_170; \
40652 uint8x8_t __reint_170 = __rev2_170; \
40653 uint32x4_t __reint1_170 = __noswap_splatq_lane_u32(*(uint32x2_t *) &__reint_170, __p3_170); \
40654  __ret_170 = __noswap_vdotq_u32(__rev0_170, __rev1_170, *(uint8x16_t *) &__reint1_170); \
40655  __ret_170 = __builtin_shufflevector(__ret_170, __ret_170, 3, 2, 1, 0); \
40656  __ret_170; \
40657 })
40658 #endif
40659 
40660 #ifdef __LITTLE_ENDIAN__
40661 #define vdotq_lane_s32(__p0_171, __p1_171, __p2_171, __p3_171) __extension__ ({ \
40662  int32x4_t __s0_171 = __p0_171; \
40663  int8x16_t __s1_171 = __p1_171; \
40664  int8x8_t __s2_171 = __p2_171; \
40665  int32x4_t __ret_171; \
40666 int8x8_t __reint_171 = __s2_171; \
40667 int32x4_t __reint1_171 = splatq_lane_s32(*(int32x2_t *) &__reint_171, __p3_171); \
40668  __ret_171 = vdotq_s32(__s0_171, __s1_171, *(int8x16_t *) &__reint1_171); \
40669  __ret_171; \
40670 })
40671 #else
40672 #define vdotq_lane_s32(__p0_172, __p1_172, __p2_172, __p3_172) __extension__ ({ \
40673  int32x4_t __s0_172 = __p0_172; \
40674  int8x16_t __s1_172 = __p1_172; \
40675  int8x8_t __s2_172 = __p2_172; \
40676  int32x4_t __rev0_172; __rev0_172 = __builtin_shufflevector(__s0_172, __s0_172, 3, 2, 1, 0); \
40677  int8x16_t __rev1_172; __rev1_172 = __builtin_shufflevector(__s1_172, __s1_172, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40678  int8x8_t __rev2_172; __rev2_172 = __builtin_shufflevector(__s2_172, __s2_172, 7, 6, 5, 4, 3, 2, 1, 0); \
40679  int32x4_t __ret_172; \
40680 int8x8_t __reint_172 = __rev2_172; \
40681 int32x4_t __reint1_172 = __noswap_splatq_lane_s32(*(int32x2_t *) &__reint_172, __p3_172); \
40682  __ret_172 = __noswap_vdotq_s32(__rev0_172, __rev1_172, *(int8x16_t *) &__reint1_172); \
40683  __ret_172 = __builtin_shufflevector(__ret_172, __ret_172, 3, 2, 1, 0); \
40684  __ret_172; \
40685 })
40686 #endif
40687 
40688 #ifdef __LITTLE_ENDIAN__
40689 #define vdot_lane_u32(__p0_173, __p1_173, __p2_173, __p3_173) __extension__ ({ \
40690  uint32x2_t __s0_173 = __p0_173; \
40691  uint8x8_t __s1_173 = __p1_173; \
40692  uint8x8_t __s2_173 = __p2_173; \
40693  uint32x2_t __ret_173; \
40694 uint8x8_t __reint_173 = __s2_173; \
40695 uint32x2_t __reint1_173 = splat_lane_u32(*(uint32x2_t *) &__reint_173, __p3_173); \
40696  __ret_173 = vdot_u32(__s0_173, __s1_173, *(uint8x8_t *) &__reint1_173); \
40697  __ret_173; \
40698 })
40699 #else
40700 #define vdot_lane_u32(__p0_174, __p1_174, __p2_174, __p3_174) __extension__ ({ \
40701  uint32x2_t __s0_174 = __p0_174; \
40702  uint8x8_t __s1_174 = __p1_174; \
40703  uint8x8_t __s2_174 = __p2_174; \
40704  uint32x2_t __rev0_174; __rev0_174 = __builtin_shufflevector(__s0_174, __s0_174, 1, 0); \
40705  uint8x8_t __rev1_174; __rev1_174 = __builtin_shufflevector(__s1_174, __s1_174, 7, 6, 5, 4, 3, 2, 1, 0); \
40706  uint8x8_t __rev2_174; __rev2_174 = __builtin_shufflevector(__s2_174, __s2_174, 7, 6, 5, 4, 3, 2, 1, 0); \
40707  uint32x2_t __ret_174; \
40708 uint8x8_t __reint_174 = __rev2_174; \
40709 uint32x2_t __reint1_174 = __noswap_splat_lane_u32(*(uint32x2_t *) &__reint_174, __p3_174); \
40710  __ret_174 = __noswap_vdot_u32(__rev0_174, __rev1_174, *(uint8x8_t *) &__reint1_174); \
40711  __ret_174 = __builtin_shufflevector(__ret_174, __ret_174, 1, 0); \
40712  __ret_174; \
40713 })
40714 #endif
40715 
40716 #ifdef __LITTLE_ENDIAN__
40717 #define vdot_lane_s32(__p0_175, __p1_175, __p2_175, __p3_175) __extension__ ({ \
40718  int32x2_t __s0_175 = __p0_175; \
40719  int8x8_t __s1_175 = __p1_175; \
40720  int8x8_t __s2_175 = __p2_175; \
40721  int32x2_t __ret_175; \
40722 int8x8_t __reint_175 = __s2_175; \
40723 int32x2_t __reint1_175 = splat_lane_s32(*(int32x2_t *) &__reint_175, __p3_175); \
40724  __ret_175 = vdot_s32(__s0_175, __s1_175, *(int8x8_t *) &__reint1_175); \
40725  __ret_175; \
40726 })
40727 #else
40728 #define vdot_lane_s32(__p0_176, __p1_176, __p2_176, __p3_176) __extension__ ({ \
40729  int32x2_t __s0_176 = __p0_176; \
40730  int8x8_t __s1_176 = __p1_176; \
40731  int8x8_t __s2_176 = __p2_176; \
40732  int32x2_t __rev0_176; __rev0_176 = __builtin_shufflevector(__s0_176, __s0_176, 1, 0); \
40733  int8x8_t __rev1_176; __rev1_176 = __builtin_shufflevector(__s1_176, __s1_176, 7, 6, 5, 4, 3, 2, 1, 0); \
40734  int8x8_t __rev2_176; __rev2_176 = __builtin_shufflevector(__s2_176, __s2_176, 7, 6, 5, 4, 3, 2, 1, 0); \
40735  int32x2_t __ret_176; \
40736 int8x8_t __reint_176 = __rev2_176; \
40737 int32x2_t __reint1_176 = __noswap_splat_lane_s32(*(int32x2_t *) &__reint_176, __p3_176); \
40738  __ret_176 = __noswap_vdot_s32(__rev0_176, __rev1_176, *(int8x8_t *) &__reint1_176); \
40739  __ret_176 = __builtin_shufflevector(__ret_176, __ret_176, 1, 0); \
40740  __ret_176; \
40741 })
40742 #endif
40743 
40744 #endif
40745 #if defined(__ARM_FEATURE_DOTPROD) && defined(__aarch64__)
40746 #ifdef __LITTLE_ENDIAN__
40747 #define vdotq_laneq_u32(__p0_177, __p1_177, __p2_177, __p3_177) __extension__ ({ \
40748  uint32x4_t __s0_177 = __p0_177; \
40749  uint8x16_t __s1_177 = __p1_177; \
40750  uint8x16_t __s2_177 = __p2_177; \
40751  uint32x4_t __ret_177; \
40752 uint8x16_t __reint_177 = __s2_177; \
40753 uint32x4_t __reint1_177 = splatq_laneq_u32(*(uint32x4_t *) &__reint_177, __p3_177); \
40754  __ret_177 = vdotq_u32(__s0_177, __s1_177, *(uint8x16_t *) &__reint1_177); \
40755  __ret_177; \
40756 })
40757 #else
40758 #define vdotq_laneq_u32(__p0_178, __p1_178, __p2_178, __p3_178) __extension__ ({ \
40759  uint32x4_t __s0_178 = __p0_178; \
40760  uint8x16_t __s1_178 = __p1_178; \
40761  uint8x16_t __s2_178 = __p2_178; \
40762  uint32x4_t __rev0_178; __rev0_178 = __builtin_shufflevector(__s0_178, __s0_178, 3, 2, 1, 0); \
40763  uint8x16_t __rev1_178; __rev1_178 = __builtin_shufflevector(__s1_178, __s1_178, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40764  uint8x16_t __rev2_178; __rev2_178 = __builtin_shufflevector(__s2_178, __s2_178, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40765  uint32x4_t __ret_178; \
40766 uint8x16_t __reint_178 = __rev2_178; \
40767 uint32x4_t __reint1_178 = __noswap_splatq_laneq_u32(*(uint32x4_t *) &__reint_178, __p3_178); \
40768  __ret_178 = __noswap_vdotq_u32(__rev0_178, __rev1_178, *(uint8x16_t *) &__reint1_178); \
40769  __ret_178 = __builtin_shufflevector(__ret_178, __ret_178, 3, 2, 1, 0); \
40770  __ret_178; \
40771 })
40772 #endif
40773 
40774 #ifdef __LITTLE_ENDIAN__
40775 #define vdotq_laneq_s32(__p0_179, __p1_179, __p2_179, __p3_179) __extension__ ({ \
40776  int32x4_t __s0_179 = __p0_179; \
40777  int8x16_t __s1_179 = __p1_179; \
40778  int8x16_t __s2_179 = __p2_179; \
40779  int32x4_t __ret_179; \
40780 int8x16_t __reint_179 = __s2_179; \
40781 int32x4_t __reint1_179 = splatq_laneq_s32(*(int32x4_t *) &__reint_179, __p3_179); \
40782  __ret_179 = vdotq_s32(__s0_179, __s1_179, *(int8x16_t *) &__reint1_179); \
40783  __ret_179; \
40784 })
40785 #else
40786 #define vdotq_laneq_s32(__p0_180, __p1_180, __p2_180, __p3_180) __extension__ ({ \
40787  int32x4_t __s0_180 = __p0_180; \
40788  int8x16_t __s1_180 = __p1_180; \
40789  int8x16_t __s2_180 = __p2_180; \
40790  int32x4_t __rev0_180; __rev0_180 = __builtin_shufflevector(__s0_180, __s0_180, 3, 2, 1, 0); \
40791  int8x16_t __rev1_180; __rev1_180 = __builtin_shufflevector(__s1_180, __s1_180, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40792  int8x16_t __rev2_180; __rev2_180 = __builtin_shufflevector(__s2_180, __s2_180, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40793  int32x4_t __ret_180; \
40794 int8x16_t __reint_180 = __rev2_180; \
40795 int32x4_t __reint1_180 = __noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_180, __p3_180); \
40796  __ret_180 = __noswap_vdotq_s32(__rev0_180, __rev1_180, *(int8x16_t *) &__reint1_180); \
40797  __ret_180 = __builtin_shufflevector(__ret_180, __ret_180, 3, 2, 1, 0); \
40798  __ret_180; \
40799 })
40800 #endif
40801 
40802 #ifdef __LITTLE_ENDIAN__
40803 #define vdot_laneq_u32(__p0_181, __p1_181, __p2_181, __p3_181) __extension__ ({ \
40804  uint32x2_t __s0_181 = __p0_181; \
40805  uint8x8_t __s1_181 = __p1_181; \
40806  uint8x16_t __s2_181 = __p2_181; \
40807  uint32x2_t __ret_181; \
40808 uint8x16_t __reint_181 = __s2_181; \
40809 uint32x2_t __reint1_181 = splat_laneq_u32(*(uint32x4_t *) &__reint_181, __p3_181); \
40810  __ret_181 = vdot_u32(__s0_181, __s1_181, *(uint8x8_t *) &__reint1_181); \
40811  __ret_181; \
40812 })
40813 #else
40814 #define vdot_laneq_u32(__p0_182, __p1_182, __p2_182, __p3_182) __extension__ ({ \
40815  uint32x2_t __s0_182 = __p0_182; \
40816  uint8x8_t __s1_182 = __p1_182; \
40817  uint8x16_t __s2_182 = __p2_182; \
40818  uint32x2_t __rev0_182; __rev0_182 = __builtin_shufflevector(__s0_182, __s0_182, 1, 0); \
40819  uint8x8_t __rev1_182; __rev1_182 = __builtin_shufflevector(__s1_182, __s1_182, 7, 6, 5, 4, 3, 2, 1, 0); \
40820  uint8x16_t __rev2_182; __rev2_182 = __builtin_shufflevector(__s2_182, __s2_182, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40821  uint32x2_t __ret_182; \
40822 uint8x16_t __reint_182 = __rev2_182; \
40823 uint32x2_t __reint1_182 = __noswap_splat_laneq_u32(*(uint32x4_t *) &__reint_182, __p3_182); \
40824  __ret_182 = __noswap_vdot_u32(__rev0_182, __rev1_182, *(uint8x8_t *) &__reint1_182); \
40825  __ret_182 = __builtin_shufflevector(__ret_182, __ret_182, 1, 0); \
40826  __ret_182; \
40827 })
40828 #endif
40829 
40830 #ifdef __LITTLE_ENDIAN__
40831 #define vdot_laneq_s32(__p0_183, __p1_183, __p2_183, __p3_183) __extension__ ({ \
40832  int32x2_t __s0_183 = __p0_183; \
40833  int8x8_t __s1_183 = __p1_183; \
40834  int8x16_t __s2_183 = __p2_183; \
40835  int32x2_t __ret_183; \
40836 int8x16_t __reint_183 = __s2_183; \
40837 int32x2_t __reint1_183 = splat_laneq_s32(*(int32x4_t *) &__reint_183, __p3_183); \
40838  __ret_183 = vdot_s32(__s0_183, __s1_183, *(int8x8_t *) &__reint1_183); \
40839  __ret_183; \
40840 })
40841 #else
40842 #define vdot_laneq_s32(__p0_184, __p1_184, __p2_184, __p3_184) __extension__ ({ \
40843  int32x2_t __s0_184 = __p0_184; \
40844  int8x8_t __s1_184 = __p1_184; \
40845  int8x16_t __s2_184 = __p2_184; \
40846  int32x2_t __rev0_184; __rev0_184 = __builtin_shufflevector(__s0_184, __s0_184, 1, 0); \
40847  int8x8_t __rev1_184; __rev1_184 = __builtin_shufflevector(__s1_184, __s1_184, 7, 6, 5, 4, 3, 2, 1, 0); \
40848  int8x16_t __rev2_184; __rev2_184 = __builtin_shufflevector(__s2_184, __s2_184, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
40849  int32x2_t __ret_184; \
40850 int8x16_t __reint_184 = __rev2_184; \
40851 int32x2_t __reint1_184 = __noswap_splat_laneq_s32(*(int32x4_t *) &__reint_184, __p3_184); \
40852  __ret_184 = __noswap_vdot_s32(__rev0_184, __rev1_184, *(int8x8_t *) &__reint1_184); \
40853  __ret_184 = __builtin_shufflevector(__ret_184, __ret_184, 1, 0); \
40854  __ret_184; \
40855 })
40856 #endif
40857 
40858 #endif
40859 #if defined(__ARM_FEATURE_FMA)
40860 #ifdef __LITTLE_ENDIAN__
40861 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40862  float32x4_t __ret;
40863  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40864  return __ret;
40865 }
40866 #else
40867 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40868  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40869  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40870  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40871  float32x4_t __ret;
40872  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
40873  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40874  return __ret;
40875 }
40876 __ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40877  float32x4_t __ret;
40878  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40879  return __ret;
40880 }
40881 #endif
40882 
40883 #ifdef __LITTLE_ENDIAN__
40884 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40885  float32x2_t __ret;
40886  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
40887  return __ret;
40888 }
40889 #else
40890 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40891  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40892  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40893  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40894  float32x2_t __ret;
40895  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
40896  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40897  return __ret;
40898 }
40899 __ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40900  float32x2_t __ret;
40901  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
40902  return __ret;
40903 }
40904 #endif
40905 
40906 #ifdef __LITTLE_ENDIAN__
40907 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
40908  float32x4_t __ret;
40909  __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
40910  return __ret;
40911 }
40912 #else
40913 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
40914  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40915  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40916  float32x4_t __ret;
40917  __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
40918  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40919  return __ret;
40920 }
40921 #endif
40922 
40923 #ifdef __LITTLE_ENDIAN__
40924 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
40925  float32x2_t __ret;
40926  __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
40927  return __ret;
40928 }
40929 #else
40930 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
40931  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40932  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40933  float32x2_t __ret;
40934  __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
40935  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40936  return __ret;
40937 }
40938 #endif
40939 
40940 #ifdef __LITTLE_ENDIAN__
40941 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40942  float32x4_t __ret;
40943  __ret = vfmaq_f32(__p0, -__p1, __p2);
40944  return __ret;
40945 }
40946 #else
40947 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
40948  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40949  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40950  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40951  float32x4_t __ret;
40952  __ret = __noswap_vfmaq_f32(__rev0, -__rev1, __rev2);
40953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40954  return __ret;
40955 }
40956 #endif
40957 
40958 #ifdef __LITTLE_ENDIAN__
40959 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40960  float32x2_t __ret;
40961  __ret = vfma_f32(__p0, -__p1, __p2);
40962  return __ret;
40963 }
40964 #else
40965 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
40966  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40967  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40968  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40969  float32x2_t __ret;
40970  __ret = __noswap_vfma_f32(__rev0, -__rev1, __rev2);
40971  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40972  return __ret;
40973 }
40974 #endif
40975 
40976 #endif
40977 #if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__)
40978 #ifdef __LITTLE_ENDIAN__
40979 __ai float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
40980  float32x4_t __ret;
40981  __ret = (float32x4_t) __builtin_neon_vfmlalq_high_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40982  return __ret;
40983 }
40984 #else
40985 __ai float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
40986  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40987  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40988  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40989  float32x4_t __ret;
40990  __ret = (float32x4_t) __builtin_neon_vfmlalq_high_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
40991  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40992  return __ret;
40993 }
40994 __ai float32x4_t __noswap_vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
40995  float32x4_t __ret;
40996  __ret = (float32x4_t) __builtin_neon_vfmlalq_high_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
40997  return __ret;
40998 }
40999 #endif
41000 
41001 #ifdef __LITTLE_ENDIAN__
41002 __ai float32x2_t vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41003  float32x2_t __ret;
41004  __ret = (float32x2_t) __builtin_neon_vfmlal_high_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41005  return __ret;
41006 }
41007 #else
41008 __ai float32x2_t vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41009  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41010  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41011  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41012  float32x2_t __ret;
41013  __ret = (float32x2_t) __builtin_neon_vfmlal_high_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
41014  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41015  return __ret;
41016 }
41017 __ai float32x2_t __noswap_vfmlal_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41018  float32x2_t __ret;
41019  __ret = (float32x2_t) __builtin_neon_vfmlal_high_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41020  return __ret;
41021 }
41022 #endif
41023 
41024 #ifdef __LITTLE_ENDIAN__
41025 __ai float32x4_t vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41026  float32x4_t __ret;
41027  __ret = (float32x4_t) __builtin_neon_vfmlalq_low_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41028  return __ret;
41029 }
41030 #else
41031 __ai float32x4_t vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41032  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41033  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41034  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41035  float32x4_t __ret;
41036  __ret = (float32x4_t) __builtin_neon_vfmlalq_low_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
41037  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41038  return __ret;
41039 }
41040 __ai float32x4_t __noswap_vfmlalq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41041  float32x4_t __ret;
41042  __ret = (float32x4_t) __builtin_neon_vfmlalq_low_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41043  return __ret;
41044 }
41045 #endif
41046 
41047 #ifdef __LITTLE_ENDIAN__
41048 __ai float32x2_t vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41049  float32x2_t __ret;
41050  __ret = (float32x2_t) __builtin_neon_vfmlal_low_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41051  return __ret;
41052 }
41053 #else
41054 __ai float32x2_t vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41055  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41056  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41057  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41058  float32x2_t __ret;
41059  __ret = (float32x2_t) __builtin_neon_vfmlal_low_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
41060  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41061  return __ret;
41062 }
41063 __ai float32x2_t __noswap_vfmlal_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41064  float32x2_t __ret;
41065  __ret = (float32x2_t) __builtin_neon_vfmlal_low_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41066  return __ret;
41067 }
41068 #endif
41069 
41070 #ifdef __LITTLE_ENDIAN__
41071 __ai float32x4_t vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41072  float32x4_t __ret;
41073  __ret = (float32x4_t) __builtin_neon_vfmlslq_high_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41074  return __ret;
41075 }
41076 #else
41077 __ai float32x4_t vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41078  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41079  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41080  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41081  float32x4_t __ret;
41082  __ret = (float32x4_t) __builtin_neon_vfmlslq_high_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
41083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41084  return __ret;
41085 }
41086 __ai float32x4_t __noswap_vfmlslq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41087  float32x4_t __ret;
41088  __ret = (float32x4_t) __builtin_neon_vfmlslq_high_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41089  return __ret;
41090 }
41091 #endif
41092 
41093 #ifdef __LITTLE_ENDIAN__
41094 __ai float32x2_t vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41095  float32x2_t __ret;
41096  __ret = (float32x2_t) __builtin_neon_vfmlsl_high_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41097  return __ret;
41098 }
41099 #else
41100 __ai float32x2_t vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41101  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41102  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41103  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41104  float32x2_t __ret;
41105  __ret = (float32x2_t) __builtin_neon_vfmlsl_high_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
41106  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41107  return __ret;
41108 }
41109 __ai float32x2_t __noswap_vfmlsl_high_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41110  float32x2_t __ret;
41111  __ret = (float32x2_t) __builtin_neon_vfmlsl_high_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41112  return __ret;
41113 }
41114 #endif
41115 
41116 #ifdef __LITTLE_ENDIAN__
41117 __ai float32x4_t vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41118  float32x4_t __ret;
41119  __ret = (float32x4_t) __builtin_neon_vfmlslq_low_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41120  return __ret;
41121 }
41122 #else
41123 __ai float32x4_t vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41124  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41125  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41126  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41127  float32x4_t __ret;
41128  __ret = (float32x4_t) __builtin_neon_vfmlslq_low_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
41129  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41130  return __ret;
41131 }
41132 __ai float32x4_t __noswap_vfmlslq_low_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) {
41133  float32x4_t __ret;
41134  __ret = (float32x4_t) __builtin_neon_vfmlslq_low_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
41135  return __ret;
41136 }
41137 #endif
41138 
41139 #ifdef __LITTLE_ENDIAN__
41140 __ai float32x2_t vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41141  float32x2_t __ret;
41142  __ret = (float32x2_t) __builtin_neon_vfmlsl_low_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41143  return __ret;
41144 }
41145 #else
41146 __ai float32x2_t vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41147  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41148  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41149  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41150  float32x2_t __ret;
41151  __ret = (float32x2_t) __builtin_neon_vfmlsl_low_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
41152  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41153  return __ret;
41154 }
41155 __ai float32x2_t __noswap_vfmlsl_low_f16(float32x2_t __p0, float16x4_t __p1, float16x4_t __p2) {
41156  float32x2_t __ret;
41157  __ret = (float32x2_t) __builtin_neon_vfmlsl_low_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
41158  return __ret;
41159 }
41160 #endif
41161 
41162 #endif
41163 #if defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC)
41164 #ifdef __LITTLE_ENDIAN__
41165 __ai float16x8_t vabdq_f16(float16x8_t __p0, float16x8_t __p1) {
41166  float16x8_t __ret;
41167  __ret = (float16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
41168  return __ret;
41169 }
41170 #else
41171 __ai float16x8_t vabdq_f16(float16x8_t __p0, float16x8_t __p1) {
41172  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41173  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41174  float16x8_t __ret;
41175  __ret = (float16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
41176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41177  return __ret;
41178 }
41179 #endif
41180 
41181 #ifdef __LITTLE_ENDIAN__
41182 __ai float16x4_t vabd_f16(float16x4_t __p0, float16x4_t __p1) {
41183  float16x4_t __ret;
41184  __ret = (float16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
41185  return __ret;
41186 }
41187 #else
41188 __ai float16x4_t vabd_f16(float16x4_t __p0, float16x4_t __p1) {
41189  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41190  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41191  float16x4_t __ret;
41192  __ret = (float16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
41193  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41194  return __ret;
41195 }
41196 #endif
41197 
41198 #ifdef __LITTLE_ENDIAN__
41199 __ai float16x8_t vabsq_f16(float16x8_t __p0) {
41200  float16x8_t __ret;
41201  __ret = (float16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 40);
41202  return __ret;
41203 }
41204 #else
41205 __ai float16x8_t vabsq_f16(float16x8_t __p0) {
41206  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41207  float16x8_t __ret;
41208  __ret = (float16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 40);
41209  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41210  return __ret;
41211 }
41212 #endif
41213 
41214 #ifdef __LITTLE_ENDIAN__
41215 __ai float16x4_t vabs_f16(float16x4_t __p0) {
41216  float16x4_t __ret;
41217  __ret = (float16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 8);
41218  return __ret;
41219 }
41220 #else
41221 __ai float16x4_t vabs_f16(float16x4_t __p0) {
41222  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41223  float16x4_t __ret;
41224  __ret = (float16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 8);
41225  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41226  return __ret;
41227 }
41228 #endif
41229 
41230 #ifdef __LITTLE_ENDIAN__
41231 __ai float16x8_t vaddq_f16(float16x8_t __p0, float16x8_t __p1) {
41232  float16x8_t __ret;
41233  __ret = __p0 + __p1;
41234  return __ret;
41235 }
41236 #else
41237 __ai float16x8_t vaddq_f16(float16x8_t __p0, float16x8_t __p1) {
41238  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41239  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41240  float16x8_t __ret;
41241  __ret = __rev0 + __rev1;
41242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41243  return __ret;
41244 }
41245 #endif
41246 
41247 #ifdef __LITTLE_ENDIAN__
41248 __ai float16x4_t vadd_f16(float16x4_t __p0, float16x4_t __p1) {
41249  float16x4_t __ret;
41250  __ret = __p0 + __p1;
41251  return __ret;
41252 }
41253 #else
41254 __ai float16x4_t vadd_f16(float16x4_t __p0, float16x4_t __p1) {
41255  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41256  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41257  float16x4_t __ret;
41258  __ret = __rev0 + __rev1;
41259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41260  return __ret;
41261 }
41262 #endif
41263 
41264 #ifdef __LITTLE_ENDIAN__
41265 __ai float16x8_t vbslq_f16(uint16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
41266  float16x8_t __ret;
41267  __ret = (float16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
41268  return __ret;
41269 }
41270 #else
41271 __ai float16x8_t vbslq_f16(uint16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
41272  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41273  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41274  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41275  float16x8_t __ret;
41276  __ret = (float16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
41277  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41278  return __ret;
41279 }
41280 #endif
41281 
41282 #ifdef __LITTLE_ENDIAN__
41283 __ai float16x4_t vbsl_f16(uint16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
41284  float16x4_t __ret;
41285  __ret = (float16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
41286  return __ret;
41287 }
41288 #else
41289 __ai float16x4_t vbsl_f16(uint16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
41290  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41291  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41292  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41293  float16x4_t __ret;
41294  __ret = (float16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
41295  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41296  return __ret;
41297 }
41298 #endif
41299 
41300 #ifdef __LITTLE_ENDIAN__
41301 __ai uint16x8_t vcageq_f16(float16x8_t __p0, float16x8_t __p1) {
41302  uint16x8_t __ret;
41303  __ret = (uint16x8_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
41304  return __ret;
41305 }
41306 #else
41307 __ai uint16x8_t vcageq_f16(float16x8_t __p0, float16x8_t __p1) {
41308  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41309  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41310  uint16x8_t __ret;
41311  __ret = (uint16x8_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
41312  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41313  return __ret;
41314 }
41315 #endif
41316 
41317 #ifdef __LITTLE_ENDIAN__
41318 __ai uint16x4_t vcage_f16(float16x4_t __p0, float16x4_t __p1) {
41319  uint16x4_t __ret;
41320  __ret = (uint16x4_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
41321  return __ret;
41322 }
41323 #else
41324 __ai uint16x4_t vcage_f16(float16x4_t __p0, float16x4_t __p1) {
41325  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41326  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41327  uint16x4_t __ret;
41328  __ret = (uint16x4_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
41329  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41330  return __ret;
41331 }
41332 #endif
41333 
41334 #ifdef __LITTLE_ENDIAN__
41335 __ai uint16x8_t vcagtq_f16(float16x8_t __p0, float16x8_t __p1) {
41336  uint16x8_t __ret;
41337  __ret = (uint16x8_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
41338  return __ret;
41339 }
41340 #else
41341 __ai uint16x8_t vcagtq_f16(float16x8_t __p0, float16x8_t __p1) {
41342  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41343  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41344  uint16x8_t __ret;
41345  __ret = (uint16x8_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
41346  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41347  return __ret;
41348 }
41349 #endif
41350 
41351 #ifdef __LITTLE_ENDIAN__
41352 __ai uint16x4_t vcagt_f16(float16x4_t __p0, float16x4_t __p1) {
41353  uint16x4_t __ret;
41354  __ret = (uint16x4_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
41355  return __ret;
41356 }
41357 #else
41358 __ai uint16x4_t vcagt_f16(float16x4_t __p0, float16x4_t __p1) {
41359  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41360  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41361  uint16x4_t __ret;
41362  __ret = (uint16x4_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
41363  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41364  return __ret;
41365 }
41366 #endif
41367 
41368 #ifdef __LITTLE_ENDIAN__
41369 __ai uint16x8_t vcaleq_f16(float16x8_t __p0, float16x8_t __p1) {
41370  uint16x8_t __ret;
41371  __ret = (uint16x8_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
41372  return __ret;
41373 }
41374 #else
41375 __ai uint16x8_t vcaleq_f16(float16x8_t __p0, float16x8_t __p1) {
41376  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41377  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41378  uint16x8_t __ret;
41379  __ret = (uint16x8_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
41380  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41381  return __ret;
41382 }
41383 #endif
41384 
41385 #ifdef __LITTLE_ENDIAN__
41386 __ai uint16x4_t vcale_f16(float16x4_t __p0, float16x4_t __p1) {
41387  uint16x4_t __ret;
41388  __ret = (uint16x4_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
41389  return __ret;
41390 }
41391 #else
41392 __ai uint16x4_t vcale_f16(float16x4_t __p0, float16x4_t __p1) {
41393  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41394  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41395  uint16x4_t __ret;
41396  __ret = (uint16x4_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
41397  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41398  return __ret;
41399 }
41400 #endif
41401 
41402 #ifdef __LITTLE_ENDIAN__
41403 __ai uint16x8_t vcaltq_f16(float16x8_t __p0, float16x8_t __p1) {
41404  uint16x8_t __ret;
41405  __ret = (uint16x8_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
41406  return __ret;
41407 }
41408 #else
41409 __ai uint16x8_t vcaltq_f16(float16x8_t __p0, float16x8_t __p1) {
41410  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41411  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41412  uint16x8_t __ret;
41413  __ret = (uint16x8_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
41414  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41415  return __ret;
41416 }
41417 #endif
41418 
41419 #ifdef __LITTLE_ENDIAN__
41420 __ai uint16x4_t vcalt_f16(float16x4_t __p0, float16x4_t __p1) {
41421  uint16x4_t __ret;
41422  __ret = (uint16x4_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
41423  return __ret;
41424 }
41425 #else
41426 __ai uint16x4_t vcalt_f16(float16x4_t __p0, float16x4_t __p1) {
41427  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41428  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41429  uint16x4_t __ret;
41430  __ret = (uint16x4_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
41431  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41432  return __ret;
41433 }
41434 #endif
41435 
41436 #ifdef __LITTLE_ENDIAN__
41437 __ai uint16x8_t vceqq_f16(float16x8_t __p0, float16x8_t __p1) {
41438  uint16x8_t __ret;
41439  __ret = (uint16x8_t)(__p0 == __p1);
41440  return __ret;
41441 }
41442 #else
41443 __ai uint16x8_t vceqq_f16(float16x8_t __p0, float16x8_t __p1) {
41444  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41445  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41446  uint16x8_t __ret;
41447  __ret = (uint16x8_t)(__rev0 == __rev1);
41448  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41449  return __ret;
41450 }
41451 #endif
41452 
41453 #ifdef __LITTLE_ENDIAN__
41454 __ai uint16x4_t vceq_f16(float16x4_t __p0, float16x4_t __p1) {
41455  uint16x4_t __ret;
41456  __ret = (uint16x4_t)(__p0 == __p1);
41457  return __ret;
41458 }
41459 #else
41460 __ai uint16x4_t vceq_f16(float16x4_t __p0, float16x4_t __p1) {
41461  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41462  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41463  uint16x4_t __ret;
41464  __ret = (uint16x4_t)(__rev0 == __rev1);
41465  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41466  return __ret;
41467 }
41468 #endif
41469 
41470 #ifdef __LITTLE_ENDIAN__
41471 __ai uint16x8_t vceqzq_f16(float16x8_t __p0) {
41472  uint16x8_t __ret;
41473  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
41474  return __ret;
41475 }
41476 #else
41477 __ai uint16x8_t vceqzq_f16(float16x8_t __p0) {
41478  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41479  uint16x8_t __ret;
41480  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
41481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41482  return __ret;
41483 }
41484 #endif
41485 
41486 #ifdef __LITTLE_ENDIAN__
41487 __ai uint16x4_t vceqz_f16(float16x4_t __p0) {
41488  uint16x4_t __ret;
41489  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
41490  return __ret;
41491 }
41492 #else
41493 __ai uint16x4_t vceqz_f16(float16x4_t __p0) {
41494  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41495  uint16x4_t __ret;
41496  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
41497  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41498  return __ret;
41499 }
41500 #endif
41501 
41502 #ifdef __LITTLE_ENDIAN__
41503 __ai uint16x8_t vcgeq_f16(float16x8_t __p0, float16x8_t __p1) {
41504  uint16x8_t __ret;
41505  __ret = (uint16x8_t)(__p0 >= __p1);
41506  return __ret;
41507 }
41508 #else
41509 __ai uint16x8_t vcgeq_f16(float16x8_t __p0, float16x8_t __p1) {
41510  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41511  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41512  uint16x8_t __ret;
41513  __ret = (uint16x8_t)(__rev0 >= __rev1);
41514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41515  return __ret;
41516 }
41517 #endif
41518 
41519 #ifdef __LITTLE_ENDIAN__
41520 __ai uint16x4_t vcge_f16(float16x4_t __p0, float16x4_t __p1) {
41521  uint16x4_t __ret;
41522  __ret = (uint16x4_t)(__p0 >= __p1);
41523  return __ret;
41524 }
41525 #else
41526 __ai uint16x4_t vcge_f16(float16x4_t __p0, float16x4_t __p1) {
41527  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41528  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41529  uint16x4_t __ret;
41530  __ret = (uint16x4_t)(__rev0 >= __rev1);
41531  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41532  return __ret;
41533 }
41534 #endif
41535 
41536 #ifdef __LITTLE_ENDIAN__
41537 __ai uint16x8_t vcgezq_f16(float16x8_t __p0) {
41538  uint16x8_t __ret;
41539  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
41540  return __ret;
41541 }
41542 #else
41543 __ai uint16x8_t vcgezq_f16(float16x8_t __p0) {
41544  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41545  uint16x8_t __ret;
41546  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
41547  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41548  return __ret;
41549 }
41550 #endif
41551 
41552 #ifdef __LITTLE_ENDIAN__
41553 __ai uint16x4_t vcgez_f16(float16x4_t __p0) {
41554  uint16x4_t __ret;
41555  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
41556  return __ret;
41557 }
41558 #else
41559 __ai uint16x4_t vcgez_f16(float16x4_t __p0) {
41560  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41561  uint16x4_t __ret;
41562  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
41563  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41564  return __ret;
41565 }
41566 #endif
41567 
41568 #ifdef __LITTLE_ENDIAN__
41569 __ai uint16x8_t vcgtq_f16(float16x8_t __p0, float16x8_t __p1) {
41570  uint16x8_t __ret;
41571  __ret = (uint16x8_t)(__p0 > __p1);
41572  return __ret;
41573 }
41574 #else
41575 __ai uint16x8_t vcgtq_f16(float16x8_t __p0, float16x8_t __p1) {
41576  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41577  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41578  uint16x8_t __ret;
41579  __ret = (uint16x8_t)(__rev0 > __rev1);
41580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41581  return __ret;
41582 }
41583 #endif
41584 
41585 #ifdef __LITTLE_ENDIAN__
41586 __ai uint16x4_t vcgt_f16(float16x4_t __p0, float16x4_t __p1) {
41587  uint16x4_t __ret;
41588  __ret = (uint16x4_t)(__p0 > __p1);
41589  return __ret;
41590 }
41591 #else
41592 __ai uint16x4_t vcgt_f16(float16x4_t __p0, float16x4_t __p1) {
41593  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41594  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41595  uint16x4_t __ret;
41596  __ret = (uint16x4_t)(__rev0 > __rev1);
41597  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41598  return __ret;
41599 }
41600 #endif
41601 
41602 #ifdef __LITTLE_ENDIAN__
41603 __ai uint16x8_t vcgtzq_f16(float16x8_t __p0) {
41604  uint16x8_t __ret;
41605  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
41606  return __ret;
41607 }
41608 #else
41609 __ai uint16x8_t vcgtzq_f16(float16x8_t __p0) {
41610  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41611  uint16x8_t __ret;
41612  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
41613  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41614  return __ret;
41615 }
41616 #endif
41617 
41618 #ifdef __LITTLE_ENDIAN__
41619 __ai uint16x4_t vcgtz_f16(float16x4_t __p0) {
41620  uint16x4_t __ret;
41621  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
41622  return __ret;
41623 }
41624 #else
41625 __ai uint16x4_t vcgtz_f16(float16x4_t __p0) {
41626  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41627  uint16x4_t __ret;
41628  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
41629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41630  return __ret;
41631 }
41632 #endif
41633 
41634 #ifdef __LITTLE_ENDIAN__
41635 __ai uint16x8_t vcleq_f16(float16x8_t __p0, float16x8_t __p1) {
41636  uint16x8_t __ret;
41637  __ret = (uint16x8_t)(__p0 <= __p1);
41638  return __ret;
41639 }
41640 #else
41641 __ai uint16x8_t vcleq_f16(float16x8_t __p0, float16x8_t __p1) {
41642  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41643  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41644  uint16x8_t __ret;
41645  __ret = (uint16x8_t)(__rev0 <= __rev1);
41646  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41647  return __ret;
41648 }
41649 #endif
41650 
41651 #ifdef __LITTLE_ENDIAN__
41652 __ai uint16x4_t vcle_f16(float16x4_t __p0, float16x4_t __p1) {
41653  uint16x4_t __ret;
41654  __ret = (uint16x4_t)(__p0 <= __p1);
41655  return __ret;
41656 }
41657 #else
41658 __ai uint16x4_t vcle_f16(float16x4_t __p0, float16x4_t __p1) {
41659  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41660  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41661  uint16x4_t __ret;
41662  __ret = (uint16x4_t)(__rev0 <= __rev1);
41663  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41664  return __ret;
41665 }
41666 #endif
41667 
41668 #ifdef __LITTLE_ENDIAN__
41669 __ai uint16x8_t vclezq_f16(float16x8_t __p0) {
41670  uint16x8_t __ret;
41671  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
41672  return __ret;
41673 }
41674 #else
41675 __ai uint16x8_t vclezq_f16(float16x8_t __p0) {
41676  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41677  uint16x8_t __ret;
41678  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
41679  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41680  return __ret;
41681 }
41682 #endif
41683 
41684 #ifdef __LITTLE_ENDIAN__
41685 __ai uint16x4_t vclez_f16(float16x4_t __p0) {
41686  uint16x4_t __ret;
41687  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
41688  return __ret;
41689 }
41690 #else
41691 __ai uint16x4_t vclez_f16(float16x4_t __p0) {
41692  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41693  uint16x4_t __ret;
41694  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
41695  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41696  return __ret;
41697 }
41698 #endif
41699 
41700 #ifdef __LITTLE_ENDIAN__
41701 __ai uint16x8_t vcltq_f16(float16x8_t __p0, float16x8_t __p1) {
41702  uint16x8_t __ret;
41703  __ret = (uint16x8_t)(__p0 < __p1);
41704  return __ret;
41705 }
41706 #else
41707 __ai uint16x8_t vcltq_f16(float16x8_t __p0, float16x8_t __p1) {
41708  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41709  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41710  uint16x8_t __ret;
41711  __ret = (uint16x8_t)(__rev0 < __rev1);
41712  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41713  return __ret;
41714 }
41715 #endif
41716 
41717 #ifdef __LITTLE_ENDIAN__
41718 __ai uint16x4_t vclt_f16(float16x4_t __p0, float16x4_t __p1) {
41719  uint16x4_t __ret;
41720  __ret = (uint16x4_t)(__p0 < __p1);
41721  return __ret;
41722 }
41723 #else
41724 __ai uint16x4_t vclt_f16(float16x4_t __p0, float16x4_t __p1) {
41725  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41726  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41727  uint16x4_t __ret;
41728  __ret = (uint16x4_t)(__rev0 < __rev1);
41729  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41730  return __ret;
41731 }
41732 #endif
41733 
41734 #ifdef __LITTLE_ENDIAN__
41735 __ai uint16x8_t vcltzq_f16(float16x8_t __p0) {
41736  uint16x8_t __ret;
41737  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
41738  return __ret;
41739 }
41740 #else
41741 __ai uint16x8_t vcltzq_f16(float16x8_t __p0) {
41742  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41743  uint16x8_t __ret;
41744  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
41745  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41746  return __ret;
41747 }
41748 #endif
41749 
41750 #ifdef __LITTLE_ENDIAN__
41751 __ai uint16x4_t vcltz_f16(float16x4_t __p0) {
41752  uint16x4_t __ret;
41753  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
41754  return __ret;
41755 }
41756 #else
41757 __ai uint16x4_t vcltz_f16(float16x4_t __p0) {
41758  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41759  uint16x4_t __ret;
41760  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
41761  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41762  return __ret;
41763 }
41764 #endif
41765 
41766 #ifdef __LITTLE_ENDIAN__
41767 __ai float16x8_t vcvtq_f16_u16(uint16x8_t __p0) {
41768  float16x8_t __ret;
41769  __ret = (float16x8_t) __builtin_neon_vcvtq_f16_v((int8x16_t)__p0, 49);
41770  return __ret;
41771 }
41772 #else
41773 __ai float16x8_t vcvtq_f16_u16(uint16x8_t __p0) {
41774  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41775  float16x8_t __ret;
41776  __ret = (float16x8_t) __builtin_neon_vcvtq_f16_v((int8x16_t)__rev0, 49);
41777  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41778  return __ret;
41779 }
41780 #endif
41781 
41782 #ifdef __LITTLE_ENDIAN__
41783 __ai float16x8_t vcvtq_f16_s16(int16x8_t __p0) {
41784  float16x8_t __ret;
41785  __ret = (float16x8_t) __builtin_neon_vcvtq_f16_v((int8x16_t)__p0, 33);
41786  return __ret;
41787 }
41788 #else
41789 __ai float16x8_t vcvtq_f16_s16(int16x8_t __p0) {
41790  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41791  float16x8_t __ret;
41792  __ret = (float16x8_t) __builtin_neon_vcvtq_f16_v((int8x16_t)__rev0, 33);
41793  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41794  return __ret;
41795 }
41796 #endif
41797 
41798 #ifdef __LITTLE_ENDIAN__
41799 __ai float16x4_t vcvt_f16_u16(uint16x4_t __p0) {
41800  float16x4_t __ret;
41801  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x8_t)__p0, 17);
41802  return __ret;
41803 }
41804 #else
41805 __ai float16x4_t vcvt_f16_u16(uint16x4_t __p0) {
41806  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41807  float16x4_t __ret;
41808  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x8_t)__rev0, 17);
41809  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41810  return __ret;
41811 }
41812 #endif
41813 
41814 #ifdef __LITTLE_ENDIAN__
41815 __ai float16x4_t vcvt_f16_s16(int16x4_t __p0) {
41816  float16x4_t __ret;
41817  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x8_t)__p0, 1);
41818  return __ret;
41819 }
41820 #else
41821 __ai float16x4_t vcvt_f16_s16(int16x4_t __p0) {
41822  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41823  float16x4_t __ret;
41824  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x8_t)__rev0, 1);
41825  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41826  return __ret;
41827 }
41828 #endif
41829 
41830 #ifdef __LITTLE_ENDIAN__
41831 #define vcvtq_n_f16_u16(__p0, __p1) __extension__ ({ \
41832  uint16x8_t __s0 = __p0; \
41833  float16x8_t __ret; \
41834  __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_v((int8x16_t)__s0, __p1, 49); \
41835  __ret; \
41836 })
41837 #else
41838 #define vcvtq_n_f16_u16(__p0, __p1) __extension__ ({ \
41839  uint16x8_t __s0 = __p0; \
41840  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
41841  float16x8_t __ret; \
41842  __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_v((int8x16_t)__rev0, __p1, 49); \
41843  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
41844  __ret; \
41845 })
41846 #endif
41847 
41848 #ifdef __LITTLE_ENDIAN__
41849 #define vcvtq_n_f16_s16(__p0, __p1) __extension__ ({ \
41850  int16x8_t __s0 = __p0; \
41851  float16x8_t __ret; \
41852  __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_v((int8x16_t)__s0, __p1, 33); \
41853  __ret; \
41854 })
41855 #else
41856 #define vcvtq_n_f16_s16(__p0, __p1) __extension__ ({ \
41857  int16x8_t __s0 = __p0; \
41858  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
41859  float16x8_t __ret; \
41860  __ret = (float16x8_t) __builtin_neon_vcvtq_n_f16_v((int8x16_t)__rev0, __p1, 33); \
41861  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
41862  __ret; \
41863 })
41864 #endif
41865 
41866 #ifdef __LITTLE_ENDIAN__
41867 #define vcvt_n_f16_u16(__p0, __p1) __extension__ ({ \
41868  uint16x4_t __s0 = __p0; \
41869  float16x4_t __ret; \
41870  __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_v((int8x8_t)__s0, __p1, 17); \
41871  __ret; \
41872 })
41873 #else
41874 #define vcvt_n_f16_u16(__p0, __p1) __extension__ ({ \
41875  uint16x4_t __s0 = __p0; \
41876  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
41877  float16x4_t __ret; \
41878  __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_v((int8x8_t)__rev0, __p1, 17); \
41879  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
41880  __ret; \
41881 })
41882 #endif
41883 
41884 #ifdef __LITTLE_ENDIAN__
41885 #define vcvt_n_f16_s16(__p0, __p1) __extension__ ({ \
41886  int16x4_t __s0 = __p0; \
41887  float16x4_t __ret; \
41888  __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_v((int8x8_t)__s0, __p1, 1); \
41889  __ret; \
41890 })
41891 #else
41892 #define vcvt_n_f16_s16(__p0, __p1) __extension__ ({ \
41893  int16x4_t __s0 = __p0; \
41894  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
41895  float16x4_t __ret; \
41896  __ret = (float16x4_t) __builtin_neon_vcvt_n_f16_v((int8x8_t)__rev0, __p1, 1); \
41897  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
41898  __ret; \
41899 })
41900 #endif
41901 
41902 #ifdef __LITTLE_ENDIAN__
41903 #define vcvtq_n_s16_f16(__p0, __p1) __extension__ ({ \
41904  float16x8_t __s0 = __p0; \
41905  int16x8_t __ret; \
41906  __ret = (int16x8_t) __builtin_neon_vcvtq_n_s16_v((int8x16_t)__s0, __p1, 33); \
41907  __ret; \
41908 })
41909 #else
41910 #define vcvtq_n_s16_f16(__p0, __p1) __extension__ ({ \
41911  float16x8_t __s0 = __p0; \
41912  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
41913  int16x8_t __ret; \
41914  __ret = (int16x8_t) __builtin_neon_vcvtq_n_s16_v((int8x16_t)__rev0, __p1, 33); \
41915  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
41916  __ret; \
41917 })
41918 #endif
41919 
41920 #ifdef __LITTLE_ENDIAN__
41921 #define vcvt_n_s16_f16(__p0, __p1) __extension__ ({ \
41922  float16x4_t __s0 = __p0; \
41923  int16x4_t __ret; \
41924  __ret = (int16x4_t) __builtin_neon_vcvt_n_s16_v((int8x8_t)__s0, __p1, 1); \
41925  __ret; \
41926 })
41927 #else
41928 #define vcvt_n_s16_f16(__p0, __p1) __extension__ ({ \
41929  float16x4_t __s0 = __p0; \
41930  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
41931  int16x4_t __ret; \
41932  __ret = (int16x4_t) __builtin_neon_vcvt_n_s16_v((int8x8_t)__rev0, __p1, 1); \
41933  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
41934  __ret; \
41935 })
41936 #endif
41937 
41938 #ifdef __LITTLE_ENDIAN__
41939 #define vcvtq_n_u16_f16(__p0, __p1) __extension__ ({ \
41940  float16x8_t __s0 = __p0; \
41941  uint16x8_t __ret; \
41942  __ret = (uint16x8_t) __builtin_neon_vcvtq_n_u16_v((int8x16_t)__s0, __p1, 49); \
41943  __ret; \
41944 })
41945 #else
41946 #define vcvtq_n_u16_f16(__p0, __p1) __extension__ ({ \
41947  float16x8_t __s0 = __p0; \
41948  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
41949  uint16x8_t __ret; \
41950  __ret = (uint16x8_t) __builtin_neon_vcvtq_n_u16_v((int8x16_t)__rev0, __p1, 49); \
41951  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
41952  __ret; \
41953 })
41954 #endif
41955 
41956 #ifdef __LITTLE_ENDIAN__
41957 #define vcvt_n_u16_f16(__p0, __p1) __extension__ ({ \
41958  float16x4_t __s0 = __p0; \
41959  uint16x4_t __ret; \
41960  __ret = (uint16x4_t) __builtin_neon_vcvt_n_u16_v((int8x8_t)__s0, __p1, 17); \
41961  __ret; \
41962 })
41963 #else
41964 #define vcvt_n_u16_f16(__p0, __p1) __extension__ ({ \
41965  float16x4_t __s0 = __p0; \
41966  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
41967  uint16x4_t __ret; \
41968  __ret = (uint16x4_t) __builtin_neon_vcvt_n_u16_v((int8x8_t)__rev0, __p1, 17); \
41969  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
41970  __ret; \
41971 })
41972 #endif
41973 
41974 #ifdef __LITTLE_ENDIAN__
41975 __ai int16x8_t vcvtq_s16_f16(float16x8_t __p0) {
41976  int16x8_t __ret;
41977  __ret = (int16x8_t) __builtin_neon_vcvtq_s16_v((int8x16_t)__p0, 33);
41978  return __ret;
41979 }
41980 #else
41981 __ai int16x8_t vcvtq_s16_f16(float16x8_t __p0) {
41982  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41983  int16x8_t __ret;
41984  __ret = (int16x8_t) __builtin_neon_vcvtq_s16_v((int8x16_t)__rev0, 33);
41985  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41986  return __ret;
41987 }
41988 #endif
41989 
41990 #ifdef __LITTLE_ENDIAN__
41991 __ai int16x4_t vcvt_s16_f16(float16x4_t __p0) {
41992  int16x4_t __ret;
41993  __ret = (int16x4_t) __builtin_neon_vcvt_s16_v((int8x8_t)__p0, 1);
41994  return __ret;
41995 }
41996 #else
41997 __ai int16x4_t vcvt_s16_f16(float16x4_t __p0) {
41998  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41999  int16x4_t __ret;
42000  __ret = (int16x4_t) __builtin_neon_vcvt_s16_v((int8x8_t)__rev0, 1);
42001  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42002  return __ret;
42003 }
42004 #endif
42005 
42006 #ifdef __LITTLE_ENDIAN__
42007 __ai uint16x8_t vcvtq_u16_f16(float16x8_t __p0) {
42008  uint16x8_t __ret;
42009  __ret = (uint16x8_t) __builtin_neon_vcvtq_u16_v((int8x16_t)__p0, 49);
42010  return __ret;
42011 }
42012 #else
42013 __ai uint16x8_t vcvtq_u16_f16(float16x8_t __p0) {
42014  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42015  uint16x8_t __ret;
42016  __ret = (uint16x8_t) __builtin_neon_vcvtq_u16_v((int8x16_t)__rev0, 49);
42017  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42018  return __ret;
42019 }
42020 #endif
42021 
42022 #ifdef __LITTLE_ENDIAN__
42023 __ai uint16x4_t vcvt_u16_f16(float16x4_t __p0) {
42024  uint16x4_t __ret;
42025  __ret = (uint16x4_t) __builtin_neon_vcvt_u16_v((int8x8_t)__p0, 17);
42026  return __ret;
42027 }
42028 #else
42029 __ai uint16x4_t vcvt_u16_f16(float16x4_t __p0) {
42030  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42031  uint16x4_t __ret;
42032  __ret = (uint16x4_t) __builtin_neon_vcvt_u16_v((int8x8_t)__rev0, 17);
42033  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42034  return __ret;
42035 }
42036 #endif
42037 
42038 #ifdef __LITTLE_ENDIAN__
42039 __ai int16x8_t vcvtaq_s16_f16(float16x8_t __p0) {
42040  int16x8_t __ret;
42041  __ret = (int16x8_t) __builtin_neon_vcvtaq_s16_v((int8x16_t)__p0, 33);
42042  return __ret;
42043 }
42044 #else
42045 __ai int16x8_t vcvtaq_s16_f16(float16x8_t __p0) {
42046  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42047  int16x8_t __ret;
42048  __ret = (int16x8_t) __builtin_neon_vcvtaq_s16_v((int8x16_t)__rev0, 33);
42049  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42050  return __ret;
42051 }
42052 #endif
42053 
42054 #ifdef __LITTLE_ENDIAN__
42055 __ai int16x4_t vcvta_s16_f16(float16x4_t __p0) {
42056  int16x4_t __ret;
42057  __ret = (int16x4_t) __builtin_neon_vcvta_s16_v((int8x8_t)__p0, 1);
42058  return __ret;
42059 }
42060 #else
42061 __ai int16x4_t vcvta_s16_f16(float16x4_t __p0) {
42062  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42063  int16x4_t __ret;
42064  __ret = (int16x4_t) __builtin_neon_vcvta_s16_v((int8x8_t)__rev0, 1);
42065  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42066  return __ret;
42067 }
42068 #endif
42069 
42070 #ifdef __LITTLE_ENDIAN__
42071 __ai uint16x8_t vcvtaq_u16_f16(float16x8_t __p0) {
42072  uint16x8_t __ret;
42073  __ret = (uint16x8_t) __builtin_neon_vcvtaq_u16_v((int8x16_t)__p0, 49);
42074  return __ret;
42075 }
42076 #else
42077 __ai uint16x8_t vcvtaq_u16_f16(float16x8_t __p0) {
42078  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42079  uint16x8_t __ret;
42080  __ret = (uint16x8_t) __builtin_neon_vcvtaq_u16_v((int8x16_t)__rev0, 49);
42081  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42082  return __ret;
42083 }
42084 #endif
42085 
42086 #ifdef __LITTLE_ENDIAN__
42087 __ai uint16x4_t vcvta_u16_f16(float16x4_t __p0) {
42088  uint16x4_t __ret;
42089  __ret = (uint16x4_t) __builtin_neon_vcvta_u16_v((int8x8_t)__p0, 17);
42090  return __ret;
42091 }
42092 #else
42093 __ai uint16x4_t vcvta_u16_f16(float16x4_t __p0) {
42094  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42095  uint16x4_t __ret;
42096  __ret = (uint16x4_t) __builtin_neon_vcvta_u16_v((int8x8_t)__rev0, 17);
42097  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42098  return __ret;
42099 }
42100 #endif
42101 
42102 #ifdef __LITTLE_ENDIAN__
42103 __ai int16x8_t vcvtmq_s16_f16(float16x8_t __p0) {
42104  int16x8_t __ret;
42105  __ret = (int16x8_t) __builtin_neon_vcvtmq_s16_v((int8x16_t)__p0, 33);
42106  return __ret;
42107 }
42108 #else
42109 __ai int16x8_t vcvtmq_s16_f16(float16x8_t __p0) {
42110  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42111  int16x8_t __ret;
42112  __ret = (int16x8_t) __builtin_neon_vcvtmq_s16_v((int8x16_t)__rev0, 33);
42113  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42114  return __ret;
42115 }
42116 #endif
42117 
42118 #ifdef __LITTLE_ENDIAN__
42119 __ai int16x4_t vcvtm_s16_f16(float16x4_t __p0) {
42120  int16x4_t __ret;
42121  __ret = (int16x4_t) __builtin_neon_vcvtm_s16_v((int8x8_t)__p0, 1);
42122  return __ret;
42123 }
42124 #else
42125 __ai int16x4_t vcvtm_s16_f16(float16x4_t __p0) {
42126  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42127  int16x4_t __ret;
42128  __ret = (int16x4_t) __builtin_neon_vcvtm_s16_v((int8x8_t)__rev0, 1);
42129  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42130  return __ret;
42131 }
42132 #endif
42133 
42134 #ifdef __LITTLE_ENDIAN__
42135 __ai uint16x8_t vcvtmq_u16_f16(float16x8_t __p0) {
42136  uint16x8_t __ret;
42137  __ret = (uint16x8_t) __builtin_neon_vcvtmq_u16_v((int8x16_t)__p0, 49);
42138  return __ret;
42139 }
42140 #else
42141 __ai uint16x8_t vcvtmq_u16_f16(float16x8_t __p0) {
42142  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42143  uint16x8_t __ret;
42144  __ret = (uint16x8_t) __builtin_neon_vcvtmq_u16_v((int8x16_t)__rev0, 49);
42145  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42146  return __ret;
42147 }
42148 #endif
42149 
42150 #ifdef __LITTLE_ENDIAN__
42151 __ai uint16x4_t vcvtm_u16_f16(float16x4_t __p0) {
42152  uint16x4_t __ret;
42153  __ret = (uint16x4_t) __builtin_neon_vcvtm_u16_v((int8x8_t)__p0, 17);
42154  return __ret;
42155 }
42156 #else
42157 __ai uint16x4_t vcvtm_u16_f16(float16x4_t __p0) {
42158  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42159  uint16x4_t __ret;
42160  __ret = (uint16x4_t) __builtin_neon_vcvtm_u16_v((int8x8_t)__rev0, 17);
42161  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42162  return __ret;
42163 }
42164 #endif
42165 
42166 #ifdef __LITTLE_ENDIAN__
42167 __ai int16x8_t vcvtnq_s16_f16(float16x8_t __p0) {
42168  int16x8_t __ret;
42169  __ret = (int16x8_t) __builtin_neon_vcvtnq_s16_v((int8x16_t)__p0, 33);
42170  return __ret;
42171 }
42172 #else
42173 __ai int16x8_t vcvtnq_s16_f16(float16x8_t __p0) {
42174  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42175  int16x8_t __ret;
42176  __ret = (int16x8_t) __builtin_neon_vcvtnq_s16_v((int8x16_t)__rev0, 33);
42177  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42178  return __ret;
42179 }
42180 #endif
42181 
42182 #ifdef __LITTLE_ENDIAN__
42183 __ai int16x4_t vcvtn_s16_f16(float16x4_t __p0) {
42184  int16x4_t __ret;
42185  __ret = (int16x4_t) __builtin_neon_vcvtn_s16_v((int8x8_t)__p0, 1);
42186  return __ret;
42187 }
42188 #else
42189 __ai int16x4_t vcvtn_s16_f16(float16x4_t __p0) {
42190  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42191  int16x4_t __ret;
42192  __ret = (int16x4_t) __builtin_neon_vcvtn_s16_v((int8x8_t)__rev0, 1);
42193  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42194  return __ret;
42195 }
42196 #endif
42197 
42198 #ifdef __LITTLE_ENDIAN__
42199 __ai uint16x8_t vcvtnq_u16_f16(float16x8_t __p0) {
42200  uint16x8_t __ret;
42201  __ret = (uint16x8_t) __builtin_neon_vcvtnq_u16_v((int8x16_t)__p0, 49);
42202  return __ret;
42203 }
42204 #else
42205 __ai uint16x8_t vcvtnq_u16_f16(float16x8_t __p0) {
42206  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42207  uint16x8_t __ret;
42208  __ret = (uint16x8_t) __builtin_neon_vcvtnq_u16_v((int8x16_t)__rev0, 49);
42209  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42210  return __ret;
42211 }
42212 #endif
42213 
42214 #ifdef __LITTLE_ENDIAN__
42215 __ai uint16x4_t vcvtn_u16_f16(float16x4_t __p0) {
42216  uint16x4_t __ret;
42217  __ret = (uint16x4_t) __builtin_neon_vcvtn_u16_v((int8x8_t)__p0, 17);
42218  return __ret;
42219 }
42220 #else
42221 __ai uint16x4_t vcvtn_u16_f16(float16x4_t __p0) {
42222  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42223  uint16x4_t __ret;
42224  __ret = (uint16x4_t) __builtin_neon_vcvtn_u16_v((int8x8_t)__rev0, 17);
42225  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42226  return __ret;
42227 }
42228 #endif
42229 
42230 #ifdef __LITTLE_ENDIAN__
42231 __ai int16x8_t vcvtpq_s16_f16(float16x8_t __p0) {
42232  int16x8_t __ret;
42233  __ret = (int16x8_t) __builtin_neon_vcvtpq_s16_v((int8x16_t)__p0, 33);
42234  return __ret;
42235 }
42236 #else
42237 __ai int16x8_t vcvtpq_s16_f16(float16x8_t __p0) {
42238  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42239  int16x8_t __ret;
42240  __ret = (int16x8_t) __builtin_neon_vcvtpq_s16_v((int8x16_t)__rev0, 33);
42241  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42242  return __ret;
42243 }
42244 #endif
42245 
42246 #ifdef __LITTLE_ENDIAN__
42247 __ai int16x4_t vcvtp_s16_f16(float16x4_t __p0) {
42248  int16x4_t __ret;
42249  __ret = (int16x4_t) __builtin_neon_vcvtp_s16_v((int8x8_t)__p0, 1);
42250  return __ret;
42251 }
42252 #else
42253 __ai int16x4_t vcvtp_s16_f16(float16x4_t __p0) {
42254  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42255  int16x4_t __ret;
42256  __ret = (int16x4_t) __builtin_neon_vcvtp_s16_v((int8x8_t)__rev0, 1);
42257  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42258  return __ret;
42259 }
42260 #endif
42261 
42262 #ifdef __LITTLE_ENDIAN__
42263 __ai uint16x8_t vcvtpq_u16_f16(float16x8_t __p0) {
42264  uint16x8_t __ret;
42265  __ret = (uint16x8_t) __builtin_neon_vcvtpq_u16_v((int8x16_t)__p0, 49);
42266  return __ret;
42267 }
42268 #else
42269 __ai uint16x8_t vcvtpq_u16_f16(float16x8_t __p0) {
42270  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42271  uint16x8_t __ret;
42272  __ret = (uint16x8_t) __builtin_neon_vcvtpq_u16_v((int8x16_t)__rev0, 49);
42273  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42274  return __ret;
42275 }
42276 #endif
42277 
42278 #ifdef __LITTLE_ENDIAN__
42279 __ai uint16x4_t vcvtp_u16_f16(float16x4_t __p0) {
42280  uint16x4_t __ret;
42281  __ret = (uint16x4_t) __builtin_neon_vcvtp_u16_v((int8x8_t)__p0, 17);
42282  return __ret;
42283 }
42284 #else
42285 __ai uint16x4_t vcvtp_u16_f16(float16x4_t __p0) {
42286  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42287  uint16x4_t __ret;
42288  __ret = (uint16x4_t) __builtin_neon_vcvtp_u16_v((int8x8_t)__rev0, 17);
42289  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42290  return __ret;
42291 }
42292 #endif
42293 
42294 #ifdef __LITTLE_ENDIAN__
42295 #define vextq_f16(__p0, __p1, __p2) __extension__ ({ \
42296  float16x8_t __s0 = __p0; \
42297  float16x8_t __s1 = __p1; \
42298  float16x8_t __ret; \
42299  __ret = (float16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 40); \
42300  __ret; \
42301 })
42302 #else
42303 #define vextq_f16(__p0, __p1, __p2) __extension__ ({ \
42304  float16x8_t __s0 = __p0; \
42305  float16x8_t __s1 = __p1; \
42306  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
42307  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
42308  float16x8_t __ret; \
42309  __ret = (float16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 40); \
42310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
42311  __ret; \
42312 })
42313 #endif
42314 
42315 #ifdef __LITTLE_ENDIAN__
42316 #define vext_f16(__p0, __p1, __p2) __extension__ ({ \
42317  float16x4_t __s0 = __p0; \
42318  float16x4_t __s1 = __p1; \
42319  float16x4_t __ret; \
42320  __ret = (float16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 8); \
42321  __ret; \
42322 })
42323 #else
42324 #define vext_f16(__p0, __p1, __p2) __extension__ ({ \
42325  float16x4_t __s0 = __p0; \
42326  float16x4_t __s1 = __p1; \
42327  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
42328  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
42329  float16x4_t __ret; \
42330  __ret = (float16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 8); \
42331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
42332  __ret; \
42333 })
42334 #endif
42335 
42336 #ifdef __LITTLE_ENDIAN__
42337 __ai float16x8_t vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
42338  float16x8_t __ret;
42339  __ret = (float16x8_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
42340  return __ret;
42341 }
42342 #else
42343 __ai float16x8_t vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
42344  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42345  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42346  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
42347  float16x8_t __ret;
42348  __ret = (float16x8_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 40);
42349  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42350  return __ret;
42351 }
42352 __ai float16x8_t __noswap_vfmaq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
42353  float16x8_t __ret;
42354  __ret = (float16x8_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 40);
42355  return __ret;
42356 }
42357 #endif
42358 
42359 #ifdef __LITTLE_ENDIAN__
42360 __ai float16x4_t vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
42361  float16x4_t __ret;
42362  __ret = (float16x4_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
42363  return __ret;
42364 }
42365 #else
42366 __ai float16x4_t vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
42367  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42368  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42369  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
42370  float16x4_t __ret;
42371  __ret = (float16x4_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 8);
42372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42373  return __ret;
42374 }
42375 __ai float16x4_t __noswap_vfma_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
42376  float16x4_t __ret;
42377  __ret = (float16x4_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 8);
42378  return __ret;
42379 }
42380 #endif
42381 
42382 #ifdef __LITTLE_ENDIAN__
42383 __ai float16x8_t vfmsq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
42384  float16x8_t __ret;
42385  __ret = vfmaq_f16(__p0, -__p1, __p2);
42386  return __ret;
42387 }
42388 #else
42389 __ai float16x8_t vfmsq_f16(float16x8_t __p0, float16x8_t __p1, float16x8_t __p2) {
42390  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42391  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42392  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
42393  float16x8_t __ret;
42394  __ret = __noswap_vfmaq_f16(__rev0, -__rev1, __rev2);
42395  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42396  return __ret;
42397 }
42398 #endif
42399 
42400 #ifdef __LITTLE_ENDIAN__
42401 __ai float16x4_t vfms_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
42402  float16x4_t __ret;
42403  __ret = vfma_f16(__p0, -__p1, __p2);
42404  return __ret;
42405 }
42406 #else
42407 __ai float16x4_t vfms_f16(float16x4_t __p0, float16x4_t __p1, float16x4_t __p2) {
42408  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42409  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42410  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
42411  float16x4_t __ret;
42412  __ret = __noswap_vfma_f16(__rev0, -__rev1, __rev2);
42413  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42414  return __ret;
42415 }
42416 #endif
42417 
42418 #ifdef __LITTLE_ENDIAN__
42419 __ai float16x8_t vmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
42420  float16x8_t __ret;
42421  __ret = (float16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
42422  return __ret;
42423 }
42424 #else
42425 __ai float16x8_t vmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
42426  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42427  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42428  float16x8_t __ret;
42429  __ret = (float16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42430  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42431  return __ret;
42432 }
42433 #endif
42434 
42435 #ifdef __LITTLE_ENDIAN__
42436 __ai float16x4_t vmax_f16(float16x4_t __p0, float16x4_t __p1) {
42437  float16x4_t __ret;
42438  __ret = (float16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42439  return __ret;
42440 }
42441 #else
42442 __ai float16x4_t vmax_f16(float16x4_t __p0, float16x4_t __p1) {
42443  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42444  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42445  float16x4_t __ret;
42446  __ret = (float16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42447  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42448  return __ret;
42449 }
42450 #endif
42451 
42452 #ifdef __LITTLE_ENDIAN__
42453 __ai float16x8_t vminq_f16(float16x8_t __p0, float16x8_t __p1) {
42454  float16x8_t __ret;
42455  __ret = (float16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
42456  return __ret;
42457 }
42458 #else
42459 __ai float16x8_t vminq_f16(float16x8_t __p0, float16x8_t __p1) {
42460  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42461  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42462  float16x8_t __ret;
42463  __ret = (float16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42464  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42465  return __ret;
42466 }
42467 #endif
42468 
42469 #ifdef __LITTLE_ENDIAN__
42470 __ai float16x4_t vmin_f16(float16x4_t __p0, float16x4_t __p1) {
42471  float16x4_t __ret;
42472  __ret = (float16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42473  return __ret;
42474 }
42475 #else
42476 __ai float16x4_t vmin_f16(float16x4_t __p0, float16x4_t __p1) {
42477  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42478  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42479  float16x4_t __ret;
42480  __ret = (float16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42481  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42482  return __ret;
42483 }
42484 #endif
42485 
42486 #ifdef __LITTLE_ENDIAN__
42487 __ai float16x8_t vmulq_f16(float16x8_t __p0, float16x8_t __p1) {
42488  float16x8_t __ret;
42489  __ret = __p0 * __p1;
42490  return __ret;
42491 }
42492 #else
42493 __ai float16x8_t vmulq_f16(float16x8_t __p0, float16x8_t __p1) {
42494  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42495  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42496  float16x8_t __ret;
42497  __ret = __rev0 * __rev1;
42498  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42499  return __ret;
42500 }
42501 #endif
42502 
42503 #ifdef __LITTLE_ENDIAN__
42504 __ai float16x4_t vmul_f16(float16x4_t __p0, float16x4_t __p1) {
42505  float16x4_t __ret;
42506  __ret = __p0 * __p1;
42507  return __ret;
42508 }
42509 #else
42510 __ai float16x4_t vmul_f16(float16x4_t __p0, float16x4_t __p1) {
42511  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42512  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42513  float16x4_t __ret;
42514  __ret = __rev0 * __rev1;
42515  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42516  return __ret;
42517 }
42518 #endif
42519 
42520 #ifdef __LITTLE_ENDIAN__
42521 #define vmulq_lane_f16(__p0_185, __p1_185, __p2_185) __extension__ ({ \
42522  float16x8_t __s0_185 = __p0_185; \
42523  float16x4_t __s1_185 = __p1_185; \
42524  float16x8_t __ret_185; \
42525  __ret_185 = __s0_185 * splatq_lane_f16(__s1_185, __p2_185); \
42526  __ret_185; \
42527 })
42528 #else
42529 #define vmulq_lane_f16(__p0_186, __p1_186, __p2_186) __extension__ ({ \
42530  float16x8_t __s0_186 = __p0_186; \
42531  float16x4_t __s1_186 = __p1_186; \
42532  float16x8_t __rev0_186; __rev0_186 = __builtin_shufflevector(__s0_186, __s0_186, 7, 6, 5, 4, 3, 2, 1, 0); \
42533  float16x4_t __rev1_186; __rev1_186 = __builtin_shufflevector(__s1_186, __s1_186, 3, 2, 1, 0); \
42534  float16x8_t __ret_186; \
42535  __ret_186 = __rev0_186 * __noswap_splatq_lane_f16(__rev1_186, __p2_186); \
42536  __ret_186 = __builtin_shufflevector(__ret_186, __ret_186, 7, 6, 5, 4, 3, 2, 1, 0); \
42537  __ret_186; \
42538 })
42539 #endif
42540 
42541 #ifdef __LITTLE_ENDIAN__
42542 #define vmul_lane_f16(__p0_187, __p1_187, __p2_187) __extension__ ({ \
42543  float16x4_t __s0_187 = __p0_187; \
42544  float16x4_t __s1_187 = __p1_187; \
42545  float16x4_t __ret_187; \
42546  __ret_187 = __s0_187 * splat_lane_f16(__s1_187, __p2_187); \
42547  __ret_187; \
42548 })
42549 #else
42550 #define vmul_lane_f16(__p0_188, __p1_188, __p2_188) __extension__ ({ \
42551  float16x4_t __s0_188 = __p0_188; \
42552  float16x4_t __s1_188 = __p1_188; \
42553  float16x4_t __rev0_188; __rev0_188 = __builtin_shufflevector(__s0_188, __s0_188, 3, 2, 1, 0); \
42554  float16x4_t __rev1_188; __rev1_188 = __builtin_shufflevector(__s1_188, __s1_188, 3, 2, 1, 0); \
42555  float16x4_t __ret_188; \
42556  __ret_188 = __rev0_188 * __noswap_splat_lane_f16(__rev1_188, __p2_188); \
42557  __ret_188 = __builtin_shufflevector(__ret_188, __ret_188, 3, 2, 1, 0); \
42558  __ret_188; \
42559 })
42560 #endif
42561 
42562 #ifdef __LITTLE_ENDIAN__
42563 #define vmulq_n_f16(__p0, __p1) __extension__ ({ \
42564  float16x8_t __s0 = __p0; \
42565  float16_t __s1 = __p1; \
42566  float16x8_t __ret; \
42567  __ret = __s0 * (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}; \
42568  __ret; \
42569 })
42570 #else
42571 #define vmulq_n_f16(__p0, __p1) __extension__ ({ \
42572  float16x8_t __s0 = __p0; \
42573  float16_t __s1 = __p1; \
42574  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
42575  float16x8_t __ret; \
42576  __ret = __rev0 * (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}; \
42577  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
42578  __ret; \
42579 })
42580 #endif
42581 
42582 #ifdef __LITTLE_ENDIAN__
42583 #define vmul_n_f16(__p0, __p1) __extension__ ({ \
42584  float16x4_t __s0 = __p0; \
42585  float16_t __s1 = __p1; \
42586  float16x4_t __ret; \
42587  __ret = __s0 * (float16x4_t) {__s1, __s1, __s1, __s1}; \
42588  __ret; \
42589 })
42590 #else
42591 #define vmul_n_f16(__p0, __p1) __extension__ ({ \
42592  float16x4_t __s0 = __p0; \
42593  float16_t __s1 = __p1; \
42594  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
42595  float16x4_t __ret; \
42596  __ret = __rev0 * (float16x4_t) {__s1, __s1, __s1, __s1}; \
42597  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
42598  __ret; \
42599 })
42600 #endif
42601 
42602 #ifdef __LITTLE_ENDIAN__
42603 __ai float16x8_t vnegq_f16(float16x8_t __p0) {
42604  float16x8_t __ret;
42605  __ret = -__p0;
42606  return __ret;
42607 }
42608 #else
42609 __ai float16x8_t vnegq_f16(float16x8_t __p0) {
42610  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42611  float16x8_t __ret;
42612  __ret = -__rev0;
42613  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42614  return __ret;
42615 }
42616 #endif
42617 
42618 #ifdef __LITTLE_ENDIAN__
42619 __ai float16x4_t vneg_f16(float16x4_t __p0) {
42620  float16x4_t __ret;
42621  __ret = -__p0;
42622  return __ret;
42623 }
42624 #else
42625 __ai float16x4_t vneg_f16(float16x4_t __p0) {
42626  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42627  float16x4_t __ret;
42628  __ret = -__rev0;
42629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42630  return __ret;
42631 }
42632 #endif
42633 
42634 #ifdef __LITTLE_ENDIAN__
42635 __ai float16x4_t vpadd_f16(float16x4_t __p0, float16x4_t __p1) {
42636  float16x4_t __ret;
42637  __ret = (float16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42638  return __ret;
42639 }
42640 #else
42641 __ai float16x4_t vpadd_f16(float16x4_t __p0, float16x4_t __p1) {
42642  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42643  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42644  float16x4_t __ret;
42645  __ret = (float16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42646  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42647  return __ret;
42648 }
42649 #endif
42650 
42651 #ifdef __LITTLE_ENDIAN__
42652 __ai float16x4_t vpmax_f16(float16x4_t __p0, float16x4_t __p1) {
42653  float16x4_t __ret;
42654  __ret = (float16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42655  return __ret;
42656 }
42657 #else
42658 __ai float16x4_t vpmax_f16(float16x4_t __p0, float16x4_t __p1) {
42659  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42660  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42661  float16x4_t __ret;
42662  __ret = (float16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42663  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42664  return __ret;
42665 }
42666 #endif
42667 
42668 #ifdef __LITTLE_ENDIAN__
42669 __ai float16x4_t vpmin_f16(float16x4_t __p0, float16x4_t __p1) {
42670  float16x4_t __ret;
42671  __ret = (float16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42672  return __ret;
42673 }
42674 #else
42675 __ai float16x4_t vpmin_f16(float16x4_t __p0, float16x4_t __p1) {
42676  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42677  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42678  float16x4_t __ret;
42679  __ret = (float16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42681  return __ret;
42682 }
42683 #endif
42684 
42685 #ifdef __LITTLE_ENDIAN__
42686 __ai float16x8_t vrecpeq_f16(float16x8_t __p0) {
42687  float16x8_t __ret;
42688  __ret = (float16x8_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 40);
42689  return __ret;
42690 }
42691 #else
42692 __ai float16x8_t vrecpeq_f16(float16x8_t __p0) {
42693  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42694  float16x8_t __ret;
42695  __ret = (float16x8_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 40);
42696  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42697  return __ret;
42698 }
42699 #endif
42700 
42701 #ifdef __LITTLE_ENDIAN__
42702 __ai float16x4_t vrecpe_f16(float16x4_t __p0) {
42703  float16x4_t __ret;
42704  __ret = (float16x4_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 8);
42705  return __ret;
42706 }
42707 #else
42708 __ai float16x4_t vrecpe_f16(float16x4_t __p0) {
42709  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42710  float16x4_t __ret;
42711  __ret = (float16x4_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 8);
42712  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42713  return __ret;
42714 }
42715 #endif
42716 
42717 #ifdef __LITTLE_ENDIAN__
42718 __ai float16x8_t vrecpsq_f16(float16x8_t __p0, float16x8_t __p1) {
42719  float16x8_t __ret;
42720  __ret = (float16x8_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
42721  return __ret;
42722 }
42723 #else
42724 __ai float16x8_t vrecpsq_f16(float16x8_t __p0, float16x8_t __p1) {
42725  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42726  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42727  float16x8_t __ret;
42728  __ret = (float16x8_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42729  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42730  return __ret;
42731 }
42732 #endif
42733 
42734 #ifdef __LITTLE_ENDIAN__
42735 __ai float16x4_t vrecps_f16(float16x4_t __p0, float16x4_t __p1) {
42736  float16x4_t __ret;
42737  __ret = (float16x4_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42738  return __ret;
42739 }
42740 #else
42741 __ai float16x4_t vrecps_f16(float16x4_t __p0, float16x4_t __p1) {
42742  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42743  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42744  float16x4_t __ret;
42745  __ret = (float16x4_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42746  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42747  return __ret;
42748 }
42749 #endif
42750 
42751 #ifdef __LITTLE_ENDIAN__
42752 __ai float16x8_t vrev64q_f16(float16x8_t __p0) {
42753  float16x8_t __ret;
42754  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
42755  return __ret;
42756 }
42757 #else
42758 __ai float16x8_t vrev64q_f16(float16x8_t __p0) {
42759  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42760  float16x8_t __ret;
42761  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
42762  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42763  return __ret;
42764 }
42765 #endif
42766 
42767 #ifdef __LITTLE_ENDIAN__
42768 __ai float16x4_t vrev64_f16(float16x4_t __p0) {
42769  float16x4_t __ret;
42770  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42771  return __ret;
42772 }
42773 #else
42774 __ai float16x4_t vrev64_f16(float16x4_t __p0) {
42775  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42776  float16x4_t __ret;
42777  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
42778  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42779  return __ret;
42780 }
42781 #endif
42782 
42783 #ifdef __LITTLE_ENDIAN__
42784 __ai float16x8_t vrsqrteq_f16(float16x8_t __p0) {
42785  float16x8_t __ret;
42786  __ret = (float16x8_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 40);
42787  return __ret;
42788 }
42789 #else
42790 __ai float16x8_t vrsqrteq_f16(float16x8_t __p0) {
42791  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42792  float16x8_t __ret;
42793  __ret = (float16x8_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 40);
42794  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42795  return __ret;
42796 }
42797 #endif
42798 
42799 #ifdef __LITTLE_ENDIAN__
42800 __ai float16x4_t vrsqrte_f16(float16x4_t __p0) {
42801  float16x4_t __ret;
42802  __ret = (float16x4_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 8);
42803  return __ret;
42804 }
42805 #else
42806 __ai float16x4_t vrsqrte_f16(float16x4_t __p0) {
42807  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42808  float16x4_t __ret;
42809  __ret = (float16x4_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 8);
42810  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42811  return __ret;
42812 }
42813 #endif
42814 
42815 #ifdef __LITTLE_ENDIAN__
42816 __ai float16x8_t vrsqrtsq_f16(float16x8_t __p0, float16x8_t __p1) {
42817  float16x8_t __ret;
42818  __ret = (float16x8_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
42819  return __ret;
42820 }
42821 #else
42822 __ai float16x8_t vrsqrtsq_f16(float16x8_t __p0, float16x8_t __p1) {
42823  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42824  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42825  float16x8_t __ret;
42826  __ret = (float16x8_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42827  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42828  return __ret;
42829 }
42830 #endif
42831 
42832 #ifdef __LITTLE_ENDIAN__
42833 __ai float16x4_t vrsqrts_f16(float16x4_t __p0, float16x4_t __p1) {
42834  float16x4_t __ret;
42835  __ret = (float16x4_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
42836  return __ret;
42837 }
42838 #else
42839 __ai float16x4_t vrsqrts_f16(float16x4_t __p0, float16x4_t __p1) {
42840  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42841  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42842  float16x4_t __ret;
42843  __ret = (float16x4_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42845  return __ret;
42846 }
42847 #endif
42848 
42849 #ifdef __LITTLE_ENDIAN__
42850 __ai float16x8_t vsubq_f16(float16x8_t __p0, float16x8_t __p1) {
42851  float16x8_t __ret;
42852  __ret = __p0 - __p1;
42853  return __ret;
42854 }
42855 #else
42856 __ai float16x8_t vsubq_f16(float16x8_t __p0, float16x8_t __p1) {
42857  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42858  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42859  float16x8_t __ret;
42860  __ret = __rev0 - __rev1;
42861  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42862  return __ret;
42863 }
42864 #endif
42865 
42866 #ifdef __LITTLE_ENDIAN__
42867 __ai float16x4_t vsub_f16(float16x4_t __p0, float16x4_t __p1) {
42868  float16x4_t __ret;
42869  __ret = __p0 - __p1;
42870  return __ret;
42871 }
42872 #else
42873 __ai float16x4_t vsub_f16(float16x4_t __p0, float16x4_t __p1) {
42874  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42875  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42876  float16x4_t __ret;
42877  __ret = __rev0 - __rev1;
42878  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42879  return __ret;
42880 }
42881 #endif
42882 
42883 #ifdef __LITTLE_ENDIAN__
42884 __ai float16x8x2_t vtrnq_f16(float16x8_t __p0, float16x8_t __p1) {
42885  float16x8x2_t __ret;
42886  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
42887  return __ret;
42888 }
42889 #else
42890 __ai float16x8x2_t vtrnq_f16(float16x8_t __p0, float16x8_t __p1) {
42891  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42892  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42893  float16x8x2_t __ret;
42894  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42895 
42896  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
42897  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
42898  return __ret;
42899 }
42900 #endif
42901 
42902 #ifdef __LITTLE_ENDIAN__
42903 __ai float16x4x2_t vtrn_f16(float16x4_t __p0, float16x4_t __p1) {
42904  float16x4x2_t __ret;
42905  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
42906  return __ret;
42907 }
42908 #else
42909 __ai float16x4x2_t vtrn_f16(float16x4_t __p0, float16x4_t __p1) {
42910  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42911  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42912  float16x4x2_t __ret;
42913  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42914 
42915  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
42916  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
42917  return __ret;
42918 }
42919 #endif
42920 
42921 #ifdef __LITTLE_ENDIAN__
42922 __ai float16x8x2_t vuzpq_f16(float16x8_t __p0, float16x8_t __p1) {
42923  float16x8x2_t __ret;
42924  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
42925  return __ret;
42926 }
42927 #else
42928 __ai float16x8x2_t vuzpq_f16(float16x8_t __p0, float16x8_t __p1) {
42929  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42930  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42931  float16x8x2_t __ret;
42932  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42933 
42934  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
42935  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
42936  return __ret;
42937 }
42938 #endif
42939 
42940 #ifdef __LITTLE_ENDIAN__
42941 __ai float16x4x2_t vuzp_f16(float16x4_t __p0, float16x4_t __p1) {
42942  float16x4x2_t __ret;
42943  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
42944  return __ret;
42945 }
42946 #else
42947 __ai float16x4x2_t vuzp_f16(float16x4_t __p0, float16x4_t __p1) {
42948  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42949  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42950  float16x4x2_t __ret;
42951  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42952 
42953  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
42954  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
42955  return __ret;
42956 }
42957 #endif
42958 
42959 #ifdef __LITTLE_ENDIAN__
42960 __ai float16x8x2_t vzipq_f16(float16x8_t __p0, float16x8_t __p1) {
42961  float16x8x2_t __ret;
42962  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 40);
42963  return __ret;
42964 }
42965 #else
42966 __ai float16x8x2_t vzipq_f16(float16x8_t __p0, float16x8_t __p1) {
42967  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42968  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
42969  float16x8x2_t __ret;
42970  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 40);
42971 
42972  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
42973  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
42974  return __ret;
42975 }
42976 #endif
42977 
42978 #ifdef __LITTLE_ENDIAN__
42979 __ai float16x4x2_t vzip_f16(float16x4_t __p0, float16x4_t __p1) {
42980  float16x4x2_t __ret;
42981  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 8);
42982  return __ret;
42983 }
42984 #else
42985 __ai float16x4x2_t vzip_f16(float16x4_t __p0, float16x4_t __p1) {
42986  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42987  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
42988  float16x4x2_t __ret;
42989  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 8);
42990 
42991  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
42992  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
42993  return __ret;
42994 }
42995 #endif
42996 
42997 #endif
42998 #if defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC) && defined(__aarch64__)
42999 #ifdef __LITTLE_ENDIAN__
43000 __ai float16x8_t vdivq_f16(float16x8_t __p0, float16x8_t __p1) {
43001  float16x8_t __ret;
43002  __ret = __p0 / __p1;
43003  return __ret;
43004 }
43005 #else
43006 __ai float16x8_t vdivq_f16(float16x8_t __p0, float16x8_t __p1) {
43007  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43008  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43009  float16x8_t __ret;
43010  __ret = __rev0 / __rev1;
43011  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43012  return __ret;
43013 }
43014 #endif
43015 
43016 #ifdef __LITTLE_ENDIAN__
43017 __ai float16x4_t vdiv_f16(float16x4_t __p0, float16x4_t __p1) {
43018  float16x4_t __ret;
43019  __ret = __p0 / __p1;
43020  return __ret;
43021 }
43022 #else
43023 __ai float16x4_t vdiv_f16(float16x4_t __p0, float16x4_t __p1) {
43024  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43025  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
43026  float16x4_t __ret;
43027  __ret = __rev0 / __rev1;
43028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43029  return __ret;
43030 }
43031 #endif
43032 
43033 #ifdef __LITTLE_ENDIAN__
43034 #define vduph_lane_f16(__p0, __p1) __extension__ ({ \
43035  float16x4_t __s0 = __p0; \
43036  float16_t __ret; \
43037  __ret = (float16_t) __builtin_neon_vduph_lane_f16((float16x4_t)__s0, __p1); \
43038  __ret; \
43039 })
43040 #else
43041 #define vduph_lane_f16(__p0, __p1) __extension__ ({ \
43042  float16x4_t __s0 = __p0; \
43043  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43044  float16_t __ret; \
43045  __ret = (float16_t) __builtin_neon_vduph_lane_f16((float16x4_t)__rev0, __p1); \
43046  __ret; \
43047 })
43048 #endif
43049 
43050 #ifdef __LITTLE_ENDIAN__
43051 #define vduph_laneq_f16(__p0, __p1) __extension__ ({ \
43052  float16x8_t __s0 = __p0; \
43053  float16_t __ret; \
43054  __ret = (float16_t) __builtin_neon_vduph_laneq_f16((float16x8_t)__s0, __p1); \
43055  __ret; \
43056 })
43057 #else
43058 #define vduph_laneq_f16(__p0, __p1) __extension__ ({ \
43059  float16x8_t __s0 = __p0; \
43060  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43061  float16_t __ret; \
43062  __ret = (float16_t) __builtin_neon_vduph_laneq_f16((float16x8_t)__rev0, __p1); \
43063  __ret; \
43064 })
43065 #endif
43066 
43067 #ifdef __LITTLE_ENDIAN__
43068 #define vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43069  float16_t __s0 = __p0; \
43070  float16_t __s1 = __p1; \
43071  float16x4_t __s2 = __p2; \
43072  float16_t __ret; \
43073  __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__s2, __p3); \
43074  __ret; \
43075 })
43076 #else
43077 #define vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43078  float16_t __s0 = __p0; \
43079  float16_t __s1 = __p1; \
43080  float16x4_t __s2 = __p2; \
43081  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
43082  float16_t __ret; \
43083  __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__rev2, __p3); \
43084  __ret; \
43085 })
43086 #define __noswap_vfmah_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43087  float16_t __s0 = __p0; \
43088  float16_t __s1 = __p1; \
43089  float16x4_t __s2 = __p2; \
43090  float16_t __ret; \
43091  __ret = (float16_t) __builtin_neon_vfmah_lane_f16(__s0, __s1, (float16x4_t)__s2, __p3); \
43092  __ret; \
43093 })
43094 #endif
43095 
43096 #ifdef __LITTLE_ENDIAN__
43097 #define vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43098  float16x8_t __s0 = __p0; \
43099  float16x8_t __s1 = __p1; \
43100  float16x4_t __s2 = __p2; \
43101  float16x8_t __ret; \
43102  __ret = (float16x8_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 40); \
43103  __ret; \
43104 })
43105 #else
43106 #define vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43107  float16x8_t __s0 = __p0; \
43108  float16x8_t __s1 = __p1; \
43109  float16x4_t __s2 = __p2; \
43110  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43111  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
43112  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
43113  float16x8_t __ret; \
43114  __ret = (float16x8_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 40); \
43115  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
43116  __ret; \
43117 })
43118 #define __noswap_vfmaq_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43119  float16x8_t __s0 = __p0; \
43120  float16x8_t __s1 = __p1; \
43121  float16x4_t __s2 = __p2; \
43122  float16x8_t __ret; \
43123  __ret = (float16x8_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 40); \
43124  __ret; \
43125 })
43126 #endif
43127 
43128 #ifdef __LITTLE_ENDIAN__
43129 #define vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43130  float16x4_t __s0 = __p0; \
43131  float16x4_t __s1 = __p1; \
43132  float16x4_t __s2 = __p2; \
43133  float16x4_t __ret; \
43134  __ret = (float16x4_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 8); \
43135  __ret; \
43136 })
43137 #else
43138 #define vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43139  float16x4_t __s0 = __p0; \
43140  float16x4_t __s1 = __p1; \
43141  float16x4_t __s2 = __p2; \
43142  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43143  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
43144  float16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
43145  float16x4_t __ret; \
43146  __ret = (float16x4_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 8); \
43147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
43148  __ret; \
43149 })
43150 #define __noswap_vfma_lane_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43151  float16x4_t __s0 = __p0; \
43152  float16x4_t __s1 = __p1; \
43153  float16x4_t __s2 = __p2; \
43154  float16x4_t __ret; \
43155  __ret = (float16x4_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 8); \
43156  __ret; \
43157 })
43158 #endif
43159 
43160 #ifdef __LITTLE_ENDIAN__
43161 #define vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43162  float16_t __s0 = __p0; \
43163  float16_t __s1 = __p1; \
43164  float16x8_t __s2 = __p2; \
43165  float16_t __ret; \
43166  __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__s2, __p3); \
43167  __ret; \
43168 })
43169 #else
43170 #define vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43171  float16_t __s0 = __p0; \
43172  float16_t __s1 = __p1; \
43173  float16x8_t __s2 = __p2; \
43174  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
43175  float16_t __ret; \
43176  __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__rev2, __p3); \
43177  __ret; \
43178 })
43179 #define __noswap_vfmah_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43180  float16_t __s0 = __p0; \
43181  float16_t __s1 = __p1; \
43182  float16x8_t __s2 = __p2; \
43183  float16_t __ret; \
43184  __ret = (float16_t) __builtin_neon_vfmah_laneq_f16(__s0, __s1, (float16x8_t)__s2, __p3); \
43185  __ret; \
43186 })
43187 #endif
43188 
43189 #ifdef __LITTLE_ENDIAN__
43190 #define vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43191  float16x8_t __s0 = __p0; \
43192  float16x8_t __s1 = __p1; \
43193  float16x8_t __s2 = __p2; \
43194  float16x8_t __ret; \
43195  __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 40); \
43196  __ret; \
43197 })
43198 #else
43199 #define vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43200  float16x8_t __s0 = __p0; \
43201  float16x8_t __s1 = __p1; \
43202  float16x8_t __s2 = __p2; \
43203  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43204  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
43205  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
43206  float16x8_t __ret; \
43207  __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 40); \
43208  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
43209  __ret; \
43210 })
43211 #define __noswap_vfmaq_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43212  float16x8_t __s0 = __p0; \
43213  float16x8_t __s1 = __p1; \
43214  float16x8_t __s2 = __p2; \
43215  float16x8_t __ret; \
43216  __ret = (float16x8_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 40); \
43217  __ret; \
43218 })
43219 #endif
43220 
43221 #ifdef __LITTLE_ENDIAN__
43222 #define vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43223  float16x4_t __s0 = __p0; \
43224  float16x4_t __s1 = __p1; \
43225  float16x8_t __s2 = __p2; \
43226  float16x4_t __ret; \
43227  __ret = (float16x4_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 8); \
43228  __ret; \
43229 })
43230 #else
43231 #define vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43232  float16x4_t __s0 = __p0; \
43233  float16x4_t __s1 = __p1; \
43234  float16x8_t __s2 = __p2; \
43235  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43236  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
43237  float16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
43238  float16x4_t __ret; \
43239  __ret = (float16x4_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 8); \
43240  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
43241  __ret; \
43242 })
43243 #define __noswap_vfma_laneq_f16(__p0, __p1, __p2, __p3) __extension__ ({ \
43244  float16x4_t __s0 = __p0; \
43245  float16x4_t __s1 = __p1; \
43246  float16x8_t __s2 = __p2; \
43247  float16x4_t __ret; \
43248  __ret = (float16x4_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 8); \
43249  __ret; \
43250 })
43251 #endif
43252 
43253 #ifdef __LITTLE_ENDIAN__
43254 #define vfmaq_n_f16(__p0, __p1, __p2) __extension__ ({ \
43255  float16x8_t __s0 = __p0; \
43256  float16x8_t __s1 = __p1; \
43257  float16_t __s2 = __p2; \
43258  float16x8_t __ret; \
43259  __ret = vfmaq_f16(__s0, __s1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
43260  __ret; \
43261 })
43262 #else
43263 #define vfmaq_n_f16(__p0, __p1, __p2) __extension__ ({ \
43264  float16x8_t __s0 = __p0; \
43265  float16x8_t __s1 = __p1; \
43266  float16_t __s2 = __p2; \
43267  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43268  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
43269  float16x8_t __ret; \
43270  __ret = __noswap_vfmaq_f16(__rev0, __rev1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
43271  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
43272  __ret; \
43273 })
43274 #endif
43275 
43276 #ifdef __LITTLE_ENDIAN__
43277 #define vfma_n_f16(__p0, __p1, __p2) __extension__ ({ \
43278  float16x4_t __s0 = __p0; \
43279  float16x4_t __s1 = __p1; \
43280  float16_t __s2 = __p2; \
43281  float16x4_t __ret; \
43282  __ret = vfma_f16(__s0, __s1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
43283  __ret; \
43284 })
43285 #else
43286 #define vfma_n_f16(__p0, __p1, __p2) __extension__ ({ \
43287  float16x4_t __s0 = __p0; \
43288  float16x4_t __s1 = __p1; \
43289  float16_t __s2 = __p2; \
43290  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43291  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
43292  float16x4_t __ret; \
43293  __ret = __noswap_vfma_f16(__rev0, __rev1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
43294  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
43295  __ret; \
43296 })
43297 #endif
43298 
43299 #ifdef __LITTLE_ENDIAN__
43300 #define vfmsh_lane_f16(__p0_189, __p1_189, __p2_189, __p3_189) __extension__ ({ \
43301  float16_t __s0_189 = __p0_189; \
43302  float16_t __s1_189 = __p1_189; \
43303  float16x4_t __s2_189 = __p2_189; \
43304  float16_t __ret_189; \
43305  __ret_189 = vfmah_lane_f16(__s0_189, -__s1_189, __s2_189, __p3_189); \
43306  __ret_189; \
43307 })
43308 #else
43309 #define vfmsh_lane_f16(__p0_190, __p1_190, __p2_190, __p3_190) __extension__ ({ \
43310  float16_t __s0_190 = __p0_190; \
43311  float16_t __s1_190 = __p1_190; \
43312  float16x4_t __s2_190 = __p2_190; \
43313  float16x4_t __rev2_190; __rev2_190 = __builtin_shufflevector(__s2_190, __s2_190, 3, 2, 1, 0); \
43314  float16_t __ret_190; \
43315  __ret_190 = __noswap_vfmah_lane_f16(__s0_190, -__s1_190, __rev2_190, __p3_190); \
43316  __ret_190; \
43317 })
43318 #endif
43319 
43320 #ifdef __LITTLE_ENDIAN__
43321 #define vfmsq_lane_f16(__p0_191, __p1_191, __p2_191, __p3_191) __extension__ ({ \
43322  float16x8_t __s0_191 = __p0_191; \
43323  float16x8_t __s1_191 = __p1_191; \
43324  float16x4_t __s2_191 = __p2_191; \
43325  float16x8_t __ret_191; \
43326  __ret_191 = vfmaq_lane_f16(__s0_191, -__s1_191, __s2_191, __p3_191); \
43327  __ret_191; \
43328 })
43329 #else
43330 #define vfmsq_lane_f16(__p0_192, __p1_192, __p2_192, __p3_192) __extension__ ({ \
43331  float16x8_t __s0_192 = __p0_192; \
43332  float16x8_t __s1_192 = __p1_192; \
43333  float16x4_t __s2_192 = __p2_192; \
43334  float16x8_t __rev0_192; __rev0_192 = __builtin_shufflevector(__s0_192, __s0_192, 7, 6, 5, 4, 3, 2, 1, 0); \
43335  float16x8_t __rev1_192; __rev1_192 = __builtin_shufflevector(__s1_192, __s1_192, 7, 6, 5, 4, 3, 2, 1, 0); \
43336  float16x4_t __rev2_192; __rev2_192 = __builtin_shufflevector(__s2_192, __s2_192, 3, 2, 1, 0); \
43337  float16x8_t __ret_192; \
43338  __ret_192 = __noswap_vfmaq_lane_f16(__rev0_192, -__rev1_192, __rev2_192, __p3_192); \
43339  __ret_192 = __builtin_shufflevector(__ret_192, __ret_192, 7, 6, 5, 4, 3, 2, 1, 0); \
43340  __ret_192; \
43341 })
43342 #endif
43343 
43344 #ifdef __LITTLE_ENDIAN__
43345 #define vfms_lane_f16(__p0_193, __p1_193, __p2_193, __p3_193) __extension__ ({ \
43346  float16x4_t __s0_193 = __p0_193; \
43347  float16x4_t __s1_193 = __p1_193; \
43348  float16x4_t __s2_193 = __p2_193; \
43349  float16x4_t __ret_193; \
43350  __ret_193 = vfma_lane_f16(__s0_193, -__s1_193, __s2_193, __p3_193); \
43351  __ret_193; \
43352 })
43353 #else
43354 #define vfms_lane_f16(__p0_194, __p1_194, __p2_194, __p3_194) __extension__ ({ \
43355  float16x4_t __s0_194 = __p0_194; \
43356  float16x4_t __s1_194 = __p1_194; \
43357  float16x4_t __s2_194 = __p2_194; \
43358  float16x4_t __rev0_194; __rev0_194 = __builtin_shufflevector(__s0_194, __s0_194, 3, 2, 1, 0); \
43359  float16x4_t __rev1_194; __rev1_194 = __builtin_shufflevector(__s1_194, __s1_194, 3, 2, 1, 0); \
43360  float16x4_t __rev2_194; __rev2_194 = __builtin_shufflevector(__s2_194, __s2_194, 3, 2, 1, 0); \
43361  float16x4_t __ret_194; \
43362  __ret_194 = __noswap_vfma_lane_f16(__rev0_194, -__rev1_194, __rev2_194, __p3_194); \
43363  __ret_194 = __builtin_shufflevector(__ret_194, __ret_194, 3, 2, 1, 0); \
43364  __ret_194; \
43365 })
43366 #endif
43367 
43368 #ifdef __LITTLE_ENDIAN__
43369 #define vfmsh_laneq_f16(__p0_195, __p1_195, __p2_195, __p3_195) __extension__ ({ \
43370  float16_t __s0_195 = __p0_195; \
43371  float16_t __s1_195 = __p1_195; \
43372  float16x8_t __s2_195 = __p2_195; \
43373  float16_t __ret_195; \
43374  __ret_195 = vfmah_laneq_f16(__s0_195, -__s1_195, __s2_195, __p3_195); \
43375  __ret_195; \
43376 })
43377 #else
43378 #define vfmsh_laneq_f16(__p0_196, __p1_196, __p2_196, __p3_196) __extension__ ({ \
43379  float16_t __s0_196 = __p0_196; \
43380  float16_t __s1_196 = __p1_196; \
43381  float16x8_t __s2_196 = __p2_196; \
43382  float16x8_t __rev2_196; __rev2_196 = __builtin_shufflevector(__s2_196, __s2_196, 7, 6, 5, 4, 3, 2, 1, 0); \
43383  float16_t __ret_196; \
43384  __ret_196 = __noswap_vfmah_laneq_f16(__s0_196, -__s1_196, __rev2_196, __p3_196); \
43385  __ret_196; \
43386 })
43387 #endif
43388 
43389 #ifdef __LITTLE_ENDIAN__
43390 #define vfmsq_laneq_f16(__p0_197, __p1_197, __p2_197, __p3_197) __extension__ ({ \
43391  float16x8_t __s0_197 = __p0_197; \
43392  float16x8_t __s1_197 = __p1_197; \
43393  float16x8_t __s2_197 = __p2_197; \
43394  float16x8_t __ret_197; \
43395  __ret_197 = vfmaq_laneq_f16(__s0_197, -__s1_197, __s2_197, __p3_197); \
43396  __ret_197; \
43397 })
43398 #else
43399 #define vfmsq_laneq_f16(__p0_198, __p1_198, __p2_198, __p3_198) __extension__ ({ \
43400  float16x8_t __s0_198 = __p0_198; \
43401  float16x8_t __s1_198 = __p1_198; \
43402  float16x8_t __s2_198 = __p2_198; \
43403  float16x8_t __rev0_198; __rev0_198 = __builtin_shufflevector(__s0_198, __s0_198, 7, 6, 5, 4, 3, 2, 1, 0); \
43404  float16x8_t __rev1_198; __rev1_198 = __builtin_shufflevector(__s1_198, __s1_198, 7, 6, 5, 4, 3, 2, 1, 0); \
43405  float16x8_t __rev2_198; __rev2_198 = __builtin_shufflevector(__s2_198, __s2_198, 7, 6, 5, 4, 3, 2, 1, 0); \
43406  float16x8_t __ret_198; \
43407  __ret_198 = __noswap_vfmaq_laneq_f16(__rev0_198, -__rev1_198, __rev2_198, __p3_198); \
43408  __ret_198 = __builtin_shufflevector(__ret_198, __ret_198, 7, 6, 5, 4, 3, 2, 1, 0); \
43409  __ret_198; \
43410 })
43411 #endif
43412 
43413 #ifdef __LITTLE_ENDIAN__
43414 #define vfms_laneq_f16(__p0_199, __p1_199, __p2_199, __p3_199) __extension__ ({ \
43415  float16x4_t __s0_199 = __p0_199; \
43416  float16x4_t __s1_199 = __p1_199; \
43417  float16x8_t __s2_199 = __p2_199; \
43418  float16x4_t __ret_199; \
43419  __ret_199 = vfma_laneq_f16(__s0_199, -__s1_199, __s2_199, __p3_199); \
43420  __ret_199; \
43421 })
43422 #else
43423 #define vfms_laneq_f16(__p0_200, __p1_200, __p2_200, __p3_200) __extension__ ({ \
43424  float16x4_t __s0_200 = __p0_200; \
43425  float16x4_t __s1_200 = __p1_200; \
43426  float16x8_t __s2_200 = __p2_200; \
43427  float16x4_t __rev0_200; __rev0_200 = __builtin_shufflevector(__s0_200, __s0_200, 3, 2, 1, 0); \
43428  float16x4_t __rev1_200; __rev1_200 = __builtin_shufflevector(__s1_200, __s1_200, 3, 2, 1, 0); \
43429  float16x8_t __rev2_200; __rev2_200 = __builtin_shufflevector(__s2_200, __s2_200, 7, 6, 5, 4, 3, 2, 1, 0); \
43430  float16x4_t __ret_200; \
43431  __ret_200 = __noswap_vfma_laneq_f16(__rev0_200, -__rev1_200, __rev2_200, __p3_200); \
43432  __ret_200 = __builtin_shufflevector(__ret_200, __ret_200, 3, 2, 1, 0); \
43433  __ret_200; \
43434 })
43435 #endif
43436 
43437 #ifdef __LITTLE_ENDIAN__
43438 #define vfmsq_n_f16(__p0, __p1, __p2) __extension__ ({ \
43439  float16x8_t __s0 = __p0; \
43440  float16x8_t __s1 = __p1; \
43441  float16_t __s2 = __p2; \
43442  float16x8_t __ret; \
43443  __ret = vfmaq_f16(__s0, -__s1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
43444  __ret; \
43445 })
43446 #else
43447 #define vfmsq_n_f16(__p0, __p1, __p2) __extension__ ({ \
43448  float16x8_t __s0 = __p0; \
43449  float16x8_t __s1 = __p1; \
43450  float16_t __s2 = __p2; \
43451  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43452  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
43453  float16x8_t __ret; \
43454  __ret = __noswap_vfmaq_f16(__rev0, -__rev1, (float16x8_t) {__s2, __s2, __s2, __s2, __s2, __s2, __s2, __s2}); \
43455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
43456  __ret; \
43457 })
43458 #endif
43459 
43460 #ifdef __LITTLE_ENDIAN__
43461 #define vfms_n_f16(__p0, __p1, __p2) __extension__ ({ \
43462  float16x4_t __s0 = __p0; \
43463  float16x4_t __s1 = __p1; \
43464  float16_t __s2 = __p2; \
43465  float16x4_t __ret; \
43466  __ret = vfma_f16(__s0, -__s1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
43467  __ret; \
43468 })
43469 #else
43470 #define vfms_n_f16(__p0, __p1, __p2) __extension__ ({ \
43471  float16x4_t __s0 = __p0; \
43472  float16x4_t __s1 = __p1; \
43473  float16_t __s2 = __p2; \
43474  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43475  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
43476  float16x4_t __ret; \
43477  __ret = __noswap_vfma_f16(__rev0, -__rev1, (float16x4_t) {__s2, __s2, __s2, __s2}); \
43478  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
43479  __ret; \
43480 })
43481 #endif
43482 
43483 #ifdef __LITTLE_ENDIAN__
43484 #define vmaxnmvq_f16(__p0) __extension__ ({ \
43485  float16x8_t __s0 = __p0; \
43486  float16_t __ret; \
43487  __ret = (float16_t) __builtin_neon_vmaxnmvq_f16((int8x16_t)__s0); \
43488  __ret; \
43489 })
43490 #else
43491 #define vmaxnmvq_f16(__p0) __extension__ ({ \
43492  float16x8_t __s0 = __p0; \
43493  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43494  float16_t __ret; \
43495  __ret = (float16_t) __builtin_neon_vmaxnmvq_f16((int8x16_t)__rev0); \
43496  __ret; \
43497 })
43498 #endif
43499 
43500 #ifdef __LITTLE_ENDIAN__
43501 #define vmaxnmv_f16(__p0) __extension__ ({ \
43502  float16x4_t __s0 = __p0; \
43503  float16_t __ret; \
43504  __ret = (float16_t) __builtin_neon_vmaxnmv_f16((int8x8_t)__s0); \
43505  __ret; \
43506 })
43507 #else
43508 #define vmaxnmv_f16(__p0) __extension__ ({ \
43509  float16x4_t __s0 = __p0; \
43510  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43511  float16_t __ret; \
43512  __ret = (float16_t) __builtin_neon_vmaxnmv_f16((int8x8_t)__rev0); \
43513  __ret; \
43514 })
43515 #endif
43516 
43517 #ifdef __LITTLE_ENDIAN__
43518 #define vmaxvq_f16(__p0) __extension__ ({ \
43519  float16x8_t __s0 = __p0; \
43520  float16_t __ret; \
43521  __ret = (float16_t) __builtin_neon_vmaxvq_f16((int8x16_t)__s0); \
43522  __ret; \
43523 })
43524 #else
43525 #define vmaxvq_f16(__p0) __extension__ ({ \
43526  float16x8_t __s0 = __p0; \
43527  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43528  float16_t __ret; \
43529  __ret = (float16_t) __builtin_neon_vmaxvq_f16((int8x16_t)__rev0); \
43530  __ret; \
43531 })
43532 #endif
43533 
43534 #ifdef __LITTLE_ENDIAN__
43535 #define vmaxv_f16(__p0) __extension__ ({ \
43536  float16x4_t __s0 = __p0; \
43537  float16_t __ret; \
43538  __ret = (float16_t) __builtin_neon_vmaxv_f16((int8x8_t)__s0); \
43539  __ret; \
43540 })
43541 #else
43542 #define vmaxv_f16(__p0) __extension__ ({ \
43543  float16x4_t __s0 = __p0; \
43544  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43545  float16_t __ret; \
43546  __ret = (float16_t) __builtin_neon_vmaxv_f16((int8x8_t)__rev0); \
43547  __ret; \
43548 })
43549 #endif
43550 
43551 #ifdef __LITTLE_ENDIAN__
43552 #define vminnmvq_f16(__p0) __extension__ ({ \
43553  float16x8_t __s0 = __p0; \
43554  float16_t __ret; \
43555  __ret = (float16_t) __builtin_neon_vminnmvq_f16((int8x16_t)__s0); \
43556  __ret; \
43557 })
43558 #else
43559 #define vminnmvq_f16(__p0) __extension__ ({ \
43560  float16x8_t __s0 = __p0; \
43561  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43562  float16_t __ret; \
43563  __ret = (float16_t) __builtin_neon_vminnmvq_f16((int8x16_t)__rev0); \
43564  __ret; \
43565 })
43566 #endif
43567 
43568 #ifdef __LITTLE_ENDIAN__
43569 #define vminnmv_f16(__p0) __extension__ ({ \
43570  float16x4_t __s0 = __p0; \
43571  float16_t __ret; \
43572  __ret = (float16_t) __builtin_neon_vminnmv_f16((int8x8_t)__s0); \
43573  __ret; \
43574 })
43575 #else
43576 #define vminnmv_f16(__p0) __extension__ ({ \
43577  float16x4_t __s0 = __p0; \
43578  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43579  float16_t __ret; \
43580  __ret = (float16_t) __builtin_neon_vminnmv_f16((int8x8_t)__rev0); \
43581  __ret; \
43582 })
43583 #endif
43584 
43585 #ifdef __LITTLE_ENDIAN__
43586 #define vminvq_f16(__p0) __extension__ ({ \
43587  float16x8_t __s0 = __p0; \
43588  float16_t __ret; \
43589  __ret = (float16_t) __builtin_neon_vminvq_f16((int8x16_t)__s0); \
43590  __ret; \
43591 })
43592 #else
43593 #define vminvq_f16(__p0) __extension__ ({ \
43594  float16x8_t __s0 = __p0; \
43595  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43596  float16_t __ret; \
43597  __ret = (float16_t) __builtin_neon_vminvq_f16((int8x16_t)__rev0); \
43598  __ret; \
43599 })
43600 #endif
43601 
43602 #ifdef __LITTLE_ENDIAN__
43603 #define vminv_f16(__p0) __extension__ ({ \
43604  float16x4_t __s0 = __p0; \
43605  float16_t __ret; \
43606  __ret = (float16_t) __builtin_neon_vminv_f16((int8x8_t)__s0); \
43607  __ret; \
43608 })
43609 #else
43610 #define vminv_f16(__p0) __extension__ ({ \
43611  float16x4_t __s0 = __p0; \
43612  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43613  float16_t __ret; \
43614  __ret = (float16_t) __builtin_neon_vminv_f16((int8x8_t)__rev0); \
43615  __ret; \
43616 })
43617 #endif
43618 
43619 #ifdef __LITTLE_ENDIAN__
43620 #define vmulq_laneq_f16(__p0_201, __p1_201, __p2_201) __extension__ ({ \
43621  float16x8_t __s0_201 = __p0_201; \
43622  float16x8_t __s1_201 = __p1_201; \
43623  float16x8_t __ret_201; \
43624  __ret_201 = __s0_201 * splatq_laneq_f16(__s1_201, __p2_201); \
43625  __ret_201; \
43626 })
43627 #else
43628 #define vmulq_laneq_f16(__p0_202, __p1_202, __p2_202) __extension__ ({ \
43629  float16x8_t __s0_202 = __p0_202; \
43630  float16x8_t __s1_202 = __p1_202; \
43631  float16x8_t __rev0_202; __rev0_202 = __builtin_shufflevector(__s0_202, __s0_202, 7, 6, 5, 4, 3, 2, 1, 0); \
43632  float16x8_t __rev1_202; __rev1_202 = __builtin_shufflevector(__s1_202, __s1_202, 7, 6, 5, 4, 3, 2, 1, 0); \
43633  float16x8_t __ret_202; \
43634  __ret_202 = __rev0_202 * __noswap_splatq_laneq_f16(__rev1_202, __p2_202); \
43635  __ret_202 = __builtin_shufflevector(__ret_202, __ret_202, 7, 6, 5, 4, 3, 2, 1, 0); \
43636  __ret_202; \
43637 })
43638 #endif
43639 
43640 #ifdef __LITTLE_ENDIAN__
43641 #define vmul_laneq_f16(__p0_203, __p1_203, __p2_203) __extension__ ({ \
43642  float16x4_t __s0_203 = __p0_203; \
43643  float16x8_t __s1_203 = __p1_203; \
43644  float16x4_t __ret_203; \
43645  __ret_203 = __s0_203 * splat_laneq_f16(__s1_203, __p2_203); \
43646  __ret_203; \
43647 })
43648 #else
43649 #define vmul_laneq_f16(__p0_204, __p1_204, __p2_204) __extension__ ({ \
43650  float16x4_t __s0_204 = __p0_204; \
43651  float16x8_t __s1_204 = __p1_204; \
43652  float16x4_t __rev0_204; __rev0_204 = __builtin_shufflevector(__s0_204, __s0_204, 3, 2, 1, 0); \
43653  float16x8_t __rev1_204; __rev1_204 = __builtin_shufflevector(__s1_204, __s1_204, 7, 6, 5, 4, 3, 2, 1, 0); \
43654  float16x4_t __ret_204; \
43655  __ret_204 = __rev0_204 * __noswap_splat_laneq_f16(__rev1_204, __p2_204); \
43656  __ret_204 = __builtin_shufflevector(__ret_204, __ret_204, 3, 2, 1, 0); \
43657  __ret_204; \
43658 })
43659 #endif
43660 
43661 #ifdef __LITTLE_ENDIAN__
43662 __ai float16x8_t vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
43663  float16x8_t __ret;
43664  __ret = (float16x8_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43665  return __ret;
43666 }
43667 #else
43668 __ai float16x8_t vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
43669  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43670  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43671  float16x8_t __ret;
43672  __ret = (float16x8_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43673  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43674  return __ret;
43675 }
43676 __ai float16x8_t __noswap_vmulxq_f16(float16x8_t __p0, float16x8_t __p1) {
43677  float16x8_t __ret;
43678  __ret = (float16x8_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43679  return __ret;
43680 }
43681 #endif
43682 
43683 #ifdef __LITTLE_ENDIAN__
43684 __ai float16x4_t vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
43685  float16x4_t __ret;
43686  __ret = (float16x4_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
43687  return __ret;
43688 }
43689 #else
43690 __ai float16x4_t vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
43691  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43692  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
43693  float16x4_t __ret;
43694  __ret = (float16x4_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
43695  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43696  return __ret;
43697 }
43698 __ai float16x4_t __noswap_vmulx_f16(float16x4_t __p0, float16x4_t __p1) {
43699  float16x4_t __ret;
43700  __ret = (float16x4_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
43701  return __ret;
43702 }
43703 #endif
43704 
43705 #ifdef __LITTLE_ENDIAN__
43706 #define vmulxh_lane_f16(__p0, __p1, __p2) __extension__ ({ \
43707  float16_t __s0 = __p0; \
43708  float16x4_t __s1 = __p1; \
43709  float16_t __ret; \
43710  __ret = (float16_t) __builtin_neon_vmulxh_lane_f16(__s0, (float16x4_t)__s1, __p2); \
43711  __ret; \
43712 })
43713 #else
43714 #define vmulxh_lane_f16(__p0, __p1, __p2) __extension__ ({ \
43715  float16_t __s0 = __p0; \
43716  float16x4_t __s1 = __p1; \
43717  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
43718  float16_t __ret; \
43719  __ret = (float16_t) __builtin_neon_vmulxh_lane_f16(__s0, (float16x4_t)__rev1, __p2); \
43720  __ret; \
43721 })
43722 #endif
43723 
43724 #ifdef __LITTLE_ENDIAN__
43725 #define vmulxq_lane_f16(__p0_205, __p1_205, __p2_205) __extension__ ({ \
43726  float16x8_t __s0_205 = __p0_205; \
43727  float16x4_t __s1_205 = __p1_205; \
43728  float16x8_t __ret_205; \
43729  __ret_205 = vmulxq_f16(__s0_205, splatq_lane_f16(__s1_205, __p2_205)); \
43730  __ret_205; \
43731 })
43732 #else
43733 #define vmulxq_lane_f16(__p0_206, __p1_206, __p2_206) __extension__ ({ \
43734  float16x8_t __s0_206 = __p0_206; \
43735  float16x4_t __s1_206 = __p1_206; \
43736  float16x8_t __rev0_206; __rev0_206 = __builtin_shufflevector(__s0_206, __s0_206, 7, 6, 5, 4, 3, 2, 1, 0); \
43737  float16x4_t __rev1_206; __rev1_206 = __builtin_shufflevector(__s1_206, __s1_206, 3, 2, 1, 0); \
43738  float16x8_t __ret_206; \
43739  __ret_206 = __noswap_vmulxq_f16(__rev0_206, __noswap_splatq_lane_f16(__rev1_206, __p2_206)); \
43740  __ret_206 = __builtin_shufflevector(__ret_206, __ret_206, 7, 6, 5, 4, 3, 2, 1, 0); \
43741  __ret_206; \
43742 })
43743 #endif
43744 
43745 #ifdef __LITTLE_ENDIAN__
43746 #define vmulx_lane_f16(__p0_207, __p1_207, __p2_207) __extension__ ({ \
43747  float16x4_t __s0_207 = __p0_207; \
43748  float16x4_t __s1_207 = __p1_207; \
43749  float16x4_t __ret_207; \
43750  __ret_207 = vmulx_f16(__s0_207, splat_lane_f16(__s1_207, __p2_207)); \
43751  __ret_207; \
43752 })
43753 #else
43754 #define vmulx_lane_f16(__p0_208, __p1_208, __p2_208) __extension__ ({ \
43755  float16x4_t __s0_208 = __p0_208; \
43756  float16x4_t __s1_208 = __p1_208; \
43757  float16x4_t __rev0_208; __rev0_208 = __builtin_shufflevector(__s0_208, __s0_208, 3, 2, 1, 0); \
43758  float16x4_t __rev1_208; __rev1_208 = __builtin_shufflevector(__s1_208, __s1_208, 3, 2, 1, 0); \
43759  float16x4_t __ret_208; \
43760  __ret_208 = __noswap_vmulx_f16(__rev0_208, __noswap_splat_lane_f16(__rev1_208, __p2_208)); \
43761  __ret_208 = __builtin_shufflevector(__ret_208, __ret_208, 3, 2, 1, 0); \
43762  __ret_208; \
43763 })
43764 #endif
43765 
43766 #ifdef __LITTLE_ENDIAN__
43767 #define vmulxh_laneq_f16(__p0, __p1, __p2) __extension__ ({ \
43768  float16_t __s0 = __p0; \
43769  float16x8_t __s1 = __p1; \
43770  float16_t __ret; \
43771  __ret = (float16_t) __builtin_neon_vmulxh_laneq_f16(__s0, (float16x8_t)__s1, __p2); \
43772  __ret; \
43773 })
43774 #else
43775 #define vmulxh_laneq_f16(__p0, __p1, __p2) __extension__ ({ \
43776  float16_t __s0 = __p0; \
43777  float16x8_t __s1 = __p1; \
43778  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
43779  float16_t __ret; \
43780  __ret = (float16_t) __builtin_neon_vmulxh_laneq_f16(__s0, (float16x8_t)__rev1, __p2); \
43781  __ret; \
43782 })
43783 #endif
43784 
43785 #ifdef __LITTLE_ENDIAN__
43786 #define vmulxq_laneq_f16(__p0_209, __p1_209, __p2_209) __extension__ ({ \
43787  float16x8_t __s0_209 = __p0_209; \
43788  float16x8_t __s1_209 = __p1_209; \
43789  float16x8_t __ret_209; \
43790  __ret_209 = vmulxq_f16(__s0_209, splatq_laneq_f16(__s1_209, __p2_209)); \
43791  __ret_209; \
43792 })
43793 #else
43794 #define vmulxq_laneq_f16(__p0_210, __p1_210, __p2_210) __extension__ ({ \
43795  float16x8_t __s0_210 = __p0_210; \
43796  float16x8_t __s1_210 = __p1_210; \
43797  float16x8_t __rev0_210; __rev0_210 = __builtin_shufflevector(__s0_210, __s0_210, 7, 6, 5, 4, 3, 2, 1, 0); \
43798  float16x8_t __rev1_210; __rev1_210 = __builtin_shufflevector(__s1_210, __s1_210, 7, 6, 5, 4, 3, 2, 1, 0); \
43799  float16x8_t __ret_210; \
43800  __ret_210 = __noswap_vmulxq_f16(__rev0_210, __noswap_splatq_laneq_f16(__rev1_210, __p2_210)); \
43801  __ret_210 = __builtin_shufflevector(__ret_210, __ret_210, 7, 6, 5, 4, 3, 2, 1, 0); \
43802  __ret_210; \
43803 })
43804 #endif
43805 
43806 #ifdef __LITTLE_ENDIAN__
43807 #define vmulx_laneq_f16(__p0_211, __p1_211, __p2_211) __extension__ ({ \
43808  float16x4_t __s0_211 = __p0_211; \
43809  float16x8_t __s1_211 = __p1_211; \
43810  float16x4_t __ret_211; \
43811  __ret_211 = vmulx_f16(__s0_211, splat_laneq_f16(__s1_211, __p2_211)); \
43812  __ret_211; \
43813 })
43814 #else
43815 #define vmulx_laneq_f16(__p0_212, __p1_212, __p2_212) __extension__ ({ \
43816  float16x4_t __s0_212 = __p0_212; \
43817  float16x8_t __s1_212 = __p1_212; \
43818  float16x4_t __rev0_212; __rev0_212 = __builtin_shufflevector(__s0_212, __s0_212, 3, 2, 1, 0); \
43819  float16x8_t __rev1_212; __rev1_212 = __builtin_shufflevector(__s1_212, __s1_212, 7, 6, 5, 4, 3, 2, 1, 0); \
43820  float16x4_t __ret_212; \
43821  __ret_212 = __noswap_vmulx_f16(__rev0_212, __noswap_splat_laneq_f16(__rev1_212, __p2_212)); \
43822  __ret_212 = __builtin_shufflevector(__ret_212, __ret_212, 3, 2, 1, 0); \
43823  __ret_212; \
43824 })
43825 #endif
43826 
43827 #ifdef __LITTLE_ENDIAN__
43828 #define vmulxq_n_f16(__p0, __p1) __extension__ ({ \
43829  float16x8_t __s0 = __p0; \
43830  float16_t __s1 = __p1; \
43831  float16x8_t __ret; \
43832  __ret = vmulxq_f16(__s0, (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}); \
43833  __ret; \
43834 })
43835 #else
43836 #define vmulxq_n_f16(__p0, __p1) __extension__ ({ \
43837  float16x8_t __s0 = __p0; \
43838  float16_t __s1 = __p1; \
43839  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
43840  float16x8_t __ret; \
43841  __ret = __noswap_vmulxq_f16(__rev0, (float16x8_t) {__s1, __s1, __s1, __s1, __s1, __s1, __s1, __s1}); \
43842  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
43843  __ret; \
43844 })
43845 #endif
43846 
43847 #ifdef __LITTLE_ENDIAN__
43848 #define vmulx_n_f16(__p0, __p1) __extension__ ({ \
43849  float16x4_t __s0 = __p0; \
43850  float16_t __s1 = __p1; \
43851  float16x4_t __ret; \
43852  __ret = vmulx_f16(__s0, (float16x4_t) {__s1, __s1, __s1, __s1}); \
43853  __ret; \
43854 })
43855 #else
43856 #define vmulx_n_f16(__p0, __p1) __extension__ ({ \
43857  float16x4_t __s0 = __p0; \
43858  float16_t __s1 = __p1; \
43859  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
43860  float16x4_t __ret; \
43861  __ret = __noswap_vmulx_f16(__rev0, (float16x4_t) {__s1, __s1, __s1, __s1}); \
43862  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
43863  __ret; \
43864 })
43865 #endif
43866 
43867 #ifdef __LITTLE_ENDIAN__
43868 __ai float16x8_t vpaddq_f16(float16x8_t __p0, float16x8_t __p1) {
43869  float16x8_t __ret;
43870  __ret = (float16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43871  return __ret;
43872 }
43873 #else
43874 __ai float16x8_t vpaddq_f16(float16x8_t __p0, float16x8_t __p1) {
43875  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43876  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43877  float16x8_t __ret;
43878  __ret = (float16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43880  return __ret;
43881 }
43882 #endif
43883 
43884 #ifdef __LITTLE_ENDIAN__
43885 __ai float16x8_t vpmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
43886  float16x8_t __ret;
43887  __ret = (float16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43888  return __ret;
43889 }
43890 #else
43891 __ai float16x8_t vpmaxq_f16(float16x8_t __p0, float16x8_t __p1) {
43892  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43893  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43894  float16x8_t __ret;
43895  __ret = (float16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43896  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43897  return __ret;
43898 }
43899 #endif
43900 
43901 #ifdef __LITTLE_ENDIAN__
43902 __ai float16x8_t vpmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
43903  float16x8_t __ret;
43904  __ret = (float16x8_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43905  return __ret;
43906 }
43907 #else
43908 __ai float16x8_t vpmaxnmq_f16(float16x8_t __p0, float16x8_t __p1) {
43909  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43910  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43911  float16x8_t __ret;
43912  __ret = (float16x8_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43913  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43914  return __ret;
43915 }
43916 #endif
43917 
43918 #ifdef __LITTLE_ENDIAN__
43919 __ai float16x4_t vpmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
43920  float16x4_t __ret;
43921  __ret = (float16x4_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
43922  return __ret;
43923 }
43924 #else
43925 __ai float16x4_t vpmaxnm_f16(float16x4_t __p0, float16x4_t __p1) {
43926  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43927  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
43928  float16x4_t __ret;
43929  __ret = (float16x4_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
43930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43931  return __ret;
43932 }
43933 #endif
43934 
43935 #ifdef __LITTLE_ENDIAN__
43936 __ai float16x8_t vpminq_f16(float16x8_t __p0, float16x8_t __p1) {
43937  float16x8_t __ret;
43938  __ret = (float16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43939  return __ret;
43940 }
43941 #else
43942 __ai float16x8_t vpminq_f16(float16x8_t __p0, float16x8_t __p1) {
43943  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43944  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43945  float16x8_t __ret;
43946  __ret = (float16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43947  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43948  return __ret;
43949 }
43950 #endif
43951 
43952 #ifdef __LITTLE_ENDIAN__
43953 __ai float16x8_t vpminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
43954  float16x8_t __ret;
43955  __ret = (float16x8_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 40);
43956  return __ret;
43957 }
43958 #else
43959 __ai float16x8_t vpminnmq_f16(float16x8_t __p0, float16x8_t __p1) {
43960  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43961  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
43962  float16x8_t __ret;
43963  __ret = (float16x8_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 40);
43964  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43965  return __ret;
43966 }
43967 #endif
43968 
43969 #ifdef __LITTLE_ENDIAN__
43970 __ai float16x4_t vpminnm_f16(float16x4_t __p0, float16x4_t __p1) {
43971  float16x4_t __ret;
43972  __ret = (float16x4_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 8);
43973  return __ret;
43974 }
43975 #else
43976 __ai float16x4_t vpminnm_f16(float16x4_t __p0, float16x4_t __p1) {
43977  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43978  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
43979  float16x4_t __ret;
43980  __ret = (float16x4_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 8);
43981  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43982  return __ret;
43983 }
43984 #endif
43985 
43986 #ifdef __LITTLE_ENDIAN__
43987 __ai float16x8_t vrndiq_f16(float16x8_t __p0) {
43988  float16x8_t __ret;
43989  __ret = (float16x8_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 40);
43990  return __ret;
43991 }
43992 #else
43993 __ai float16x8_t vrndiq_f16(float16x8_t __p0) {
43994  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43995  float16x8_t __ret;
43996  __ret = (float16x8_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 40);
43997  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43998  return __ret;
43999 }
44000 #endif
44001 
44002 #ifdef __LITTLE_ENDIAN__
44003 __ai float16x4_t vrndi_f16(float16x4_t __p0) {
44004  float16x4_t __ret;
44005  __ret = (float16x4_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 8);
44006  return __ret;
44007 }
44008 #else
44009 __ai float16x4_t vrndi_f16(float16x4_t __p0) {
44010  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44011  float16x4_t __ret;
44012  __ret = (float16x4_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 8);
44013  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44014  return __ret;
44015 }
44016 #endif
44017 
44018 #ifdef __LITTLE_ENDIAN__
44019 __ai float16x8_t vsqrtq_f16(float16x8_t __p0) {
44020  float16x8_t __ret;
44021  __ret = (float16x8_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 40);
44022  return __ret;
44023 }
44024 #else
44025 __ai float16x8_t vsqrtq_f16(float16x8_t __p0) {
44026  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44027  float16x8_t __ret;
44028  __ret = (float16x8_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 40);
44029  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44030  return __ret;
44031 }
44032 #endif
44033 
44034 #ifdef __LITTLE_ENDIAN__
44035 __ai float16x4_t vsqrt_f16(float16x4_t __p0) {
44036  float16x4_t __ret;
44037  __ret = (float16x4_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 8);
44038  return __ret;
44039 }
44040 #else
44041 __ai float16x4_t vsqrt_f16(float16x4_t __p0) {
44042  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44043  float16x4_t __ret;
44044  __ret = (float16x4_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 8);
44045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44046  return __ret;
44047 }
44048 #endif
44049 
44050 #ifdef __LITTLE_ENDIAN__
44051 __ai float16x8_t vtrn1q_f16(float16x8_t __p0, float16x8_t __p1) {
44052  float16x8_t __ret;
44053  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
44054  return __ret;
44055 }
44056 #else
44057 __ai float16x8_t vtrn1q_f16(float16x8_t __p0, float16x8_t __p1) {
44058  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44059  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44060  float16x8_t __ret;
44061  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
44062  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44063  return __ret;
44064 }
44065 #endif
44066 
44067 #ifdef __LITTLE_ENDIAN__
44068 __ai float16x4_t vtrn1_f16(float16x4_t __p0, float16x4_t __p1) {
44069  float16x4_t __ret;
44070  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
44071  return __ret;
44072 }
44073 #else
44074 __ai float16x4_t vtrn1_f16(float16x4_t __p0, float16x4_t __p1) {
44075  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44076  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44077  float16x4_t __ret;
44078  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
44079  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44080  return __ret;
44081 }
44082 #endif
44083 
44084 #ifdef __LITTLE_ENDIAN__
44085 __ai float16x8_t vtrn2q_f16(float16x8_t __p0, float16x8_t __p1) {
44086  float16x8_t __ret;
44087  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
44088  return __ret;
44089 }
44090 #else
44091 __ai float16x8_t vtrn2q_f16(float16x8_t __p0, float16x8_t __p1) {
44092  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44093  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44094  float16x8_t __ret;
44095  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
44096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44097  return __ret;
44098 }
44099 #endif
44100 
44101 #ifdef __LITTLE_ENDIAN__
44102 __ai float16x4_t vtrn2_f16(float16x4_t __p0, float16x4_t __p1) {
44103  float16x4_t __ret;
44104  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
44105  return __ret;
44106 }
44107 #else
44108 __ai float16x4_t vtrn2_f16(float16x4_t __p0, float16x4_t __p1) {
44109  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44110  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44111  float16x4_t __ret;
44112  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
44113  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44114  return __ret;
44115 }
44116 #endif
44117 
44118 #ifdef __LITTLE_ENDIAN__
44119 __ai float16x8_t vuzp1q_f16(float16x8_t __p0, float16x8_t __p1) {
44120  float16x8_t __ret;
44121  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
44122  return __ret;
44123 }
44124 #else
44125 __ai float16x8_t vuzp1q_f16(float16x8_t __p0, float16x8_t __p1) {
44126  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44127  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44128  float16x8_t __ret;
44129  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
44130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44131  return __ret;
44132 }
44133 #endif
44134 
44135 #ifdef __LITTLE_ENDIAN__
44136 __ai float16x4_t vuzp1_f16(float16x4_t __p0, float16x4_t __p1) {
44137  float16x4_t __ret;
44138  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
44139  return __ret;
44140 }
44141 #else
44142 __ai float16x4_t vuzp1_f16(float16x4_t __p0, float16x4_t __p1) {
44143  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44144  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44145  float16x4_t __ret;
44146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
44147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44148  return __ret;
44149 }
44150 #endif
44151 
44152 #ifdef __LITTLE_ENDIAN__
44153 __ai float16x8_t vuzp2q_f16(float16x8_t __p0, float16x8_t __p1) {
44154  float16x8_t __ret;
44155  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
44156  return __ret;
44157 }
44158 #else
44159 __ai float16x8_t vuzp2q_f16(float16x8_t __p0, float16x8_t __p1) {
44160  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44161  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44162  float16x8_t __ret;
44163  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
44164  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44165  return __ret;
44166 }
44167 #endif
44168 
44169 #ifdef __LITTLE_ENDIAN__
44170 __ai float16x4_t vuzp2_f16(float16x4_t __p0, float16x4_t __p1) {
44171  float16x4_t __ret;
44172  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
44173  return __ret;
44174 }
44175 #else
44176 __ai float16x4_t vuzp2_f16(float16x4_t __p0, float16x4_t __p1) {
44177  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44178  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44179  float16x4_t __ret;
44180  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
44181  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44182  return __ret;
44183 }
44184 #endif
44185 
44186 #ifdef __LITTLE_ENDIAN__
44187 __ai float16x8_t vzip1q_f16(float16x8_t __p0, float16x8_t __p1) {
44188  float16x8_t __ret;
44189  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
44190  return __ret;
44191 }
44192 #else
44193 __ai float16x8_t vzip1q_f16(float16x8_t __p0, float16x8_t __p1) {
44194  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44195  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44196  float16x8_t __ret;
44197  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
44198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44199  return __ret;
44200 }
44201 #endif
44202 
44203 #ifdef __LITTLE_ENDIAN__
44204 __ai float16x4_t vzip1_f16(float16x4_t __p0, float16x4_t __p1) {
44205  float16x4_t __ret;
44206  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
44207  return __ret;
44208 }
44209 #else
44210 __ai float16x4_t vzip1_f16(float16x4_t __p0, float16x4_t __p1) {
44211  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44212  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44213  float16x4_t __ret;
44214  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
44215  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44216  return __ret;
44217 }
44218 #endif
44219 
44220 #ifdef __LITTLE_ENDIAN__
44221 __ai float16x8_t vzip2q_f16(float16x8_t __p0, float16x8_t __p1) {
44222  float16x8_t __ret;
44223  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
44224  return __ret;
44225 }
44226 #else
44227 __ai float16x8_t vzip2q_f16(float16x8_t __p0, float16x8_t __p1) {
44228  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44229  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44230  float16x8_t __ret;
44231  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
44232  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44233  return __ret;
44234 }
44235 #endif
44236 
44237 #ifdef __LITTLE_ENDIAN__
44238 __ai float16x4_t vzip2_f16(float16x4_t __p0, float16x4_t __p1) {
44239  float16x4_t __ret;
44240  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
44241  return __ret;
44242 }
44243 #else
44244 __ai float16x4_t vzip2_f16(float16x4_t __p0, float16x4_t __p1) {
44245  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44246  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44247  float16x4_t __ret;
44248  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
44249  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44250  return __ret;
44251 }
44252 #endif
44253 
44254 #endif
44255 #if defined(__ARM_FEATURE_MATMUL_INT8)
44256 #ifdef __LITTLE_ENDIAN__
44257 __ai uint32x4_t vmmlaq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
44258  uint32x4_t __ret;
44259  __ret = (uint32x4_t) __builtin_neon_vmmlaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
44260  return __ret;
44261 }
44262 #else
44263 __ai uint32x4_t vmmlaq_u32(uint32x4_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
44264  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44265  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44266  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44267  uint32x4_t __ret;
44268  __ret = (uint32x4_t) __builtin_neon_vmmlaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
44269  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44270  return __ret;
44271 }
44272 #endif
44273 
44274 #ifdef __LITTLE_ENDIAN__
44275 __ai int32x4_t vmmlaq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
44276  int32x4_t __ret;
44277  __ret = (int32x4_t) __builtin_neon_vmmlaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
44278  return __ret;
44279 }
44280 #else
44281 __ai int32x4_t vmmlaq_s32(int32x4_t __p0, int8x16_t __p1, int8x16_t __p2) {
44282  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44283  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44284  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44285  int32x4_t __ret;
44286  __ret = (int32x4_t) __builtin_neon_vmmlaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
44287  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44288  return __ret;
44289 }
44290 #endif
44291 
44292 #ifdef __LITTLE_ENDIAN__
44293 __ai int32x4_t vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
44294  int32x4_t __ret;
44295  __ret = (int32x4_t) __builtin_neon_vusdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
44296  return __ret;
44297 }
44298 #else
44299 __ai int32x4_t vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
44300  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44301  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44302  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44303  int32x4_t __ret;
44304  __ret = (int32x4_t) __builtin_neon_vusdotq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
44305  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44306  return __ret;
44307 }
44308 __ai int32x4_t __noswap_vusdotq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
44309  int32x4_t __ret;
44310  __ret = (int32x4_t) __builtin_neon_vusdotq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
44311  return __ret;
44312 }
44313 #endif
44314 
44315 #ifdef __LITTLE_ENDIAN__
44316 __ai int32x2_t vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
44317  int32x2_t __ret;
44318  __ret = (int32x2_t) __builtin_neon_vusdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
44319  return __ret;
44320 }
44321 #else
44322 __ai int32x2_t vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
44323  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44324  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44325  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
44326  int32x2_t __ret;
44327  __ret = (int32x2_t) __builtin_neon_vusdot_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
44328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44329  return __ret;
44330 }
44331 __ai int32x2_t __noswap_vusdot_s32(int32x2_t __p0, uint8x8_t __p1, int8x8_t __p2) {
44332  int32x2_t __ret;
44333  __ret = (int32x2_t) __builtin_neon_vusdot_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
44334  return __ret;
44335 }
44336 #endif
44337 
44338 #ifdef __LITTLE_ENDIAN__
44339 #define vusdotq_lane_s32(__p0_213, __p1_213, __p2_213, __p3_213) __extension__ ({ \
44340  int32x4_t __s0_213 = __p0_213; \
44341  uint8x16_t __s1_213 = __p1_213; \
44342  int8x8_t __s2_213 = __p2_213; \
44343  int32x4_t __ret_213; \
44344 int8x8_t __reint_213 = __s2_213; \
44345  __ret_213 = vusdotq_s32(__s0_213, __s1_213, (int8x16_t)(splatq_lane_s32(*(int32x2_t *) &__reint_213, __p3_213))); \
44346  __ret_213; \
44347 })
44348 #else
44349 #define vusdotq_lane_s32(__p0_214, __p1_214, __p2_214, __p3_214) __extension__ ({ \
44350  int32x4_t __s0_214 = __p0_214; \
44351  uint8x16_t __s1_214 = __p1_214; \
44352  int8x8_t __s2_214 = __p2_214; \
44353  int32x4_t __rev0_214; __rev0_214 = __builtin_shufflevector(__s0_214, __s0_214, 3, 2, 1, 0); \
44354  uint8x16_t __rev1_214; __rev1_214 = __builtin_shufflevector(__s1_214, __s1_214, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44355  int8x8_t __rev2_214; __rev2_214 = __builtin_shufflevector(__s2_214, __s2_214, 7, 6, 5, 4, 3, 2, 1, 0); \
44356  int32x4_t __ret_214; \
44357 int8x8_t __reint_214 = __rev2_214; \
44358  __ret_214 = __noswap_vusdotq_s32(__rev0_214, __rev1_214, (int8x16_t)(__noswap_splatq_lane_s32(*(int32x2_t *) &__reint_214, __p3_214))); \
44359  __ret_214 = __builtin_shufflevector(__ret_214, __ret_214, 3, 2, 1, 0); \
44360  __ret_214; \
44361 })
44362 #endif
44363 
44364 #ifdef __LITTLE_ENDIAN__
44365 #define vusdot_lane_s32(__p0_215, __p1_215, __p2_215, __p3_215) __extension__ ({ \
44366  int32x2_t __s0_215 = __p0_215; \
44367  uint8x8_t __s1_215 = __p1_215; \
44368  int8x8_t __s2_215 = __p2_215; \
44369  int32x2_t __ret_215; \
44370 int8x8_t __reint_215 = __s2_215; \
44371  __ret_215 = vusdot_s32(__s0_215, __s1_215, (int8x8_t)(splat_lane_s32(*(int32x2_t *) &__reint_215, __p3_215))); \
44372  __ret_215; \
44373 })
44374 #else
44375 #define vusdot_lane_s32(__p0_216, __p1_216, __p2_216, __p3_216) __extension__ ({ \
44376  int32x2_t __s0_216 = __p0_216; \
44377  uint8x8_t __s1_216 = __p1_216; \
44378  int8x8_t __s2_216 = __p2_216; \
44379  int32x2_t __rev0_216; __rev0_216 = __builtin_shufflevector(__s0_216, __s0_216, 1, 0); \
44380  uint8x8_t __rev1_216; __rev1_216 = __builtin_shufflevector(__s1_216, __s1_216, 7, 6, 5, 4, 3, 2, 1, 0); \
44381  int8x8_t __rev2_216; __rev2_216 = __builtin_shufflevector(__s2_216, __s2_216, 7, 6, 5, 4, 3, 2, 1, 0); \
44382  int32x2_t __ret_216; \
44383 int8x8_t __reint_216 = __rev2_216; \
44384  __ret_216 = __noswap_vusdot_s32(__rev0_216, __rev1_216, (int8x8_t)(__noswap_splat_lane_s32(*(int32x2_t *) &__reint_216, __p3_216))); \
44385  __ret_216 = __builtin_shufflevector(__ret_216, __ret_216, 1, 0); \
44386  __ret_216; \
44387 })
44388 #endif
44389 
44390 #ifdef __LITTLE_ENDIAN__
44391 __ai int32x4_t vusmmlaq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
44392  int32x4_t __ret;
44393  __ret = (int32x4_t) __builtin_neon_vusmmlaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
44394  return __ret;
44395 }
44396 #else
44397 __ai int32x4_t vusmmlaq_s32(int32x4_t __p0, uint8x16_t __p1, int8x16_t __p2) {
44398  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44399  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44400  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
44401  int32x4_t __ret;
44402  __ret = (int32x4_t) __builtin_neon_vusmmlaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
44403  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44404  return __ret;
44405 }
44406 #endif
44407 
44408 #endif
44409 #if defined(__ARM_FEATURE_QRDMX)
44410 #ifdef __LITTLE_ENDIAN__
44411 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
44412  int32x4_t __ret;
44413  __ret = vqaddq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
44414  return __ret;
44415 }
44416 #else
44417 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
44418  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44419  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44420  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
44421  int32x4_t __ret;
44422  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
44423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44424  return __ret;
44425 }
44426 #endif
44427 
44428 #ifdef __LITTLE_ENDIAN__
44429 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
44430  int16x8_t __ret;
44431  __ret = vqaddq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
44432  return __ret;
44433 }
44434 #else
44435 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
44436  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44437  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44438  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
44439  int16x8_t __ret;
44440  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
44441  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44442  return __ret;
44443 }
44444 #endif
44445 
44446 #ifdef __LITTLE_ENDIAN__
44447 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
44448  int32x2_t __ret;
44449  __ret = vqadd_s32(__p0, vqrdmulh_s32(__p1, __p2));
44450  return __ret;
44451 }
44452 #else
44453 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
44454  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44455  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44456  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
44457  int32x2_t __ret;
44458  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
44459  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44460  return __ret;
44461 }
44462 #endif
44463 
44464 #ifdef __LITTLE_ENDIAN__
44465 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
44466  int16x4_t __ret;
44467  __ret = vqadd_s16(__p0, vqrdmulh_s16(__p1, __p2));
44468  return __ret;
44469 }
44470 #else
44471 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
44472  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44473  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44474  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
44475  int16x4_t __ret;
44476  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
44477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44478  return __ret;
44479 }
44480 #endif
44481 
44482 #ifdef __LITTLE_ENDIAN__
44483 #define vqrdmlahq_lane_s32(__p0_217, __p1_217, __p2_217, __p3_217) __extension__ ({ \
44484  int32x4_t __s0_217 = __p0_217; \
44485  int32x4_t __s1_217 = __p1_217; \
44486  int32x2_t __s2_217 = __p2_217; \
44487  int32x4_t __ret_217; \
44488  __ret_217 = vqaddq_s32(__s0_217, vqrdmulhq_s32(__s1_217, splatq_lane_s32(__s2_217, __p3_217))); \
44489  __ret_217; \
44490 })
44491 #else
44492 #define vqrdmlahq_lane_s32(__p0_218, __p1_218, __p2_218, __p3_218) __extension__ ({ \
44493  int32x4_t __s0_218 = __p0_218; \
44494  int32x4_t __s1_218 = __p1_218; \
44495  int32x2_t __s2_218 = __p2_218; \
44496  int32x4_t __rev0_218; __rev0_218 = __builtin_shufflevector(__s0_218, __s0_218, 3, 2, 1, 0); \
44497  int32x4_t __rev1_218; __rev1_218 = __builtin_shufflevector(__s1_218, __s1_218, 3, 2, 1, 0); \
44498  int32x2_t __rev2_218; __rev2_218 = __builtin_shufflevector(__s2_218, __s2_218, 1, 0); \
44499  int32x4_t __ret_218; \
44500  __ret_218 = __noswap_vqaddq_s32(__rev0_218, __noswap_vqrdmulhq_s32(__rev1_218, __noswap_splatq_lane_s32(__rev2_218, __p3_218))); \
44501  __ret_218 = __builtin_shufflevector(__ret_218, __ret_218, 3, 2, 1, 0); \
44502  __ret_218; \
44503 })
44504 #endif
44505 
44506 #ifdef __LITTLE_ENDIAN__
44507 #define vqrdmlahq_lane_s16(__p0_219, __p1_219, __p2_219, __p3_219) __extension__ ({ \
44508  int16x8_t __s0_219 = __p0_219; \
44509  int16x8_t __s1_219 = __p1_219; \
44510  int16x4_t __s2_219 = __p2_219; \
44511  int16x8_t __ret_219; \
44512  __ret_219 = vqaddq_s16(__s0_219, vqrdmulhq_s16(__s1_219, splatq_lane_s16(__s2_219, __p3_219))); \
44513  __ret_219; \
44514 })
44515 #else
44516 #define vqrdmlahq_lane_s16(__p0_220, __p1_220, __p2_220, __p3_220) __extension__ ({ \
44517  int16x8_t __s0_220 = __p0_220; \
44518  int16x8_t __s1_220 = __p1_220; \
44519  int16x4_t __s2_220 = __p2_220; \
44520  int16x8_t __rev0_220; __rev0_220 = __builtin_shufflevector(__s0_220, __s0_220, 7, 6, 5, 4, 3, 2, 1, 0); \
44521  int16x8_t __rev1_220; __rev1_220 = __builtin_shufflevector(__s1_220, __s1_220, 7, 6, 5, 4, 3, 2, 1, 0); \
44522  int16x4_t __rev2_220; __rev2_220 = __builtin_shufflevector(__s2_220, __s2_220, 3, 2, 1, 0); \
44523  int16x8_t __ret_220; \
44524  __ret_220 = __noswap_vqaddq_s16(__rev0_220, __noswap_vqrdmulhq_s16(__rev1_220, __noswap_splatq_lane_s16(__rev2_220, __p3_220))); \
44525  __ret_220 = __builtin_shufflevector(__ret_220, __ret_220, 7, 6, 5, 4, 3, 2, 1, 0); \
44526  __ret_220; \
44527 })
44528 #endif
44529 
44530 #ifdef __LITTLE_ENDIAN__
44531 #define vqrdmlah_lane_s32(__p0_221, __p1_221, __p2_221, __p3_221) __extension__ ({ \
44532  int32x2_t __s0_221 = __p0_221; \
44533  int32x2_t __s1_221 = __p1_221; \
44534  int32x2_t __s2_221 = __p2_221; \
44535  int32x2_t __ret_221; \
44536  __ret_221 = vqadd_s32(__s0_221, vqrdmulh_s32(__s1_221, splat_lane_s32(__s2_221, __p3_221))); \
44537  __ret_221; \
44538 })
44539 #else
44540 #define vqrdmlah_lane_s32(__p0_222, __p1_222, __p2_222, __p3_222) __extension__ ({ \
44541  int32x2_t __s0_222 = __p0_222; \
44542  int32x2_t __s1_222 = __p1_222; \
44543  int32x2_t __s2_222 = __p2_222; \
44544  int32x2_t __rev0_222; __rev0_222 = __builtin_shufflevector(__s0_222, __s0_222, 1, 0); \
44545  int32x2_t __rev1_222; __rev1_222 = __builtin_shufflevector(__s1_222, __s1_222, 1, 0); \
44546  int32x2_t __rev2_222; __rev2_222 = __builtin_shufflevector(__s2_222, __s2_222, 1, 0); \
44547  int32x2_t __ret_222; \
44548  __ret_222 = __noswap_vqadd_s32(__rev0_222, __noswap_vqrdmulh_s32(__rev1_222, __noswap_splat_lane_s32(__rev2_222, __p3_222))); \
44549  __ret_222 = __builtin_shufflevector(__ret_222, __ret_222, 1, 0); \
44550  __ret_222; \
44551 })
44552 #endif
44553 
44554 #ifdef __LITTLE_ENDIAN__
44555 #define vqrdmlah_lane_s16(__p0_223, __p1_223, __p2_223, __p3_223) __extension__ ({ \
44556  int16x4_t __s0_223 = __p0_223; \
44557  int16x4_t __s1_223 = __p1_223; \
44558  int16x4_t __s2_223 = __p2_223; \
44559  int16x4_t __ret_223; \
44560  __ret_223 = vqadd_s16(__s0_223, vqrdmulh_s16(__s1_223, splat_lane_s16(__s2_223, __p3_223))); \
44561  __ret_223; \
44562 })
44563 #else
44564 #define vqrdmlah_lane_s16(__p0_224, __p1_224, __p2_224, __p3_224) __extension__ ({ \
44565  int16x4_t __s0_224 = __p0_224; \
44566  int16x4_t __s1_224 = __p1_224; \
44567  int16x4_t __s2_224 = __p2_224; \
44568  int16x4_t __rev0_224; __rev0_224 = __builtin_shufflevector(__s0_224, __s0_224, 3, 2, 1, 0); \
44569  int16x4_t __rev1_224; __rev1_224 = __builtin_shufflevector(__s1_224, __s1_224, 3, 2, 1, 0); \
44570  int16x4_t __rev2_224; __rev2_224 = __builtin_shufflevector(__s2_224, __s2_224, 3, 2, 1, 0); \
44571  int16x4_t __ret_224; \
44572  __ret_224 = __noswap_vqadd_s16(__rev0_224, __noswap_vqrdmulh_s16(__rev1_224, __noswap_splat_lane_s16(__rev2_224, __p3_224))); \
44573  __ret_224 = __builtin_shufflevector(__ret_224, __ret_224, 3, 2, 1, 0); \
44574  __ret_224; \
44575 })
44576 #endif
44577 
44578 #ifdef __LITTLE_ENDIAN__
44579 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
44580  int32x4_t __ret;
44581  __ret = vqsubq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
44582  return __ret;
44583 }
44584 #else
44585 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
44586  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44587  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44588  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
44589  int32x4_t __ret;
44590  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
44591  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44592  return __ret;
44593 }
44594 #endif
44595 
44596 #ifdef __LITTLE_ENDIAN__
44597 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
44598  int16x8_t __ret;
44599  __ret = vqsubq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
44600  return __ret;
44601 }
44602 #else
44603 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
44604  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44605  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
44606  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
44607  int16x8_t __ret;
44608  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
44609  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44610  return __ret;
44611 }
44612 #endif
44613 
44614 #ifdef __LITTLE_ENDIAN__
44615 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
44616  int32x2_t __ret;
44617  __ret = vqsub_s32(__p0, vqrdmulh_s32(__p1, __p2));
44618  return __ret;
44619 }
44620 #else
44621 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
44622  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44623  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44624  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
44625  int32x2_t __ret;
44626  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
44627  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44628  return __ret;
44629 }
44630 #endif
44631 
44632 #ifdef __LITTLE_ENDIAN__
44633 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
44634  int16x4_t __ret;
44635  __ret = vqsub_s16(__p0, vqrdmulh_s16(__p1, __p2));
44636  return __ret;
44637 }
44638 #else
44639 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
44640  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44641  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44642  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
44643  int16x4_t __ret;
44644  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
44645  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44646  return __ret;
44647 }
44648 #endif
44649 
44650 #ifdef __LITTLE_ENDIAN__
44651 #define vqrdmlshq_lane_s32(__p0_225, __p1_225, __p2_225, __p3_225) __extension__ ({ \
44652  int32x4_t __s0_225 = __p0_225; \
44653  int32x4_t __s1_225 = __p1_225; \
44654  int32x2_t __s2_225 = __p2_225; \
44655  int32x4_t __ret_225; \
44656  __ret_225 = vqsubq_s32(__s0_225, vqrdmulhq_s32(__s1_225, splatq_lane_s32(__s2_225, __p3_225))); \
44657  __ret_225; \
44658 })
44659 #else
44660 #define vqrdmlshq_lane_s32(__p0_226, __p1_226, __p2_226, __p3_226) __extension__ ({ \
44661  int32x4_t __s0_226 = __p0_226; \
44662  int32x4_t __s1_226 = __p1_226; \
44663  int32x2_t __s2_226 = __p2_226; \
44664  int32x4_t __rev0_226; __rev0_226 = __builtin_shufflevector(__s0_226, __s0_226, 3, 2, 1, 0); \
44665  int32x4_t __rev1_226; __rev1_226 = __builtin_shufflevector(__s1_226, __s1_226, 3, 2, 1, 0); \
44666  int32x2_t __rev2_226; __rev2_226 = __builtin_shufflevector(__s2_226, __s2_226, 1, 0); \
44667  int32x4_t __ret_226; \
44668  __ret_226 = __noswap_vqsubq_s32(__rev0_226, __noswap_vqrdmulhq_s32(__rev1_226, __noswap_splatq_lane_s32(__rev2_226, __p3_226))); \
44669  __ret_226 = __builtin_shufflevector(__ret_226, __ret_226, 3, 2, 1, 0); \
44670  __ret_226; \
44671 })
44672 #endif
44673 
44674 #ifdef __LITTLE_ENDIAN__
44675 #define vqrdmlshq_lane_s16(__p0_227, __p1_227, __p2_227, __p3_227) __extension__ ({ \
44676  int16x8_t __s0_227 = __p0_227; \
44677  int16x8_t __s1_227 = __p1_227; \
44678  int16x4_t __s2_227 = __p2_227; \
44679  int16x8_t __ret_227; \
44680  __ret_227 = vqsubq_s16(__s0_227, vqrdmulhq_s16(__s1_227, splatq_lane_s16(__s2_227, __p3_227))); \
44681  __ret_227; \
44682 })
44683 #else
44684 #define vqrdmlshq_lane_s16(__p0_228, __p1_228, __p2_228, __p3_228) __extension__ ({ \
44685  int16x8_t __s0_228 = __p0_228; \
44686  int16x8_t __s1_228 = __p1_228; \
44687  int16x4_t __s2_228 = __p2_228; \
44688  int16x8_t __rev0_228; __rev0_228 = __builtin_shufflevector(__s0_228, __s0_228, 7, 6, 5, 4, 3, 2, 1, 0); \
44689  int16x8_t __rev1_228; __rev1_228 = __builtin_shufflevector(__s1_228, __s1_228, 7, 6, 5, 4, 3, 2, 1, 0); \
44690  int16x4_t __rev2_228; __rev2_228 = __builtin_shufflevector(__s2_228, __s2_228, 3, 2, 1, 0); \
44691  int16x8_t __ret_228; \
44692  __ret_228 = __noswap_vqsubq_s16(__rev0_228, __noswap_vqrdmulhq_s16(__rev1_228, __noswap_splatq_lane_s16(__rev2_228, __p3_228))); \
44693  __ret_228 = __builtin_shufflevector(__ret_228, __ret_228, 7, 6, 5, 4, 3, 2, 1, 0); \
44694  __ret_228; \
44695 })
44696 #endif
44697 
44698 #ifdef __LITTLE_ENDIAN__
44699 #define vqrdmlsh_lane_s32(__p0_229, __p1_229, __p2_229, __p3_229) __extension__ ({ \
44700  int32x2_t __s0_229 = __p0_229; \
44701  int32x2_t __s1_229 = __p1_229; \
44702  int32x2_t __s2_229 = __p2_229; \
44703  int32x2_t __ret_229; \
44704  __ret_229 = vqsub_s32(__s0_229, vqrdmulh_s32(__s1_229, splat_lane_s32(__s2_229, __p3_229))); \
44705  __ret_229; \
44706 })
44707 #else
44708 #define vqrdmlsh_lane_s32(__p0_230, __p1_230, __p2_230, __p3_230) __extension__ ({ \
44709  int32x2_t __s0_230 = __p0_230; \
44710  int32x2_t __s1_230 = __p1_230; \
44711  int32x2_t __s2_230 = __p2_230; \
44712  int32x2_t __rev0_230; __rev0_230 = __builtin_shufflevector(__s0_230, __s0_230, 1, 0); \
44713  int32x2_t __rev1_230; __rev1_230 = __builtin_shufflevector(__s1_230, __s1_230, 1, 0); \
44714  int32x2_t __rev2_230; __rev2_230 = __builtin_shufflevector(__s2_230, __s2_230, 1, 0); \
44715  int32x2_t __ret_230; \
44716  __ret_230 = __noswap_vqsub_s32(__rev0_230, __noswap_vqrdmulh_s32(__rev1_230, __noswap_splat_lane_s32(__rev2_230, __p3_230))); \
44717  __ret_230 = __builtin_shufflevector(__ret_230, __ret_230, 1, 0); \
44718  __ret_230; \
44719 })
44720 #endif
44721 
44722 #ifdef __LITTLE_ENDIAN__
44723 #define vqrdmlsh_lane_s16(__p0_231, __p1_231, __p2_231, __p3_231) __extension__ ({ \
44724  int16x4_t __s0_231 = __p0_231; \
44725  int16x4_t __s1_231 = __p1_231; \
44726  int16x4_t __s2_231 = __p2_231; \
44727  int16x4_t __ret_231; \
44728  __ret_231 = vqsub_s16(__s0_231, vqrdmulh_s16(__s1_231, splat_lane_s16(__s2_231, __p3_231))); \
44729  __ret_231; \
44730 })
44731 #else
44732 #define vqrdmlsh_lane_s16(__p0_232, __p1_232, __p2_232, __p3_232) __extension__ ({ \
44733  int16x4_t __s0_232 = __p0_232; \
44734  int16x4_t __s1_232 = __p1_232; \
44735  int16x4_t __s2_232 = __p2_232; \
44736  int16x4_t __rev0_232; __rev0_232 = __builtin_shufflevector(__s0_232, __s0_232, 3, 2, 1, 0); \
44737  int16x4_t __rev1_232; __rev1_232 = __builtin_shufflevector(__s1_232, __s1_232, 3, 2, 1, 0); \
44738  int16x4_t __rev2_232; __rev2_232 = __builtin_shufflevector(__s2_232, __s2_232, 3, 2, 1, 0); \
44739  int16x4_t __ret_232; \
44740  __ret_232 = __noswap_vqsub_s16(__rev0_232, __noswap_vqrdmulh_s16(__rev1_232, __noswap_splat_lane_s16(__rev2_232, __p3_232))); \
44741  __ret_232 = __builtin_shufflevector(__ret_232, __ret_232, 3, 2, 1, 0); \
44742  __ret_232; \
44743 })
44744 #endif
44745 
44746 #endif
44747 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
44748 #ifdef __LITTLE_ENDIAN__
44749 #define vqrdmlahq_laneq_s32(__p0_233, __p1_233, __p2_233, __p3_233) __extension__ ({ \
44750  int32x4_t __s0_233 = __p0_233; \
44751  int32x4_t __s1_233 = __p1_233; \
44752  int32x4_t __s2_233 = __p2_233; \
44753  int32x4_t __ret_233; \
44754  __ret_233 = vqaddq_s32(__s0_233, vqrdmulhq_s32(__s1_233, splatq_laneq_s32(__s2_233, __p3_233))); \
44755  __ret_233; \
44756 })
44757 #else
44758 #define vqrdmlahq_laneq_s32(__p0_234, __p1_234, __p2_234, __p3_234) __extension__ ({ \
44759  int32x4_t __s0_234 = __p0_234; \
44760  int32x4_t __s1_234 = __p1_234; \
44761  int32x4_t __s2_234 = __p2_234; \
44762  int32x4_t __rev0_234; __rev0_234 = __builtin_shufflevector(__s0_234, __s0_234, 3, 2, 1, 0); \
44763  int32x4_t __rev1_234; __rev1_234 = __builtin_shufflevector(__s1_234, __s1_234, 3, 2, 1, 0); \
44764  int32x4_t __rev2_234; __rev2_234 = __builtin_shufflevector(__s2_234, __s2_234, 3, 2, 1, 0); \
44765  int32x4_t __ret_234; \
44766  __ret_234 = __noswap_vqaddq_s32(__rev0_234, __noswap_vqrdmulhq_s32(__rev1_234, __noswap_splatq_laneq_s32(__rev2_234, __p3_234))); \
44767  __ret_234 = __builtin_shufflevector(__ret_234, __ret_234, 3, 2, 1, 0); \
44768  __ret_234; \
44769 })
44770 #endif
44771 
44772 #ifdef __LITTLE_ENDIAN__
44773 #define vqrdmlahq_laneq_s16(__p0_235, __p1_235, __p2_235, __p3_235) __extension__ ({ \
44774  int16x8_t __s0_235 = __p0_235; \
44775  int16x8_t __s1_235 = __p1_235; \
44776  int16x8_t __s2_235 = __p2_235; \
44777  int16x8_t __ret_235; \
44778  __ret_235 = vqaddq_s16(__s0_235, vqrdmulhq_s16(__s1_235, splatq_laneq_s16(__s2_235, __p3_235))); \
44779  __ret_235; \
44780 })
44781 #else
44782 #define vqrdmlahq_laneq_s16(__p0_236, __p1_236, __p2_236, __p3_236) __extension__ ({ \
44783  int16x8_t __s0_236 = __p0_236; \
44784  int16x8_t __s1_236 = __p1_236; \
44785  int16x8_t __s2_236 = __p2_236; \
44786  int16x8_t __rev0_236; __rev0_236 = __builtin_shufflevector(__s0_236, __s0_236, 7, 6, 5, 4, 3, 2, 1, 0); \
44787  int16x8_t __rev1_236; __rev1_236 = __builtin_shufflevector(__s1_236, __s1_236, 7, 6, 5, 4, 3, 2, 1, 0); \
44788  int16x8_t __rev2_236; __rev2_236 = __builtin_shufflevector(__s2_236, __s2_236, 7, 6, 5, 4, 3, 2, 1, 0); \
44789  int16x8_t __ret_236; \
44790  __ret_236 = __noswap_vqaddq_s16(__rev0_236, __noswap_vqrdmulhq_s16(__rev1_236, __noswap_splatq_laneq_s16(__rev2_236, __p3_236))); \
44791  __ret_236 = __builtin_shufflevector(__ret_236, __ret_236, 7, 6, 5, 4, 3, 2, 1, 0); \
44792  __ret_236; \
44793 })
44794 #endif
44795 
44796 #ifdef __LITTLE_ENDIAN__
44797 #define vqrdmlah_laneq_s32(__p0_237, __p1_237, __p2_237, __p3_237) __extension__ ({ \
44798  int32x2_t __s0_237 = __p0_237; \
44799  int32x2_t __s1_237 = __p1_237; \
44800  int32x4_t __s2_237 = __p2_237; \
44801  int32x2_t __ret_237; \
44802  __ret_237 = vqadd_s32(__s0_237, vqrdmulh_s32(__s1_237, splat_laneq_s32(__s2_237, __p3_237))); \
44803  __ret_237; \
44804 })
44805 #else
44806 #define vqrdmlah_laneq_s32(__p0_238, __p1_238, __p2_238, __p3_238) __extension__ ({ \
44807  int32x2_t __s0_238 = __p0_238; \
44808  int32x2_t __s1_238 = __p1_238; \
44809  int32x4_t __s2_238 = __p2_238; \
44810  int32x2_t __rev0_238; __rev0_238 = __builtin_shufflevector(__s0_238, __s0_238, 1, 0); \
44811  int32x2_t __rev1_238; __rev1_238 = __builtin_shufflevector(__s1_238, __s1_238, 1, 0); \
44812  int32x4_t __rev2_238; __rev2_238 = __builtin_shufflevector(__s2_238, __s2_238, 3, 2, 1, 0); \
44813  int32x2_t __ret_238; \
44814  __ret_238 = __noswap_vqadd_s32(__rev0_238, __noswap_vqrdmulh_s32(__rev1_238, __noswap_splat_laneq_s32(__rev2_238, __p3_238))); \
44815  __ret_238 = __builtin_shufflevector(__ret_238, __ret_238, 1, 0); \
44816  __ret_238; \
44817 })
44818 #endif
44819 
44820 #ifdef __LITTLE_ENDIAN__
44821 #define vqrdmlah_laneq_s16(__p0_239, __p1_239, __p2_239, __p3_239) __extension__ ({ \
44822  int16x4_t __s0_239 = __p0_239; \
44823  int16x4_t __s1_239 = __p1_239; \
44824  int16x8_t __s2_239 = __p2_239; \
44825  int16x4_t __ret_239; \
44826  __ret_239 = vqadd_s16(__s0_239, vqrdmulh_s16(__s1_239, splat_laneq_s16(__s2_239, __p3_239))); \
44827  __ret_239; \
44828 })
44829 #else
44830 #define vqrdmlah_laneq_s16(__p0_240, __p1_240, __p2_240, __p3_240) __extension__ ({ \
44831  int16x4_t __s0_240 = __p0_240; \
44832  int16x4_t __s1_240 = __p1_240; \
44833  int16x8_t __s2_240 = __p2_240; \
44834  int16x4_t __rev0_240; __rev0_240 = __builtin_shufflevector(__s0_240, __s0_240, 3, 2, 1, 0); \
44835  int16x4_t __rev1_240; __rev1_240 = __builtin_shufflevector(__s1_240, __s1_240, 3, 2, 1, 0); \
44836  int16x8_t __rev2_240; __rev2_240 = __builtin_shufflevector(__s2_240, __s2_240, 7, 6, 5, 4, 3, 2, 1, 0); \
44837  int16x4_t __ret_240; \
44838  __ret_240 = __noswap_vqadd_s16(__rev0_240, __noswap_vqrdmulh_s16(__rev1_240, __noswap_splat_laneq_s16(__rev2_240, __p3_240))); \
44839  __ret_240 = __builtin_shufflevector(__ret_240, __ret_240, 3, 2, 1, 0); \
44840  __ret_240; \
44841 })
44842 #endif
44843 
44844 #ifdef __LITTLE_ENDIAN__
44845 #define vqrdmlshq_laneq_s32(__p0_241, __p1_241, __p2_241, __p3_241) __extension__ ({ \
44846  int32x4_t __s0_241 = __p0_241; \
44847  int32x4_t __s1_241 = __p1_241; \
44848  int32x4_t __s2_241 = __p2_241; \
44849  int32x4_t __ret_241; \
44850  __ret_241 = vqsubq_s32(__s0_241, vqrdmulhq_s32(__s1_241, splatq_laneq_s32(__s2_241, __p3_241))); \
44851  __ret_241; \
44852 })
44853 #else
44854 #define vqrdmlshq_laneq_s32(__p0_242, __p1_242, __p2_242, __p3_242) __extension__ ({ \
44855  int32x4_t __s0_242 = __p0_242; \
44856  int32x4_t __s1_242 = __p1_242; \
44857  int32x4_t __s2_242 = __p2_242; \
44858  int32x4_t __rev0_242; __rev0_242 = __builtin_shufflevector(__s0_242, __s0_242, 3, 2, 1, 0); \
44859  int32x4_t __rev1_242; __rev1_242 = __builtin_shufflevector(__s1_242, __s1_242, 3, 2, 1, 0); \
44860  int32x4_t __rev2_242; __rev2_242 = __builtin_shufflevector(__s2_242, __s2_242, 3, 2, 1, 0); \
44861  int32x4_t __ret_242; \
44862  __ret_242 = __noswap_vqsubq_s32(__rev0_242, __noswap_vqrdmulhq_s32(__rev1_242, __noswap_splatq_laneq_s32(__rev2_242, __p3_242))); \
44863  __ret_242 = __builtin_shufflevector(__ret_242, __ret_242, 3, 2, 1, 0); \
44864  __ret_242; \
44865 })
44866 #endif
44867 
44868 #ifdef __LITTLE_ENDIAN__
44869 #define vqrdmlshq_laneq_s16(__p0_243, __p1_243, __p2_243, __p3_243) __extension__ ({ \
44870  int16x8_t __s0_243 = __p0_243; \
44871  int16x8_t __s1_243 = __p1_243; \
44872  int16x8_t __s2_243 = __p2_243; \
44873  int16x8_t __ret_243; \
44874  __ret_243 = vqsubq_s16(__s0_243, vqrdmulhq_s16(__s1_243, splatq_laneq_s16(__s2_243, __p3_243))); \
44875  __ret_243; \
44876 })
44877 #else
44878 #define vqrdmlshq_laneq_s16(__p0_244, __p1_244, __p2_244, __p3_244) __extension__ ({ \
44879  int16x8_t __s0_244 = __p0_244; \
44880  int16x8_t __s1_244 = __p1_244; \
44881  int16x8_t __s2_244 = __p2_244; \
44882  int16x8_t __rev0_244; __rev0_244 = __builtin_shufflevector(__s0_244, __s0_244, 7, 6, 5, 4, 3, 2, 1, 0); \
44883  int16x8_t __rev1_244; __rev1_244 = __builtin_shufflevector(__s1_244, __s1_244, 7, 6, 5, 4, 3, 2, 1, 0); \
44884  int16x8_t __rev2_244; __rev2_244 = __builtin_shufflevector(__s2_244, __s2_244, 7, 6, 5, 4, 3, 2, 1, 0); \
44885  int16x8_t __ret_244; \
44886  __ret_244 = __noswap_vqsubq_s16(__rev0_244, __noswap_vqrdmulhq_s16(__rev1_244, __noswap_splatq_laneq_s16(__rev2_244, __p3_244))); \
44887  __ret_244 = __builtin_shufflevector(__ret_244, __ret_244, 7, 6, 5, 4, 3, 2, 1, 0); \
44888  __ret_244; \
44889 })
44890 #endif
44891 
44892 #ifdef __LITTLE_ENDIAN__
44893 #define vqrdmlsh_laneq_s32(__p0_245, __p1_245, __p2_245, __p3_245) __extension__ ({ \
44894  int32x2_t __s0_245 = __p0_245; \
44895  int32x2_t __s1_245 = __p1_245; \
44896  int32x4_t __s2_245 = __p2_245; \
44897  int32x2_t __ret_245; \
44898  __ret_245 = vqsub_s32(__s0_245, vqrdmulh_s32(__s1_245, splat_laneq_s32(__s2_245, __p3_245))); \
44899  __ret_245; \
44900 })
44901 #else
44902 #define vqrdmlsh_laneq_s32(__p0_246, __p1_246, __p2_246, __p3_246) __extension__ ({ \
44903  int32x2_t __s0_246 = __p0_246; \
44904  int32x2_t __s1_246 = __p1_246; \
44905  int32x4_t __s2_246 = __p2_246; \
44906  int32x2_t __rev0_246; __rev0_246 = __builtin_shufflevector(__s0_246, __s0_246, 1, 0); \
44907  int32x2_t __rev1_246; __rev1_246 = __builtin_shufflevector(__s1_246, __s1_246, 1, 0); \
44908  int32x4_t __rev2_246; __rev2_246 = __builtin_shufflevector(__s2_246, __s2_246, 3, 2, 1, 0); \
44909  int32x2_t __ret_246; \
44910  __ret_246 = __noswap_vqsub_s32(__rev0_246, __noswap_vqrdmulh_s32(__rev1_246, __noswap_splat_laneq_s32(__rev2_246, __p3_246))); \
44911  __ret_246 = __builtin_shufflevector(__ret_246, __ret_246, 1, 0); \
44912  __ret_246; \
44913 })
44914 #endif
44915 
44916 #ifdef __LITTLE_ENDIAN__
44917 #define vqrdmlsh_laneq_s16(__p0_247, __p1_247, __p2_247, __p3_247) __extension__ ({ \
44918  int16x4_t __s0_247 = __p0_247; \
44919  int16x4_t __s1_247 = __p1_247; \
44920  int16x8_t __s2_247 = __p2_247; \
44921  int16x4_t __ret_247; \
44922  __ret_247 = vqsub_s16(__s0_247, vqrdmulh_s16(__s1_247, splat_laneq_s16(__s2_247, __p3_247))); \
44923  __ret_247; \
44924 })
44925 #else
44926 #define vqrdmlsh_laneq_s16(__p0_248, __p1_248, __p2_248, __p3_248) __extension__ ({ \
44927  int16x4_t __s0_248 = __p0_248; \
44928  int16x4_t __s1_248 = __p1_248; \
44929  int16x8_t __s2_248 = __p2_248; \
44930  int16x4_t __rev0_248; __rev0_248 = __builtin_shufflevector(__s0_248, __s0_248, 3, 2, 1, 0); \
44931  int16x4_t __rev1_248; __rev1_248 = __builtin_shufflevector(__s1_248, __s1_248, 3, 2, 1, 0); \
44932  int16x8_t __rev2_248; __rev2_248 = __builtin_shufflevector(__s2_248, __s2_248, 7, 6, 5, 4, 3, 2, 1, 0); \
44933  int16x4_t __ret_248; \
44934  __ret_248 = __noswap_vqsub_s16(__rev0_248, __noswap_vqrdmulh_s16(__rev1_248, __noswap_splat_laneq_s16(__rev2_248, __p3_248))); \
44935  __ret_248 = __builtin_shufflevector(__ret_248, __ret_248, 3, 2, 1, 0); \
44936  __ret_248; \
44937 })
44938 #endif
44939 
44940 #endif
44941 #if defined(__aarch64__)
44942 #ifdef __LITTLE_ENDIAN__
44943 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
44944  float64x2_t __ret;
44945  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
44946  return __ret;
44947 }
44948 #else
44949 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
44950  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44951  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44952  float64x2_t __ret;
44953  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
44954  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44955  return __ret;
44956 }
44957 #endif
44958 
44959 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
44960  float64x1_t __ret;
44961  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
44962  return __ret;
44963 }
44964 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
44965  float64_t __ret;
44966  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
44967  return __ret;
44968 }
44969 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
44970  float32_t __ret;
44971  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
44972  return __ret;
44973 }
44974 #ifdef __LITTLE_ENDIAN__
44975 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
44976  float64x2_t __ret;
44977  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
44978  return __ret;
44979 }
44980 #else
44981 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
44982  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44983  float64x2_t __ret;
44984  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
44985  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44986  return __ret;
44987 }
44988 #endif
44989 
44990 #ifdef __LITTLE_ENDIAN__
44991 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
44992  int64x2_t __ret;
44993  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
44994  return __ret;
44995 }
44996 #else
44997 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
44998  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44999  int64x2_t __ret;
45000  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
45001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45002  return __ret;
45003 }
45004 #endif
45005 
45006 __ai float64x1_t vabs_f64(float64x1_t __p0) {
45007  float64x1_t __ret;
45008  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
45009  return __ret;
45010 }
45011 __ai int64x1_t vabs_s64(int64x1_t __p0) {
45012  int64x1_t __ret;
45013  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
45014  return __ret;
45015 }
45016 __ai int64_t vabsd_s64(int64_t __p0) {
45017  int64_t __ret;
45018  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
45019  return __ret;
45020 }
45021 #ifdef __LITTLE_ENDIAN__
45022 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
45023  float64x2_t __ret;
45024  __ret = __p0 + __p1;
45025  return __ret;
45026 }
45027 #else
45028 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
45029  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45030  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45031  float64x2_t __ret;
45032  __ret = __rev0 + __rev1;
45033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45034  return __ret;
45035 }
45036 #endif
45037 
45038 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
45039  float64x1_t __ret;
45040  __ret = __p0 + __p1;
45041  return __ret;
45042 }
45043 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
45044  uint64_t __ret;
45045  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
45046  return __ret;
45047 }
45048 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
45049  int64_t __ret;
45050  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
45051  return __ret;
45052 }
45053 #ifdef __LITTLE_ENDIAN__
45054 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
45055  uint16x8_t __ret;
45056  __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
45057  return __ret;
45058 }
45059 #else
45060 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
45061  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45062  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45063  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
45064  uint16x8_t __ret;
45065  __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
45066  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45067  return __ret;
45068 }
45069 #endif
45070 
45071 #ifdef __LITTLE_ENDIAN__
45072 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
45073  uint32x4_t __ret;
45074  __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
45075  return __ret;
45076 }
45077 #else
45078 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
45079  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45080  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45081  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
45082  uint32x4_t __ret;
45083  __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
45084  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45085  return __ret;
45086 }
45087 #endif
45088 
45089 #ifdef __LITTLE_ENDIAN__
45090 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
45091  uint8x16_t __ret;
45092  __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
45093  return __ret;
45094 }
45095 #else
45096 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
45097  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45098  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
45099  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
45100  uint8x16_t __ret;
45101  __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
45102  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45103  return __ret;
45104 }
45105 #endif
45106 
45107 #ifdef __LITTLE_ENDIAN__
45108 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
45109  int16x8_t __ret;
45110  __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
45111  return __ret;
45112 }
45113 #else
45114 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
45115  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45116  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45117  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
45118  int16x8_t __ret;
45119  __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
45120  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45121  return __ret;
45122 }
45123 #endif
45124 
45125 #ifdef __LITTLE_ENDIAN__
45126 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
45127  int32x4_t __ret;
45128  __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
45129  return __ret;
45130 }
45131 #else
45132 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
45133  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45134  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45135  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
45136  int32x4_t __ret;
45137  __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
45138  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45139  return __ret;
45140 }
45141 #endif
45142 
45143 #ifdef __LITTLE_ENDIAN__
45144 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
45145  int8x16_t __ret;
45146  __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
45147  return __ret;
45148 }
45149 #else
45150 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
45151  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45152  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
45153  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
45154  int8x16_t __ret;
45155  __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
45156  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45157  return __ret;
45158 }
45159 #endif
45160 
45161 #ifdef __LITTLE_ENDIAN__
45162 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
45163  uint16_t __ret;
45164  __ret = (uint16_t) __builtin_neon_vaddlvq_u8(__p0);
45165  return __ret;
45166 }
45167 #else
45168 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
45169  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45170  uint16_t __ret;
45171  __ret = (uint16_t) __builtin_neon_vaddlvq_u8(__rev0);
45172  return __ret;
45173 }
45174 #endif
45175 
45176 #ifdef __LITTLE_ENDIAN__
45177 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
45178  uint64_t __ret;
45179  __ret = (uint64_t) __builtin_neon_vaddlvq_u32(__p0);
45180  return __ret;
45181 }
45182 #else
45183 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
45184  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45185  uint64_t __ret;
45186  __ret = (uint64_t) __builtin_neon_vaddlvq_u32(__rev0);
45187  return __ret;
45188 }
45189 #endif
45190 
45191 #ifdef __LITTLE_ENDIAN__
45192 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
45193  uint32_t __ret;
45194  __ret = (uint32_t) __builtin_neon_vaddlvq_u16(__p0);
45195  return __ret;
45196 }
45197 #else
45198 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
45199  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45200  uint32_t __ret;
45201  __ret = (uint32_t) __builtin_neon_vaddlvq_u16(__rev0);
45202  return __ret;
45203 }
45204 #endif
45205 
45206 #ifdef __LITTLE_ENDIAN__
45207 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
45208  int16_t __ret;
45209  __ret = (int16_t) __builtin_neon_vaddlvq_s8(__p0);
45210  return __ret;
45211 }
45212 #else
45213 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
45214  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45215  int16_t __ret;
45216  __ret = (int16_t) __builtin_neon_vaddlvq_s8(__rev0);
45217  return __ret;
45218 }
45219 #endif
45220 
45221 #ifdef __LITTLE_ENDIAN__
45222 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
45223  int64_t __ret;
45224  __ret = (int64_t) __builtin_neon_vaddlvq_s32(__p0);
45225  return __ret;
45226 }
45227 #else
45228 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
45229  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45230  int64_t __ret;
45231  __ret = (int64_t) __builtin_neon_vaddlvq_s32(__rev0);
45232  return __ret;
45233 }
45234 #endif
45235 
45236 #ifdef __LITTLE_ENDIAN__
45237 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
45238  int32_t __ret;
45239  __ret = (int32_t) __builtin_neon_vaddlvq_s16(__p0);
45240  return __ret;
45241 }
45242 #else
45243 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
45244  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45245  int32_t __ret;
45246  __ret = (int32_t) __builtin_neon_vaddlvq_s16(__rev0);
45247  return __ret;
45248 }
45249 #endif
45250 
45251 #ifdef __LITTLE_ENDIAN__
45252 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
45253  uint16_t __ret;
45254  __ret = (uint16_t) __builtin_neon_vaddlv_u8(__p0);
45255  return __ret;
45256 }
45257 #else
45258 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
45259  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45260  uint16_t __ret;
45261  __ret = (uint16_t) __builtin_neon_vaddlv_u8(__rev0);
45262  return __ret;
45263 }
45264 #endif
45265 
45266 #ifdef __LITTLE_ENDIAN__
45267 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
45268  uint64_t __ret;
45269  __ret = (uint64_t) __builtin_neon_vaddlv_u32(__p0);
45270  return __ret;
45271 }
45272 #else
45273 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
45274  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45275  uint64_t __ret;
45276  __ret = (uint64_t) __builtin_neon_vaddlv_u32(__rev0);
45277  return __ret;
45278 }
45279 #endif
45280 
45281 #ifdef __LITTLE_ENDIAN__
45282 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
45283  uint32_t __ret;
45284  __ret = (uint32_t) __builtin_neon_vaddlv_u16(__p0);
45285  return __ret;
45286 }
45287 #else
45288 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
45289  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45290  uint32_t __ret;
45291  __ret = (uint32_t) __builtin_neon_vaddlv_u16(__rev0);
45292  return __ret;
45293 }
45294 #endif
45295 
45296 #ifdef __LITTLE_ENDIAN__
45297 __ai int16_t vaddlv_s8(int8x8_t __p0) {
45298  int16_t __ret;
45299  __ret = (int16_t) __builtin_neon_vaddlv_s8(__p0);
45300  return __ret;
45301 }
45302 #else
45303 __ai int16_t vaddlv_s8(int8x8_t __p0) {
45304  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45305  int16_t __ret;
45306  __ret = (int16_t) __builtin_neon_vaddlv_s8(__rev0);
45307  return __ret;
45308 }
45309 #endif
45310 
45311 #ifdef __LITTLE_ENDIAN__
45312 __ai int64_t vaddlv_s32(int32x2_t __p0) {
45313  int64_t __ret;
45314  __ret = (int64_t) __builtin_neon_vaddlv_s32(__p0);
45315  return __ret;
45316 }
45317 #else
45318 __ai int64_t vaddlv_s32(int32x2_t __p0) {
45319  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45320  int64_t __ret;
45321  __ret = (int64_t) __builtin_neon_vaddlv_s32(__rev0);
45322  return __ret;
45323 }
45324 #endif
45325 
45326 #ifdef __LITTLE_ENDIAN__
45327 __ai int32_t vaddlv_s16(int16x4_t __p0) {
45328  int32_t __ret;
45329  __ret = (int32_t) __builtin_neon_vaddlv_s16(__p0);
45330  return __ret;
45331 }
45332 #else
45333 __ai int32_t vaddlv_s16(int16x4_t __p0) {
45334  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45335  int32_t __ret;
45336  __ret = (int32_t) __builtin_neon_vaddlv_s16(__rev0);
45337  return __ret;
45338 }
45339 #endif
45340 
45341 #ifdef __LITTLE_ENDIAN__
45342 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
45343  uint8_t __ret;
45344  __ret = (uint8_t) __builtin_neon_vaddvq_u8(__p0);
45345  return __ret;
45346 }
45347 #else
45348 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
45349  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45350  uint8_t __ret;
45351  __ret = (uint8_t) __builtin_neon_vaddvq_u8(__rev0);
45352  return __ret;
45353 }
45354 #endif
45355 
45356 #ifdef __LITTLE_ENDIAN__
45357 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
45358  uint32_t __ret;
45359  __ret = (uint32_t) __builtin_neon_vaddvq_u32(__p0);
45360  return __ret;
45361 }
45362 #else
45363 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
45364  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45365  uint32_t __ret;
45366  __ret = (uint32_t) __builtin_neon_vaddvq_u32(__rev0);
45367  return __ret;
45368 }
45369 #endif
45370 
45371 #ifdef __LITTLE_ENDIAN__
45372 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
45373  uint64_t __ret;
45374  __ret = (uint64_t) __builtin_neon_vaddvq_u64(__p0);
45375  return __ret;
45376 }
45377 #else
45378 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
45379  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45380  uint64_t __ret;
45381  __ret = (uint64_t) __builtin_neon_vaddvq_u64(__rev0);
45382  return __ret;
45383 }
45384 #endif
45385 
45386 #ifdef __LITTLE_ENDIAN__
45387 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
45388  uint16_t __ret;
45389  __ret = (uint16_t) __builtin_neon_vaddvq_u16(__p0);
45390  return __ret;
45391 }
45392 #else
45393 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
45394  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45395  uint16_t __ret;
45396  __ret = (uint16_t) __builtin_neon_vaddvq_u16(__rev0);
45397  return __ret;
45398 }
45399 #endif
45400 
45401 #ifdef __LITTLE_ENDIAN__
45402 __ai int8_t vaddvq_s8(int8x16_t __p0) {
45403  int8_t __ret;
45404  __ret = (int8_t) __builtin_neon_vaddvq_s8(__p0);
45405  return __ret;
45406 }
45407 #else
45408 __ai int8_t vaddvq_s8(int8x16_t __p0) {
45409  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45410  int8_t __ret;
45411  __ret = (int8_t) __builtin_neon_vaddvq_s8(__rev0);
45412  return __ret;
45413 }
45414 #endif
45415 
45416 #ifdef __LITTLE_ENDIAN__
45417 __ai float64_t vaddvq_f64(float64x2_t __p0) {
45418  float64_t __ret;
45419  __ret = (float64_t) __builtin_neon_vaddvq_f64(__p0);
45420  return __ret;
45421 }
45422 #else
45423 __ai float64_t vaddvq_f64(float64x2_t __p0) {
45424  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45425  float64_t __ret;
45426  __ret = (float64_t) __builtin_neon_vaddvq_f64(__rev0);
45427  return __ret;
45428 }
45429 #endif
45430 
45431 #ifdef __LITTLE_ENDIAN__
45432 __ai float32_t vaddvq_f32(float32x4_t __p0) {
45433  float32_t __ret;
45434  __ret = (float32_t) __builtin_neon_vaddvq_f32(__p0);
45435  return __ret;
45436 }
45437 #else
45438 __ai float32_t vaddvq_f32(float32x4_t __p0) {
45439  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45440  float32_t __ret;
45441  __ret = (float32_t) __builtin_neon_vaddvq_f32(__rev0);
45442  return __ret;
45443 }
45444 #endif
45445 
45446 #ifdef __LITTLE_ENDIAN__
45447 __ai int32_t vaddvq_s32(int32x4_t __p0) {
45448  int32_t __ret;
45449  __ret = (int32_t) __builtin_neon_vaddvq_s32(__p0);
45450  return __ret;
45451 }
45452 #else
45453 __ai int32_t vaddvq_s32(int32x4_t __p0) {
45454  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45455  int32_t __ret;
45456  __ret = (int32_t) __builtin_neon_vaddvq_s32(__rev0);
45457  return __ret;
45458 }
45459 #endif
45460 
45461 #ifdef __LITTLE_ENDIAN__
45462 __ai int64_t vaddvq_s64(int64x2_t __p0) {
45463  int64_t __ret;
45464  __ret = (int64_t) __builtin_neon_vaddvq_s64(__p0);
45465  return __ret;
45466 }
45467 #else
45468 __ai int64_t vaddvq_s64(int64x2_t __p0) {
45469  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45470  int64_t __ret;
45471  __ret = (int64_t) __builtin_neon_vaddvq_s64(__rev0);
45472  return __ret;
45473 }
45474 #endif
45475 
45476 #ifdef __LITTLE_ENDIAN__
45477 __ai int16_t vaddvq_s16(int16x8_t __p0) {
45478  int16_t __ret;
45479  __ret = (int16_t) __builtin_neon_vaddvq_s16(__p0);
45480  return __ret;
45481 }
45482 #else
45483 __ai int16_t vaddvq_s16(int16x8_t __p0) {
45484  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45485  int16_t __ret;
45486  __ret = (int16_t) __builtin_neon_vaddvq_s16(__rev0);
45487  return __ret;
45488 }
45489 #endif
45490 
45491 #ifdef __LITTLE_ENDIAN__
45492 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
45493  uint8_t __ret;
45494  __ret = (uint8_t) __builtin_neon_vaddv_u8(__p0);
45495  return __ret;
45496 }
45497 #else
45498 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
45499  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45500  uint8_t __ret;
45501  __ret = (uint8_t) __builtin_neon_vaddv_u8(__rev0);
45502  return __ret;
45503 }
45504 #endif
45505 
45506 #ifdef __LITTLE_ENDIAN__
45507 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
45508  uint32_t __ret;
45509  __ret = (uint32_t) __builtin_neon_vaddv_u32(__p0);
45510  return __ret;
45511 }
45512 #else
45513 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
45514  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45515  uint32_t __ret;
45516  __ret = (uint32_t) __builtin_neon_vaddv_u32(__rev0);
45517  return __ret;
45518 }
45519 #endif
45520 
45521 #ifdef __LITTLE_ENDIAN__
45522 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
45523  uint16_t __ret;
45524  __ret = (uint16_t) __builtin_neon_vaddv_u16(__p0);
45525  return __ret;
45526 }
45527 #else
45528 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
45529  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45530  uint16_t __ret;
45531  __ret = (uint16_t) __builtin_neon_vaddv_u16(__rev0);
45532  return __ret;
45533 }
45534 #endif
45535 
45536 #ifdef __LITTLE_ENDIAN__
45537 __ai int8_t vaddv_s8(int8x8_t __p0) {
45538  int8_t __ret;
45539  __ret = (int8_t) __builtin_neon_vaddv_s8(__p0);
45540  return __ret;
45541 }
45542 #else
45543 __ai int8_t vaddv_s8(int8x8_t __p0) {
45544  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45545  int8_t __ret;
45546  __ret = (int8_t) __builtin_neon_vaddv_s8(__rev0);
45547  return __ret;
45548 }
45549 #endif
45550 
45551 #ifdef __LITTLE_ENDIAN__
45552 __ai float32_t vaddv_f32(float32x2_t __p0) {
45553  float32_t __ret;
45554  __ret = (float32_t) __builtin_neon_vaddv_f32(__p0);
45555  return __ret;
45556 }
45557 #else
45558 __ai float32_t vaddv_f32(float32x2_t __p0) {
45559  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45560  float32_t __ret;
45561  __ret = (float32_t) __builtin_neon_vaddv_f32(__rev0);
45562  return __ret;
45563 }
45564 #endif
45565 
45566 #ifdef __LITTLE_ENDIAN__
45567 __ai int32_t vaddv_s32(int32x2_t __p0) {
45568  int32_t __ret;
45569  __ret = (int32_t) __builtin_neon_vaddv_s32(__p0);
45570  return __ret;
45571 }
45572 #else
45573 __ai int32_t vaddv_s32(int32x2_t __p0) {
45574  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45575  int32_t __ret;
45576  __ret = (int32_t) __builtin_neon_vaddv_s32(__rev0);
45577  return __ret;
45578 }
45579 #endif
45580 
45581 #ifdef __LITTLE_ENDIAN__
45582 __ai int16_t vaddv_s16(int16x4_t __p0) {
45583  int16_t __ret;
45584  __ret = (int16_t) __builtin_neon_vaddv_s16(__p0);
45585  return __ret;
45586 }
45587 #else
45588 __ai int16_t vaddv_s16(int16x4_t __p0) {
45589  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45590  int16_t __ret;
45591  __ret = (int16_t) __builtin_neon_vaddv_s16(__rev0);
45592  return __ret;
45593 }
45594 #endif
45595 
45596 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
45597  poly64x1_t __ret;
45598  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
45599  return __ret;
45600 }
45601 #ifdef __LITTLE_ENDIAN__
45602 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
45603  poly64x2_t __ret;
45604  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
45605  return __ret;
45606 }
45607 #else
45608 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
45609  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45610  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45611  poly64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
45612  poly64x2_t __ret;
45613  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
45614  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45615  return __ret;
45616 }
45617 #endif
45618 
45619 #ifdef __LITTLE_ENDIAN__
45620 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
45621  float64x2_t __ret;
45622  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
45623  return __ret;
45624 }
45625 #else
45626 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
45627  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45628  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45629  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
45630  float64x2_t __ret;
45631  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
45632  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45633  return __ret;
45634 }
45635 #endif
45636 
45637 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
45638  float64x1_t __ret;
45639  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
45640  return __ret;
45641 }
45642 #ifdef __LITTLE_ENDIAN__
45643 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
45644  uint64x2_t __ret;
45645  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
45646  return __ret;
45647 }
45648 #else
45649 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
45650  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45651  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45652  uint64x2_t __ret;
45653  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
45654  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45655  return __ret;
45656 }
45657 #endif
45658 
45659 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
45660  uint64x1_t __ret;
45661  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
45662  return __ret;
45663 }
45664 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
45665  uint64_t __ret;
45666  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
45667  return __ret;
45668 }
45669 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
45670  uint32_t __ret;
45671  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
45672  return __ret;
45673 }
45674 #ifdef __LITTLE_ENDIAN__
45675 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
45676  uint64x2_t __ret;
45677  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
45678  return __ret;
45679 }
45680 #else
45681 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
45682  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45683  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45684  uint64x2_t __ret;
45685  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
45686  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45687  return __ret;
45688 }
45689 #endif
45690 
45691 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
45692  uint64x1_t __ret;
45693  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
45694  return __ret;
45695 }
45696 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
45697  uint64_t __ret;
45698  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
45699  return __ret;
45700 }
45701 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
45702  uint32_t __ret;
45703  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
45704  return __ret;
45705 }
45706 #ifdef __LITTLE_ENDIAN__
45707 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
45708  uint64x2_t __ret;
45709  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
45710  return __ret;
45711 }
45712 #else
45713 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
45714  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45715  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45716  uint64x2_t __ret;
45717  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
45718  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45719  return __ret;
45720 }
45721 #endif
45722 
45723 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
45724  uint64x1_t __ret;
45725  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
45726  return __ret;
45727 }
45728 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
45729  uint64_t __ret;
45730  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
45731  return __ret;
45732 }
45733 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
45734  uint32_t __ret;
45735  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
45736  return __ret;
45737 }
45738 #ifdef __LITTLE_ENDIAN__
45739 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
45740  uint64x2_t __ret;
45741  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
45742  return __ret;
45743 }
45744 #else
45745 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
45746  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45747  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45748  uint64x2_t __ret;
45749  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
45750  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45751  return __ret;
45752 }
45753 #endif
45754 
45755 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
45756  uint64x1_t __ret;
45757  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
45758  return __ret;
45759 }
45760 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
45761  uint64_t __ret;
45762  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
45763  return __ret;
45764 }
45765 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
45766  uint32_t __ret;
45767  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
45768  return __ret;
45769 }
45770 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
45771  uint64x1_t __ret;
45772  __ret = (uint64x1_t)(__p0 == __p1);
45773  return __ret;
45774 }
45775 #ifdef __LITTLE_ENDIAN__
45776 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
45777  uint64x2_t __ret;
45778  __ret = (uint64x2_t)(__p0 == __p1);
45779  return __ret;
45780 }
45781 #else
45782 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
45783  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45784  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45785  uint64x2_t __ret;
45786  __ret = (uint64x2_t)(__rev0 == __rev1);
45787  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45788  return __ret;
45789 }
45790 #endif
45791 
45792 #ifdef __LITTLE_ENDIAN__
45793 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
45794  uint64x2_t __ret;
45795  __ret = (uint64x2_t)(__p0 == __p1);
45796  return __ret;
45797 }
45798 #else
45799 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
45800  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45801  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45802  uint64x2_t __ret;
45803  __ret = (uint64x2_t)(__rev0 == __rev1);
45804  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45805  return __ret;
45806 }
45807 #endif
45808 
45809 #ifdef __LITTLE_ENDIAN__
45810 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
45811  uint64x2_t __ret;
45812  __ret = (uint64x2_t)(__p0 == __p1);
45813  return __ret;
45814 }
45815 #else
45816 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
45817  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45818  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45819  uint64x2_t __ret;
45820  __ret = (uint64x2_t)(__rev0 == __rev1);
45821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45822  return __ret;
45823 }
45824 #endif
45825 
45826 #ifdef __LITTLE_ENDIAN__
45827 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
45828  uint64x2_t __ret;
45829  __ret = (uint64x2_t)(__p0 == __p1);
45830  return __ret;
45831 }
45832 #else
45833 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
45834  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45835  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45836  uint64x2_t __ret;
45837  __ret = (uint64x2_t)(__rev0 == __rev1);
45838  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45839  return __ret;
45840 }
45841 #endif
45842 
45843 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
45844  uint64x1_t __ret;
45845  __ret = (uint64x1_t)(__p0 == __p1);
45846  return __ret;
45847 }
45848 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
45849  uint64x1_t __ret;
45850  __ret = (uint64x1_t)(__p0 == __p1);
45851  return __ret;
45852 }
45853 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
45854  uint64x1_t __ret;
45855  __ret = (uint64x1_t)(__p0 == __p1);
45856  return __ret;
45857 }
45858 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
45859  uint64_t __ret;
45860  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
45861  return __ret;
45862 }
45863 __ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
45864  int64_t __ret;
45865  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
45866  return __ret;
45867 }
45868 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
45869  uint64_t __ret;
45870  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
45871  return __ret;
45872 }
45873 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
45874  uint32_t __ret;
45875  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
45876  return __ret;
45877 }
45878 #ifdef __LITTLE_ENDIAN__
45879 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
45880  uint8x8_t __ret;
45881  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
45882  return __ret;
45883 }
45884 #else
45885 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
45886  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45887  uint8x8_t __ret;
45888  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
45889  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45890  return __ret;
45891 }
45892 #endif
45893 
45894 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
45895  uint64x1_t __ret;
45896  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
45897  return __ret;
45898 }
45899 #ifdef __LITTLE_ENDIAN__
45900 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
45901  uint16x4_t __ret;
45902  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
45903  return __ret;
45904 }
45905 #else
45906 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
45907  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45908  uint16x4_t __ret;
45909  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
45910  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45911  return __ret;
45912 }
45913 #endif
45914 
45915 #ifdef __LITTLE_ENDIAN__
45916 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
45917  uint8x16_t __ret;
45918  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
45919  return __ret;
45920 }
45921 #else
45922 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
45923  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45924  uint8x16_t __ret;
45925  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
45926  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45927  return __ret;
45928 }
45929 #endif
45930 
45931 #ifdef __LITTLE_ENDIAN__
45932 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
45933  uint64x2_t __ret;
45934  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
45935  return __ret;
45936 }
45937 #else
45938 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
45939  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45940  uint64x2_t __ret;
45941  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
45942  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45943  return __ret;
45944 }
45945 #endif
45946 
45947 #ifdef __LITTLE_ENDIAN__
45948 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
45949  uint16x8_t __ret;
45950  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
45951  return __ret;
45952 }
45953 #else
45954 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
45955  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45956  uint16x8_t __ret;
45957  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
45958  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45959  return __ret;
45960 }
45961 #endif
45962 
45963 #ifdef __LITTLE_ENDIAN__
45964 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
45965  uint8x16_t __ret;
45966  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
45967  return __ret;
45968 }
45969 #else
45970 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
45971  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45972  uint8x16_t __ret;
45973  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
45974  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
45975  return __ret;
45976 }
45977 #endif
45978 
45979 #ifdef __LITTLE_ENDIAN__
45980 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
45981  uint32x4_t __ret;
45982  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
45983  return __ret;
45984 }
45985 #else
45986 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
45987  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45988  uint32x4_t __ret;
45989  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
45990  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45991  return __ret;
45992 }
45993 #endif
45994 
45995 #ifdef __LITTLE_ENDIAN__
45996 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
45997  uint64x2_t __ret;
45998  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
45999  return __ret;
46000 }
46001 #else
46002 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
46003  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46004  uint64x2_t __ret;
46005  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
46006  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46007  return __ret;
46008 }
46009 #endif
46010 
46011 #ifdef __LITTLE_ENDIAN__
46012 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
46013  uint16x8_t __ret;
46014  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
46015  return __ret;
46016 }
46017 #else
46018 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
46019  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46020  uint16x8_t __ret;
46021  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
46022  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46023  return __ret;
46024 }
46025 #endif
46026 
46027 #ifdef __LITTLE_ENDIAN__
46028 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
46029  uint8x16_t __ret;
46030  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
46031  return __ret;
46032 }
46033 #else
46034 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
46035  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46036  uint8x16_t __ret;
46037  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
46038  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46039  return __ret;
46040 }
46041 #endif
46042 
46043 #ifdef __LITTLE_ENDIAN__
46044 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
46045  uint64x2_t __ret;
46046  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
46047  return __ret;
46048 }
46049 #else
46050 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
46051  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46052  uint64x2_t __ret;
46053  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
46054  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46055  return __ret;
46056 }
46057 #endif
46058 
46059 #ifdef __LITTLE_ENDIAN__
46060 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
46061  uint32x4_t __ret;
46062  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
46063  return __ret;
46064 }
46065 #else
46066 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
46067  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46068  uint32x4_t __ret;
46069  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
46070  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46071  return __ret;
46072 }
46073 #endif
46074 
46075 #ifdef __LITTLE_ENDIAN__
46076 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
46077  uint32x4_t __ret;
46078  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
46079  return __ret;
46080 }
46081 #else
46082 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
46083  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46084  uint32x4_t __ret;
46085  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
46086  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46087  return __ret;
46088 }
46089 #endif
46090 
46091 #ifdef __LITTLE_ENDIAN__
46092 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
46093  uint64x2_t __ret;
46094  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
46095  return __ret;
46096 }
46097 #else
46098 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
46099  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46100  uint64x2_t __ret;
46101  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
46102  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46103  return __ret;
46104 }
46105 #endif
46106 
46107 #ifdef __LITTLE_ENDIAN__
46108 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
46109  uint16x8_t __ret;
46110  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
46111  return __ret;
46112 }
46113 #else
46114 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
46115  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46116  uint16x8_t __ret;
46117  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
46118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46119  return __ret;
46120 }
46121 #endif
46122 
46123 #ifdef __LITTLE_ENDIAN__
46124 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
46125  uint8x8_t __ret;
46126  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
46127  return __ret;
46128 }
46129 #else
46130 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
46131  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46132  uint8x8_t __ret;
46133  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
46134  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46135  return __ret;
46136 }
46137 #endif
46138 
46139 #ifdef __LITTLE_ENDIAN__
46140 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
46141  uint32x2_t __ret;
46142  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
46143  return __ret;
46144 }
46145 #else
46146 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
46147  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46148  uint32x2_t __ret;
46149  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
46150  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46151  return __ret;
46152 }
46153 #endif
46154 
46155 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
46156  uint64x1_t __ret;
46157  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
46158  return __ret;
46159 }
46160 #ifdef __LITTLE_ENDIAN__
46161 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
46162  uint16x4_t __ret;
46163  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
46164  return __ret;
46165 }
46166 #else
46167 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
46168  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46169  uint16x4_t __ret;
46170  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
46171  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46172  return __ret;
46173 }
46174 #endif
46175 
46176 #ifdef __LITTLE_ENDIAN__
46177 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
46178  uint8x8_t __ret;
46179  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
46180  return __ret;
46181 }
46182 #else
46183 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
46184  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46185  uint8x8_t __ret;
46186  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
46187  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46188  return __ret;
46189 }
46190 #endif
46191 
46192 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
46193  uint64x1_t __ret;
46194  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
46195  return __ret;
46196 }
46197 #ifdef __LITTLE_ENDIAN__
46198 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
46199  uint32x2_t __ret;
46200  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
46201  return __ret;
46202 }
46203 #else
46204 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
46205  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46206  uint32x2_t __ret;
46207  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
46208  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46209  return __ret;
46210 }
46211 #endif
46212 
46213 #ifdef __LITTLE_ENDIAN__
46214 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
46215  uint32x2_t __ret;
46216  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
46217  return __ret;
46218 }
46219 #else
46220 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
46221  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46222  uint32x2_t __ret;
46223  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
46224  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46225  return __ret;
46226 }
46227 #endif
46228 
46229 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
46230  uint64x1_t __ret;
46231  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
46232  return __ret;
46233 }
46234 #ifdef __LITTLE_ENDIAN__
46235 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
46236  uint16x4_t __ret;
46237  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
46238  return __ret;
46239 }
46240 #else
46241 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
46242  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46243  uint16x4_t __ret;
46244  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
46245  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46246  return __ret;
46247 }
46248 #endif
46249 
46250 __ai uint64_t vceqzd_u64(uint64_t __p0) {
46251  uint64_t __ret;
46252  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
46253  return __ret;
46254 }
46255 __ai int64_t vceqzd_s64(int64_t __p0) {
46256  int64_t __ret;
46257  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
46258  return __ret;
46259 }
46260 __ai uint64_t vceqzd_f64(float64_t __p0) {
46261  uint64_t __ret;
46262  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
46263  return __ret;
46264 }
46265 __ai uint32_t vceqzs_f32(float32_t __p0) {
46266  uint32_t __ret;
46267  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
46268  return __ret;
46269 }
46270 #ifdef __LITTLE_ENDIAN__
46271 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46272  uint64x2_t __ret;
46273  __ret = (uint64x2_t)(__p0 >= __p1);
46274  return __ret;
46275 }
46276 #else
46277 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46278  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46279  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46280  uint64x2_t __ret;
46281  __ret = (uint64x2_t)(__rev0 >= __rev1);
46282  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46283  return __ret;
46284 }
46285 #endif
46286 
46287 #ifdef __LITTLE_ENDIAN__
46288 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
46289  uint64x2_t __ret;
46290  __ret = (uint64x2_t)(__p0 >= __p1);
46291  return __ret;
46292 }
46293 #else
46294 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
46295  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46296  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46297  uint64x2_t __ret;
46298  __ret = (uint64x2_t)(__rev0 >= __rev1);
46299  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46300  return __ret;
46301 }
46302 #endif
46303 
46304 #ifdef __LITTLE_ENDIAN__
46305 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
46306  uint64x2_t __ret;
46307  __ret = (uint64x2_t)(__p0 >= __p1);
46308  return __ret;
46309 }
46310 #else
46311 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
46312  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46313  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46314  uint64x2_t __ret;
46315  __ret = (uint64x2_t)(__rev0 >= __rev1);
46316  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46317  return __ret;
46318 }
46319 #endif
46320 
46321 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
46322  uint64x1_t __ret;
46323  __ret = (uint64x1_t)(__p0 >= __p1);
46324  return __ret;
46325 }
46326 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
46327  uint64x1_t __ret;
46328  __ret = (uint64x1_t)(__p0 >= __p1);
46329  return __ret;
46330 }
46331 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
46332  uint64x1_t __ret;
46333  __ret = (uint64x1_t)(__p0 >= __p1);
46334  return __ret;
46335 }
46336 __ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
46337  int64_t __ret;
46338  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
46339  return __ret;
46340 }
46341 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
46342  uint64_t __ret;
46343  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
46344  return __ret;
46345 }
46346 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
46347  uint64_t __ret;
46348  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
46349  return __ret;
46350 }
46351 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
46352  uint32_t __ret;
46353  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
46354  return __ret;
46355 }
46356 #ifdef __LITTLE_ENDIAN__
46357 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
46358  uint8x16_t __ret;
46359  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
46360  return __ret;
46361 }
46362 #else
46363 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
46364  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46365  uint8x16_t __ret;
46366  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
46367  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46368  return __ret;
46369 }
46370 #endif
46371 
46372 #ifdef __LITTLE_ENDIAN__
46373 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
46374  uint64x2_t __ret;
46375  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
46376  return __ret;
46377 }
46378 #else
46379 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
46380  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46381  uint64x2_t __ret;
46382  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
46383  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46384  return __ret;
46385 }
46386 #endif
46387 
46388 #ifdef __LITTLE_ENDIAN__
46389 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
46390  uint32x4_t __ret;
46391  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
46392  return __ret;
46393 }
46394 #else
46395 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
46396  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46397  uint32x4_t __ret;
46398  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
46399  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46400  return __ret;
46401 }
46402 #endif
46403 
46404 #ifdef __LITTLE_ENDIAN__
46405 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
46406  uint32x4_t __ret;
46407  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
46408  return __ret;
46409 }
46410 #else
46411 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
46412  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46413  uint32x4_t __ret;
46414  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
46415  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46416  return __ret;
46417 }
46418 #endif
46419 
46420 #ifdef __LITTLE_ENDIAN__
46421 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
46422  uint64x2_t __ret;
46423  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
46424  return __ret;
46425 }
46426 #else
46427 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
46428  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46429  uint64x2_t __ret;
46430  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
46431  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46432  return __ret;
46433 }
46434 #endif
46435 
46436 #ifdef __LITTLE_ENDIAN__
46437 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
46438  uint16x8_t __ret;
46439  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
46440  return __ret;
46441 }
46442 #else
46443 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
46444  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46445  uint16x8_t __ret;
46446  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
46447  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46448  return __ret;
46449 }
46450 #endif
46451 
46452 #ifdef __LITTLE_ENDIAN__
46453 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
46454  uint8x8_t __ret;
46455  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
46456  return __ret;
46457 }
46458 #else
46459 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
46460  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46461  uint8x8_t __ret;
46462  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
46463  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46464  return __ret;
46465 }
46466 #endif
46467 
46468 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
46469  uint64x1_t __ret;
46470  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
46471  return __ret;
46472 }
46473 #ifdef __LITTLE_ENDIAN__
46474 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
46475  uint32x2_t __ret;
46476  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
46477  return __ret;
46478 }
46479 #else
46480 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
46481  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46482  uint32x2_t __ret;
46483  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
46484  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46485  return __ret;
46486 }
46487 #endif
46488 
46489 #ifdef __LITTLE_ENDIAN__
46490 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
46491  uint32x2_t __ret;
46492  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
46493  return __ret;
46494 }
46495 #else
46496 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
46497  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46498  uint32x2_t __ret;
46499  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
46500  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46501  return __ret;
46502 }
46503 #endif
46504 
46505 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
46506  uint64x1_t __ret;
46507  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
46508  return __ret;
46509 }
46510 #ifdef __LITTLE_ENDIAN__
46511 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
46512  uint16x4_t __ret;
46513  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
46514  return __ret;
46515 }
46516 #else
46517 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
46518  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46519  uint16x4_t __ret;
46520  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
46521  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46522  return __ret;
46523 }
46524 #endif
46525 
46526 __ai int64_t vcgezd_s64(int64_t __p0) {
46527  int64_t __ret;
46528  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
46529  return __ret;
46530 }
46531 __ai uint64_t vcgezd_f64(float64_t __p0) {
46532  uint64_t __ret;
46533  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
46534  return __ret;
46535 }
46536 __ai uint32_t vcgezs_f32(float32_t __p0) {
46537  uint32_t __ret;
46538  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
46539  return __ret;
46540 }
46541 #ifdef __LITTLE_ENDIAN__
46542 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46543  uint64x2_t __ret;
46544  __ret = (uint64x2_t)(__p0 > __p1);
46545  return __ret;
46546 }
46547 #else
46548 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46549  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46550  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46551  uint64x2_t __ret;
46552  __ret = (uint64x2_t)(__rev0 > __rev1);
46553  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46554  return __ret;
46555 }
46556 #endif
46557 
46558 #ifdef __LITTLE_ENDIAN__
46559 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
46560  uint64x2_t __ret;
46561  __ret = (uint64x2_t)(__p0 > __p1);
46562  return __ret;
46563 }
46564 #else
46565 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
46566  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46567  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46568  uint64x2_t __ret;
46569  __ret = (uint64x2_t)(__rev0 > __rev1);
46570  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46571  return __ret;
46572 }
46573 #endif
46574 
46575 #ifdef __LITTLE_ENDIAN__
46576 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
46577  uint64x2_t __ret;
46578  __ret = (uint64x2_t)(__p0 > __p1);
46579  return __ret;
46580 }
46581 #else
46582 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
46583  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46584  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46585  uint64x2_t __ret;
46586  __ret = (uint64x2_t)(__rev0 > __rev1);
46587  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46588  return __ret;
46589 }
46590 #endif
46591 
46592 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
46593  uint64x1_t __ret;
46594  __ret = (uint64x1_t)(__p0 > __p1);
46595  return __ret;
46596 }
46597 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
46598  uint64x1_t __ret;
46599  __ret = (uint64x1_t)(__p0 > __p1);
46600  return __ret;
46601 }
46602 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
46603  uint64x1_t __ret;
46604  __ret = (uint64x1_t)(__p0 > __p1);
46605  return __ret;
46606 }
46607 __ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
46608  int64_t __ret;
46609  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
46610  return __ret;
46611 }
46612 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
46613  uint64_t __ret;
46614  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
46615  return __ret;
46616 }
46617 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
46618  uint64_t __ret;
46619  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
46620  return __ret;
46621 }
46622 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
46623  uint32_t __ret;
46624  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
46625  return __ret;
46626 }
46627 #ifdef __LITTLE_ENDIAN__
46628 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
46629  uint8x16_t __ret;
46630  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
46631  return __ret;
46632 }
46633 #else
46634 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
46635  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46636  uint8x16_t __ret;
46637  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
46638  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46639  return __ret;
46640 }
46641 #endif
46642 
46643 #ifdef __LITTLE_ENDIAN__
46644 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
46645  uint64x2_t __ret;
46646  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
46647  return __ret;
46648 }
46649 #else
46650 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
46651  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46652  uint64x2_t __ret;
46653  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
46654  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46655  return __ret;
46656 }
46657 #endif
46658 
46659 #ifdef __LITTLE_ENDIAN__
46660 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
46661  uint32x4_t __ret;
46662  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
46663  return __ret;
46664 }
46665 #else
46666 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
46667  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46668  uint32x4_t __ret;
46669  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
46670  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46671  return __ret;
46672 }
46673 #endif
46674 
46675 #ifdef __LITTLE_ENDIAN__
46676 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
46677  uint32x4_t __ret;
46678  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
46679  return __ret;
46680 }
46681 #else
46682 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
46683  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46684  uint32x4_t __ret;
46685  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
46686  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46687  return __ret;
46688 }
46689 #endif
46690 
46691 #ifdef __LITTLE_ENDIAN__
46692 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
46693  uint64x2_t __ret;
46694  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
46695  return __ret;
46696 }
46697 #else
46698 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
46699  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46700  uint64x2_t __ret;
46701  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
46702  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46703  return __ret;
46704 }
46705 #endif
46706 
46707 #ifdef __LITTLE_ENDIAN__
46708 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
46709  uint16x8_t __ret;
46710  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
46711  return __ret;
46712 }
46713 #else
46714 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
46715  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46716  uint16x8_t __ret;
46717  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
46718  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46719  return __ret;
46720 }
46721 #endif
46722 
46723 #ifdef __LITTLE_ENDIAN__
46724 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
46725  uint8x8_t __ret;
46726  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
46727  return __ret;
46728 }
46729 #else
46730 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
46731  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46732  uint8x8_t __ret;
46733  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
46734  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46735  return __ret;
46736 }
46737 #endif
46738 
46739 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
46740  uint64x1_t __ret;
46741  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
46742  return __ret;
46743 }
46744 #ifdef __LITTLE_ENDIAN__
46745 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
46746  uint32x2_t __ret;
46747  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
46748  return __ret;
46749 }
46750 #else
46751 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
46752  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46753  uint32x2_t __ret;
46754  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
46755  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46756  return __ret;
46757 }
46758 #endif
46759 
46760 #ifdef __LITTLE_ENDIAN__
46761 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
46762  uint32x2_t __ret;
46763  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
46764  return __ret;
46765 }
46766 #else
46767 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
46768  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46769  uint32x2_t __ret;
46770  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
46771  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46772  return __ret;
46773 }
46774 #endif
46775 
46776 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
46777  uint64x1_t __ret;
46778  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
46779  return __ret;
46780 }
46781 #ifdef __LITTLE_ENDIAN__
46782 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
46783  uint16x4_t __ret;
46784  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
46785  return __ret;
46786 }
46787 #else
46788 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
46789  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46790  uint16x4_t __ret;
46791  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
46792  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46793  return __ret;
46794 }
46795 #endif
46796 
46797 __ai int64_t vcgtzd_s64(int64_t __p0) {
46798  int64_t __ret;
46799  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
46800  return __ret;
46801 }
46802 __ai uint64_t vcgtzd_f64(float64_t __p0) {
46803  uint64_t __ret;
46804  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
46805  return __ret;
46806 }
46807 __ai uint32_t vcgtzs_f32(float32_t __p0) {
46808  uint32_t __ret;
46809  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
46810  return __ret;
46811 }
46812 #ifdef __LITTLE_ENDIAN__
46813 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46814  uint64x2_t __ret;
46815  __ret = (uint64x2_t)(__p0 <= __p1);
46816  return __ret;
46817 }
46818 #else
46819 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
46820  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46821  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46822  uint64x2_t __ret;
46823  __ret = (uint64x2_t)(__rev0 <= __rev1);
46824  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46825  return __ret;
46826 }
46827 #endif
46828 
46829 #ifdef __LITTLE_ENDIAN__
46830 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
46831  uint64x2_t __ret;
46832  __ret = (uint64x2_t)(__p0 <= __p1);
46833  return __ret;
46834 }
46835 #else
46836 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
46837  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46838  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46839  uint64x2_t __ret;
46840  __ret = (uint64x2_t)(__rev0 <= __rev1);
46841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46842  return __ret;
46843 }
46844 #endif
46845 
46846 #ifdef __LITTLE_ENDIAN__
46847 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
46848  uint64x2_t __ret;
46849  __ret = (uint64x2_t)(__p0 <= __p1);
46850  return __ret;
46851 }
46852 #else
46853 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
46854  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46855  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46856  uint64x2_t __ret;
46857  __ret = (uint64x2_t)(__rev0 <= __rev1);
46858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46859  return __ret;
46860 }
46861 #endif
46862 
46863 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
46864  uint64x1_t __ret;
46865  __ret = (uint64x1_t)(__p0 <= __p1);
46866  return __ret;
46867 }
46868 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
46869  uint64x1_t __ret;
46870  __ret = (uint64x1_t)(__p0 <= __p1);
46871  return __ret;
46872 }
46873 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
46874  uint64x1_t __ret;
46875  __ret = (uint64x1_t)(__p0 <= __p1);
46876  return __ret;
46877 }
46878 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
46879  uint64_t __ret;
46880  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
46881  return __ret;
46882 }
46883 __ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
46884  int64_t __ret;
46885  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
46886  return __ret;
46887 }
46888 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
46889  uint64_t __ret;
46890  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
46891  return __ret;
46892 }
46893 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
46894  uint32_t __ret;
46895  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
46896  return __ret;
46897 }
46898 #ifdef __LITTLE_ENDIAN__
46899 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
46900  uint8x16_t __ret;
46901  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
46902  return __ret;
46903 }
46904 #else
46905 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
46906  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46907  uint8x16_t __ret;
46908  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
46909  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
46910  return __ret;
46911 }
46912 #endif
46913 
46914 #ifdef __LITTLE_ENDIAN__
46915 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
46916  uint64x2_t __ret;
46917  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
46918  return __ret;
46919 }
46920 #else
46921 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
46922  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46923  uint64x2_t __ret;
46924  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
46925  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46926  return __ret;
46927 }
46928 #endif
46929 
46930 #ifdef __LITTLE_ENDIAN__
46931 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
46932  uint32x4_t __ret;
46933  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
46934  return __ret;
46935 }
46936 #else
46937 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
46938  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46939  uint32x4_t __ret;
46940  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
46941  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46942  return __ret;
46943 }
46944 #endif
46945 
46946 #ifdef __LITTLE_ENDIAN__
46947 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
46948  uint32x4_t __ret;
46949  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
46950  return __ret;
46951 }
46952 #else
46953 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
46954  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46955  uint32x4_t __ret;
46956  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
46957  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46958  return __ret;
46959 }
46960 #endif
46961 
46962 #ifdef __LITTLE_ENDIAN__
46963 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
46964  uint64x2_t __ret;
46965  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
46966  return __ret;
46967 }
46968 #else
46969 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
46970  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46971  uint64x2_t __ret;
46972  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
46973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46974  return __ret;
46975 }
46976 #endif
46977 
46978 #ifdef __LITTLE_ENDIAN__
46979 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
46980  uint16x8_t __ret;
46981  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
46982  return __ret;
46983 }
46984 #else
46985 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
46986  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
46987  uint16x8_t __ret;
46988  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
46989  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
46990  return __ret;
46991 }
46992 #endif
46993 
46994 #ifdef __LITTLE_ENDIAN__
46995 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
46996  uint8x8_t __ret;
46997  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
46998  return __ret;
46999 }
47000 #else
47001 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
47002  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
47003  uint8x8_t __ret;
47004  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
47005  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
47006  return __ret;
47007 }
47008 #endif
47009 
47010 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
47011  uint64x1_t __ret;
47012  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
47013  return __ret;
47014 }
47015 #ifdef __LITTLE_ENDIAN__
47016 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
47017  uint32x2_t __ret;
47018  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
47019  return __ret;
47020 }
47021 #else
47022 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
47023  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47024  uint32x2_t __ret;
47025  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
47026  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47027  return __ret;
47028 }
47029 #endif
47030 
47031 #ifdef __LITTLE_ENDIAN__
47032 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
47033  uint32x2_t __ret;
47034  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
47035  return __ret;
47036 }
47037 #else
47038 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
47039  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47040  uint32x2_t __ret;
47041  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
47042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47043  return __ret;
47044 }
47045 #endif
47046 
47047 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
47048  uint64x1_t __ret;
47049  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
47050  return __ret;
47051 }
47052 #ifdef __LITTLE_ENDIAN__
47053 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
47054  uint16x4_t __ret;
47055  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
47056  return __ret;
47057 }
47058 #else
47059 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
47060  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47061  uint16x4_t __ret;
47062  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
47063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47064  return __ret;
47065 }
47066 #endif
47067 
47068 __ai int64_t vclezd_s64(int64_t __p0) {
47069  int64_t __ret;
47070  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
47071  return __ret;
47072 }
47073 __ai uint64_t vclezd_f64(float64_t __p0) {
47074  uint64_t __ret;
47075  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
47076  return __ret;
47077 }
47078 __ai uint32_t vclezs_f32(float32_t __p0) {
47079  uint32_t __ret;
47080  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
47081  return __ret;
47082 }
47083 #ifdef __LITTLE_ENDIAN__
47084 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
47085  uint64x2_t __ret;
47086  __ret = (uint64x2_t)(__p0 < __p1);
47087  return __ret;
47088 }
47089 #else
47090 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
47091  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47092  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47093  uint64x2_t __ret;
47094  __ret = (uint64x2_t)(__rev0 < __rev1);
47095  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47096  return __ret;
47097 }
47098 #endif
47099 
47100 #ifdef __LITTLE_ENDIAN__
47101 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
47102  uint64x2_t __ret;
47103  __ret = (uint64x2_t)(__p0 < __p1);
47104  return __ret;
47105 }
47106 #else
47107 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
47108  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47109  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47110  uint64x2_t __ret;
47111  __ret = (uint64x2_t)(__rev0 < __rev1);
47112  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47113  return __ret;
47114 }
47115 #endif
47116 
47117 #ifdef __LITTLE_ENDIAN__
47118 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
47119  uint64x2_t __ret;
47120  __ret = (uint64x2_t)(__p0 < __p1);
47121  return __ret;
47122 }
47123 #else
47124 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
47125  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47126  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47127  uint64x2_t __ret;
47128  __ret = (uint64x2_t)(__rev0 < __rev1);
47129  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47130  return __ret;
47131 }
47132 #endif
47133 
47134 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
47135  uint64x1_t __ret;
47136  __ret = (uint64x1_t)(__p0 < __p1);
47137  return __ret;
47138 }
47139 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
47140  uint64x1_t __ret;
47141  __ret = (uint64x1_t)(__p0 < __p1);
47142  return __ret;
47143 }
47144 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
47145  uint64x1_t __ret;
47146  __ret = (uint64x1_t)(__p0 < __p1);
47147  return __ret;
47148 }
47149 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
47150  uint64_t __ret;
47151  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
47152  return __ret;
47153 }
47154 __ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
47155  int64_t __ret;
47156  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
47157  return __ret;
47158 }
47159 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
47160  uint64_t __ret;
47161  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
47162  return __ret;
47163 }
47164 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
47165  uint32_t __ret;
47166  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
47167  return __ret;
47168 }
47169 #ifdef __LITTLE_ENDIAN__
47170 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
47171  uint8x16_t __ret;
47172  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
47173  return __ret;
47174 }
47175 #else
47176 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
47177  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
47178  uint8x16_t __ret;
47179  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
47180  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
47181  return __ret;
47182 }
47183 #endif
47184 
47185 #ifdef __LITTLE_ENDIAN__
47186 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
47187  uint64x2_t __ret;
47188  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
47189  return __ret;
47190 }
47191 #else
47192 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
47193  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47194  uint64x2_t __ret;
47195  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
47196  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47197  return __ret;
47198 }
47199 #endif
47200 
47201 #ifdef __LITTLE_ENDIAN__
47202 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
47203  uint32x4_t __ret;
47204  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
47205  return __ret;
47206 }
47207 #else
47208 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
47209  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47210  uint32x4_t __ret;
47211  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
47212  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47213  return __ret;
47214 }
47215 #endif
47216 
47217 #ifdef __LITTLE_ENDIAN__
47218 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
47219  uint32x4_t __ret;
47220  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
47221  return __ret;
47222 }
47223 #else
47224 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
47225  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47226  uint32x4_t __ret;
47227  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
47228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47229  return __ret;
47230 }
47231 #endif
47232 
47233 #ifdef __LITTLE_ENDIAN__
47234 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
47235  uint64x2_t __ret;
47236  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
47237  return __ret;
47238 }
47239 #else
47240 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
47241  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47242  uint64x2_t __ret;
47243  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
47244  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47245  return __ret;
47246 }
47247 #endif
47248 
47249 #ifdef __LITTLE_ENDIAN__
47250 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
47251  uint16x8_t __ret;
47252  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
47253  return __ret;
47254 }
47255 #else
47256 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
47257  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
47258  uint16x8_t __ret;
47259  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
47260  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
47261  return __ret;
47262 }
47263 #endif
47264 
47265 #ifdef __LITTLE_ENDIAN__
47266 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
47267  uint8x8_t __ret;
47268  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
47269  return __ret;
47270 }
47271 #else
47272 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
47273  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
47274  uint8x8_t __ret;
47275  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
47276  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
47277  return __ret;
47278 }
47279 #endif
47280 
47281 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
47282  uint64x1_t __ret;
47283  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
47284  return __ret;
47285 }
47286 #ifdef __LITTLE_ENDIAN__
47287 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
47288  uint32x2_t __ret;
47289  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
47290  return __ret;
47291 }
47292 #else
47293 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
47294  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47295  uint32x2_t __ret;
47296  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
47297  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47298  return __ret;
47299 }
47300 #endif
47301 
47302 #ifdef __LITTLE_ENDIAN__
47303 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
47304  uint32x2_t __ret;
47305  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
47306  return __ret;
47307 }
47308 #else
47309 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
47310  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47311  uint32x2_t __ret;
47312  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
47313  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47314  return __ret;
47315 }
47316 #endif
47317 
47318 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
47319  uint64x1_t __ret;
47320  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
47321  return __ret;
47322 }
47323 #ifdef __LITTLE_ENDIAN__
47324 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
47325  uint16x4_t __ret;
47326  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
47327  return __ret;
47328 }
47329 #else
47330 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
47331  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47332  uint16x4_t __ret;
47333  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
47334  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47335  return __ret;
47336 }
47337 #endif
47338 
47339 __ai int64_t vcltzd_s64(int64_t __p0) {
47340  int64_t __ret;
47341  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
47342  return __ret;
47343 }
47344 __ai uint64_t vcltzd_f64(float64_t __p0) {
47345  uint64_t __ret;
47346  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
47347  return __ret;
47348 }
47349 __ai uint32_t vcltzs_f32(float32_t __p0) {
47350  uint32_t __ret;
47351  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
47352  return __ret;
47353 }
47354 #ifdef __LITTLE_ENDIAN__
47355 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
47356  poly64x2_t __ret;
47357  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
47358  return __ret;
47359 }
47360 #else
47361 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
47362  poly64x2_t __ret;
47363  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
47364  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47365  return __ret;
47366 }
47367 #endif
47368 
47369 #ifdef __LITTLE_ENDIAN__
47370 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
47371  float64x2_t __ret;
47372  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
47373  return __ret;
47374 }
47375 #else
47376 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
47377  float64x2_t __ret;
47378  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
47379  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47380  return __ret;
47381 }
47382 #endif
47383 
47384 #ifdef __LITTLE_ENDIAN__
47385 #define vcopyq_lane_p8(__p0_249, __p1_249, __p2_249, __p3_249) __extension__ ({ \
47386  poly8x16_t __s0_249 = __p0_249; \
47387  poly8x8_t __s2_249 = __p2_249; \
47388  poly8x16_t __ret_249; \
47389  __ret_249 = vsetq_lane_p8(vget_lane_p8(__s2_249, __p3_249), __s0_249, __p1_249); \
47390  __ret_249; \
47391 })
47392 #else
47393 #define vcopyq_lane_p8(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
47394  poly8x16_t __s0_250 = __p0_250; \
47395  poly8x8_t __s2_250 = __p2_250; \
47396  poly8x16_t __rev0_250; __rev0_250 = __builtin_shufflevector(__s0_250, __s0_250, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47397  poly8x8_t __rev2_250; __rev2_250 = __builtin_shufflevector(__s2_250, __s2_250, 7, 6, 5, 4, 3, 2, 1, 0); \
47398  poly8x16_t __ret_250; \
47399  __ret_250 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_250, __p3_250), __rev0_250, __p1_250); \
47400  __ret_250 = __builtin_shufflevector(__ret_250, __ret_250, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47401  __ret_250; \
47402 })
47403 #endif
47404 
47405 #ifdef __LITTLE_ENDIAN__
47406 #define vcopyq_lane_p16(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
47407  poly16x8_t __s0_251 = __p0_251; \
47408  poly16x4_t __s2_251 = __p2_251; \
47409  poly16x8_t __ret_251; \
47410  __ret_251 = vsetq_lane_p16(vget_lane_p16(__s2_251, __p3_251), __s0_251, __p1_251); \
47411  __ret_251; \
47412 })
47413 #else
47414 #define vcopyq_lane_p16(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
47415  poly16x8_t __s0_252 = __p0_252; \
47416  poly16x4_t __s2_252 = __p2_252; \
47417  poly16x8_t __rev0_252; __rev0_252 = __builtin_shufflevector(__s0_252, __s0_252, 7, 6, 5, 4, 3, 2, 1, 0); \
47418  poly16x4_t __rev2_252; __rev2_252 = __builtin_shufflevector(__s2_252, __s2_252, 3, 2, 1, 0); \
47419  poly16x8_t __ret_252; \
47420  __ret_252 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_252, __p3_252), __rev0_252, __p1_252); \
47421  __ret_252 = __builtin_shufflevector(__ret_252, __ret_252, 7, 6, 5, 4, 3, 2, 1, 0); \
47422  __ret_252; \
47423 })
47424 #endif
47425 
47426 #ifdef __LITTLE_ENDIAN__
47427 #define vcopyq_lane_u8(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
47428  uint8x16_t __s0_253 = __p0_253; \
47429  uint8x8_t __s2_253 = __p2_253; \
47430  uint8x16_t __ret_253; \
47431  __ret_253 = vsetq_lane_u8(vget_lane_u8(__s2_253, __p3_253), __s0_253, __p1_253); \
47432  __ret_253; \
47433 })
47434 #else
47435 #define vcopyq_lane_u8(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
47436  uint8x16_t __s0_254 = __p0_254; \
47437  uint8x8_t __s2_254 = __p2_254; \
47438  uint8x16_t __rev0_254; __rev0_254 = __builtin_shufflevector(__s0_254, __s0_254, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47439  uint8x8_t __rev2_254; __rev2_254 = __builtin_shufflevector(__s2_254, __s2_254, 7, 6, 5, 4, 3, 2, 1, 0); \
47440  uint8x16_t __ret_254; \
47441  __ret_254 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_254, __p3_254), __rev0_254, __p1_254); \
47442  __ret_254 = __builtin_shufflevector(__ret_254, __ret_254, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47443  __ret_254; \
47444 })
47445 #endif
47446 
47447 #ifdef __LITTLE_ENDIAN__
47448 #define vcopyq_lane_u32(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
47449  uint32x4_t __s0_255 = __p0_255; \
47450  uint32x2_t __s2_255 = __p2_255; \
47451  uint32x4_t __ret_255; \
47452  __ret_255 = vsetq_lane_u32(vget_lane_u32(__s2_255, __p3_255), __s0_255, __p1_255); \
47453  __ret_255; \
47454 })
47455 #else
47456 #define vcopyq_lane_u32(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
47457  uint32x4_t __s0_256 = __p0_256; \
47458  uint32x2_t __s2_256 = __p2_256; \
47459  uint32x4_t __rev0_256; __rev0_256 = __builtin_shufflevector(__s0_256, __s0_256, 3, 2, 1, 0); \
47460  uint32x2_t __rev2_256; __rev2_256 = __builtin_shufflevector(__s2_256, __s2_256, 1, 0); \
47461  uint32x4_t __ret_256; \
47462  __ret_256 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_256, __p3_256), __rev0_256, __p1_256); \
47463  __ret_256 = __builtin_shufflevector(__ret_256, __ret_256, 3, 2, 1, 0); \
47464  __ret_256; \
47465 })
47466 #endif
47467 
47468 #ifdef __LITTLE_ENDIAN__
47469 #define vcopyq_lane_u64(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
47470  uint64x2_t __s0_257 = __p0_257; \
47471  uint64x1_t __s2_257 = __p2_257; \
47472  uint64x2_t __ret_257; \
47473  __ret_257 = vsetq_lane_u64(vget_lane_u64(__s2_257, __p3_257), __s0_257, __p1_257); \
47474  __ret_257; \
47475 })
47476 #else
47477 #define vcopyq_lane_u64(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
47478  uint64x2_t __s0_258 = __p0_258; \
47479  uint64x1_t __s2_258 = __p2_258; \
47480  uint64x2_t __rev0_258; __rev0_258 = __builtin_shufflevector(__s0_258, __s0_258, 1, 0); \
47481  uint64x2_t __ret_258; \
47482  __ret_258 = __noswap_vsetq_lane_u64(vget_lane_u64(__s2_258, __p3_258), __rev0_258, __p1_258); \
47483  __ret_258 = __builtin_shufflevector(__ret_258, __ret_258, 1, 0); \
47484  __ret_258; \
47485 })
47486 #endif
47487 
47488 #ifdef __LITTLE_ENDIAN__
47489 #define vcopyq_lane_u16(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
47490  uint16x8_t __s0_259 = __p0_259; \
47491  uint16x4_t __s2_259 = __p2_259; \
47492  uint16x8_t __ret_259; \
47493  __ret_259 = vsetq_lane_u16(vget_lane_u16(__s2_259, __p3_259), __s0_259, __p1_259); \
47494  __ret_259; \
47495 })
47496 #else
47497 #define vcopyq_lane_u16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
47498  uint16x8_t __s0_260 = __p0_260; \
47499  uint16x4_t __s2_260 = __p2_260; \
47500  uint16x8_t __rev0_260; __rev0_260 = __builtin_shufflevector(__s0_260, __s0_260, 7, 6, 5, 4, 3, 2, 1, 0); \
47501  uint16x4_t __rev2_260; __rev2_260 = __builtin_shufflevector(__s2_260, __s2_260, 3, 2, 1, 0); \
47502  uint16x8_t __ret_260; \
47503  __ret_260 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_260, __p3_260), __rev0_260, __p1_260); \
47504  __ret_260 = __builtin_shufflevector(__ret_260, __ret_260, 7, 6, 5, 4, 3, 2, 1, 0); \
47505  __ret_260; \
47506 })
47507 #endif
47508 
47509 #ifdef __LITTLE_ENDIAN__
47510 #define vcopyq_lane_s8(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
47511  int8x16_t __s0_261 = __p0_261; \
47512  int8x8_t __s2_261 = __p2_261; \
47513  int8x16_t __ret_261; \
47514  __ret_261 = vsetq_lane_s8(vget_lane_s8(__s2_261, __p3_261), __s0_261, __p1_261); \
47515  __ret_261; \
47516 })
47517 #else
47518 #define vcopyq_lane_s8(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
47519  int8x16_t __s0_262 = __p0_262; \
47520  int8x8_t __s2_262 = __p2_262; \
47521  int8x16_t __rev0_262; __rev0_262 = __builtin_shufflevector(__s0_262, __s0_262, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47522  int8x8_t __rev2_262; __rev2_262 = __builtin_shufflevector(__s2_262, __s2_262, 7, 6, 5, 4, 3, 2, 1, 0); \
47523  int8x16_t __ret_262; \
47524  __ret_262 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_262, __p3_262), __rev0_262, __p1_262); \
47525  __ret_262 = __builtin_shufflevector(__ret_262, __ret_262, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47526  __ret_262; \
47527 })
47528 #endif
47529 
47530 #ifdef __LITTLE_ENDIAN__
47531 #define vcopyq_lane_f32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
47532  float32x4_t __s0_263 = __p0_263; \
47533  float32x2_t __s2_263 = __p2_263; \
47534  float32x4_t __ret_263; \
47535  __ret_263 = vsetq_lane_f32(vget_lane_f32(__s2_263, __p3_263), __s0_263, __p1_263); \
47536  __ret_263; \
47537 })
47538 #else
47539 #define vcopyq_lane_f32(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
47540  float32x4_t __s0_264 = __p0_264; \
47541  float32x2_t __s2_264 = __p2_264; \
47542  float32x4_t __rev0_264; __rev0_264 = __builtin_shufflevector(__s0_264, __s0_264, 3, 2, 1, 0); \
47543  float32x2_t __rev2_264; __rev2_264 = __builtin_shufflevector(__s2_264, __s2_264, 1, 0); \
47544  float32x4_t __ret_264; \
47545  __ret_264 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_264, __p3_264), __rev0_264, __p1_264); \
47546  __ret_264 = __builtin_shufflevector(__ret_264, __ret_264, 3, 2, 1, 0); \
47547  __ret_264; \
47548 })
47549 #endif
47550 
47551 #ifdef __LITTLE_ENDIAN__
47552 #define vcopyq_lane_s32(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
47553  int32x4_t __s0_265 = __p0_265; \
47554  int32x2_t __s2_265 = __p2_265; \
47555  int32x4_t __ret_265; \
47556  __ret_265 = vsetq_lane_s32(vget_lane_s32(__s2_265, __p3_265), __s0_265, __p1_265); \
47557  __ret_265; \
47558 })
47559 #else
47560 #define vcopyq_lane_s32(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
47561  int32x4_t __s0_266 = __p0_266; \
47562  int32x2_t __s2_266 = __p2_266; \
47563  int32x4_t __rev0_266; __rev0_266 = __builtin_shufflevector(__s0_266, __s0_266, 3, 2, 1, 0); \
47564  int32x2_t __rev2_266; __rev2_266 = __builtin_shufflevector(__s2_266, __s2_266, 1, 0); \
47565  int32x4_t __ret_266; \
47566  __ret_266 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_266, __p3_266), __rev0_266, __p1_266); \
47567  __ret_266 = __builtin_shufflevector(__ret_266, __ret_266, 3, 2, 1, 0); \
47568  __ret_266; \
47569 })
47570 #endif
47571 
47572 #ifdef __LITTLE_ENDIAN__
47573 #define vcopyq_lane_s64(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
47574  int64x2_t __s0_267 = __p0_267; \
47575  int64x1_t __s2_267 = __p2_267; \
47576  int64x2_t __ret_267; \
47577  __ret_267 = vsetq_lane_s64(vget_lane_s64(__s2_267, __p3_267), __s0_267, __p1_267); \
47578  __ret_267; \
47579 })
47580 #else
47581 #define vcopyq_lane_s64(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
47582  int64x2_t __s0_268 = __p0_268; \
47583  int64x1_t __s2_268 = __p2_268; \
47584  int64x2_t __rev0_268; __rev0_268 = __builtin_shufflevector(__s0_268, __s0_268, 1, 0); \
47585  int64x2_t __ret_268; \
47586  __ret_268 = __noswap_vsetq_lane_s64(vget_lane_s64(__s2_268, __p3_268), __rev0_268, __p1_268); \
47587  __ret_268 = __builtin_shufflevector(__ret_268, __ret_268, 1, 0); \
47588  __ret_268; \
47589 })
47590 #endif
47591 
47592 #ifdef __LITTLE_ENDIAN__
47593 #define vcopyq_lane_s16(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
47594  int16x8_t __s0_269 = __p0_269; \
47595  int16x4_t __s2_269 = __p2_269; \
47596  int16x8_t __ret_269; \
47597  __ret_269 = vsetq_lane_s16(vget_lane_s16(__s2_269, __p3_269), __s0_269, __p1_269); \
47598  __ret_269; \
47599 })
47600 #else
47601 #define vcopyq_lane_s16(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
47602  int16x8_t __s0_270 = __p0_270; \
47603  int16x4_t __s2_270 = __p2_270; \
47604  int16x8_t __rev0_270; __rev0_270 = __builtin_shufflevector(__s0_270, __s0_270, 7, 6, 5, 4, 3, 2, 1, 0); \
47605  int16x4_t __rev2_270; __rev2_270 = __builtin_shufflevector(__s2_270, __s2_270, 3, 2, 1, 0); \
47606  int16x8_t __ret_270; \
47607  __ret_270 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_270, __p3_270), __rev0_270, __p1_270); \
47608  __ret_270 = __builtin_shufflevector(__ret_270, __ret_270, 7, 6, 5, 4, 3, 2, 1, 0); \
47609  __ret_270; \
47610 })
47611 #endif
47612 
47613 #ifdef __LITTLE_ENDIAN__
47614 #define vcopy_lane_p8(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
47615  poly8x8_t __s0_271 = __p0_271; \
47616  poly8x8_t __s2_271 = __p2_271; \
47617  poly8x8_t __ret_271; \
47618  __ret_271 = vset_lane_p8(vget_lane_p8(__s2_271, __p3_271), __s0_271, __p1_271); \
47619  __ret_271; \
47620 })
47621 #else
47622 #define vcopy_lane_p8(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
47623  poly8x8_t __s0_272 = __p0_272; \
47624  poly8x8_t __s2_272 = __p2_272; \
47625  poly8x8_t __rev0_272; __rev0_272 = __builtin_shufflevector(__s0_272, __s0_272, 7, 6, 5, 4, 3, 2, 1, 0); \
47626  poly8x8_t __rev2_272; __rev2_272 = __builtin_shufflevector(__s2_272, __s2_272, 7, 6, 5, 4, 3, 2, 1, 0); \
47627  poly8x8_t __ret_272; \
47628  __ret_272 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_272, __p3_272), __rev0_272, __p1_272); \
47629  __ret_272 = __builtin_shufflevector(__ret_272, __ret_272, 7, 6, 5, 4, 3, 2, 1, 0); \
47630  __ret_272; \
47631 })
47632 #endif
47633 
47634 #ifdef __LITTLE_ENDIAN__
47635 #define vcopy_lane_p16(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
47636  poly16x4_t __s0_273 = __p0_273; \
47637  poly16x4_t __s2_273 = __p2_273; \
47638  poly16x4_t __ret_273; \
47639  __ret_273 = vset_lane_p16(vget_lane_p16(__s2_273, __p3_273), __s0_273, __p1_273); \
47640  __ret_273; \
47641 })
47642 #else
47643 #define vcopy_lane_p16(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
47644  poly16x4_t __s0_274 = __p0_274; \
47645  poly16x4_t __s2_274 = __p2_274; \
47646  poly16x4_t __rev0_274; __rev0_274 = __builtin_shufflevector(__s0_274, __s0_274, 3, 2, 1, 0); \
47647  poly16x4_t __rev2_274; __rev2_274 = __builtin_shufflevector(__s2_274, __s2_274, 3, 2, 1, 0); \
47648  poly16x4_t __ret_274; \
47649  __ret_274 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_274, __p3_274), __rev0_274, __p1_274); \
47650  __ret_274 = __builtin_shufflevector(__ret_274, __ret_274, 3, 2, 1, 0); \
47651  __ret_274; \
47652 })
47653 #endif
47654 
47655 #ifdef __LITTLE_ENDIAN__
47656 #define vcopy_lane_u8(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
47657  uint8x8_t __s0_275 = __p0_275; \
47658  uint8x8_t __s2_275 = __p2_275; \
47659  uint8x8_t __ret_275; \
47660  __ret_275 = vset_lane_u8(vget_lane_u8(__s2_275, __p3_275), __s0_275, __p1_275); \
47661  __ret_275; \
47662 })
47663 #else
47664 #define vcopy_lane_u8(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
47665  uint8x8_t __s0_276 = __p0_276; \
47666  uint8x8_t __s2_276 = __p2_276; \
47667  uint8x8_t __rev0_276; __rev0_276 = __builtin_shufflevector(__s0_276, __s0_276, 7, 6, 5, 4, 3, 2, 1, 0); \
47668  uint8x8_t __rev2_276; __rev2_276 = __builtin_shufflevector(__s2_276, __s2_276, 7, 6, 5, 4, 3, 2, 1, 0); \
47669  uint8x8_t __ret_276; \
47670  __ret_276 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_276, __p3_276), __rev0_276, __p1_276); \
47671  __ret_276 = __builtin_shufflevector(__ret_276, __ret_276, 7, 6, 5, 4, 3, 2, 1, 0); \
47672  __ret_276; \
47673 })
47674 #endif
47675 
47676 #ifdef __LITTLE_ENDIAN__
47677 #define vcopy_lane_u32(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
47678  uint32x2_t __s0_277 = __p0_277; \
47679  uint32x2_t __s2_277 = __p2_277; \
47680  uint32x2_t __ret_277; \
47681  __ret_277 = vset_lane_u32(vget_lane_u32(__s2_277, __p3_277), __s0_277, __p1_277); \
47682  __ret_277; \
47683 })
47684 #else
47685 #define vcopy_lane_u32(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
47686  uint32x2_t __s0_278 = __p0_278; \
47687  uint32x2_t __s2_278 = __p2_278; \
47688  uint32x2_t __rev0_278; __rev0_278 = __builtin_shufflevector(__s0_278, __s0_278, 1, 0); \
47689  uint32x2_t __rev2_278; __rev2_278 = __builtin_shufflevector(__s2_278, __s2_278, 1, 0); \
47690  uint32x2_t __ret_278; \
47691  __ret_278 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_278, __p3_278), __rev0_278, __p1_278); \
47692  __ret_278 = __builtin_shufflevector(__ret_278, __ret_278, 1, 0); \
47693  __ret_278; \
47694 })
47695 #endif
47696 
47697 #define vcopy_lane_u64(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
47698  uint64x1_t __s0_279 = __p0_279; \
47699  uint64x1_t __s2_279 = __p2_279; \
47700  uint64x1_t __ret_279; \
47701  __ret_279 = vset_lane_u64(vget_lane_u64(__s2_279, __p3_279), __s0_279, __p1_279); \
47702  __ret_279; \
47703 })
47704 #ifdef __LITTLE_ENDIAN__
47705 #define vcopy_lane_u16(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
47706  uint16x4_t __s0_280 = __p0_280; \
47707  uint16x4_t __s2_280 = __p2_280; \
47708  uint16x4_t __ret_280; \
47709  __ret_280 = vset_lane_u16(vget_lane_u16(__s2_280, __p3_280), __s0_280, __p1_280); \
47710  __ret_280; \
47711 })
47712 #else
47713 #define vcopy_lane_u16(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
47714  uint16x4_t __s0_281 = __p0_281; \
47715  uint16x4_t __s2_281 = __p2_281; \
47716  uint16x4_t __rev0_281; __rev0_281 = __builtin_shufflevector(__s0_281, __s0_281, 3, 2, 1, 0); \
47717  uint16x4_t __rev2_281; __rev2_281 = __builtin_shufflevector(__s2_281, __s2_281, 3, 2, 1, 0); \
47718  uint16x4_t __ret_281; \
47719  __ret_281 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_281, __p3_281), __rev0_281, __p1_281); \
47720  __ret_281 = __builtin_shufflevector(__ret_281, __ret_281, 3, 2, 1, 0); \
47721  __ret_281; \
47722 })
47723 #endif
47724 
47725 #ifdef __LITTLE_ENDIAN__
47726 #define vcopy_lane_s8(__p0_282, __p1_282, __p2_282, __p3_282) __extension__ ({ \
47727  int8x8_t __s0_282 = __p0_282; \
47728  int8x8_t __s2_282 = __p2_282; \
47729  int8x8_t __ret_282; \
47730  __ret_282 = vset_lane_s8(vget_lane_s8(__s2_282, __p3_282), __s0_282, __p1_282); \
47731  __ret_282; \
47732 })
47733 #else
47734 #define vcopy_lane_s8(__p0_283, __p1_283, __p2_283, __p3_283) __extension__ ({ \
47735  int8x8_t __s0_283 = __p0_283; \
47736  int8x8_t __s2_283 = __p2_283; \
47737  int8x8_t __rev0_283; __rev0_283 = __builtin_shufflevector(__s0_283, __s0_283, 7, 6, 5, 4, 3, 2, 1, 0); \
47738  int8x8_t __rev2_283; __rev2_283 = __builtin_shufflevector(__s2_283, __s2_283, 7, 6, 5, 4, 3, 2, 1, 0); \
47739  int8x8_t __ret_283; \
47740  __ret_283 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_283, __p3_283), __rev0_283, __p1_283); \
47741  __ret_283 = __builtin_shufflevector(__ret_283, __ret_283, 7, 6, 5, 4, 3, 2, 1, 0); \
47742  __ret_283; \
47743 })
47744 #endif
47745 
47746 #ifdef __LITTLE_ENDIAN__
47747 #define vcopy_lane_f32(__p0_284, __p1_284, __p2_284, __p3_284) __extension__ ({ \
47748  float32x2_t __s0_284 = __p0_284; \
47749  float32x2_t __s2_284 = __p2_284; \
47750  float32x2_t __ret_284; \
47751  __ret_284 = vset_lane_f32(vget_lane_f32(__s2_284, __p3_284), __s0_284, __p1_284); \
47752  __ret_284; \
47753 })
47754 #else
47755 #define vcopy_lane_f32(__p0_285, __p1_285, __p2_285, __p3_285) __extension__ ({ \
47756  float32x2_t __s0_285 = __p0_285; \
47757  float32x2_t __s2_285 = __p2_285; \
47758  float32x2_t __rev0_285; __rev0_285 = __builtin_shufflevector(__s0_285, __s0_285, 1, 0); \
47759  float32x2_t __rev2_285; __rev2_285 = __builtin_shufflevector(__s2_285, __s2_285, 1, 0); \
47760  float32x2_t __ret_285; \
47761  __ret_285 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_285, __p3_285), __rev0_285, __p1_285); \
47762  __ret_285 = __builtin_shufflevector(__ret_285, __ret_285, 1, 0); \
47763  __ret_285; \
47764 })
47765 #endif
47766 
47767 #ifdef __LITTLE_ENDIAN__
47768 #define vcopy_lane_s32(__p0_286, __p1_286, __p2_286, __p3_286) __extension__ ({ \
47769  int32x2_t __s0_286 = __p0_286; \
47770  int32x2_t __s2_286 = __p2_286; \
47771  int32x2_t __ret_286; \
47772  __ret_286 = vset_lane_s32(vget_lane_s32(__s2_286, __p3_286), __s0_286, __p1_286); \
47773  __ret_286; \
47774 })
47775 #else
47776 #define vcopy_lane_s32(__p0_287, __p1_287, __p2_287, __p3_287) __extension__ ({ \
47777  int32x2_t __s0_287 = __p0_287; \
47778  int32x2_t __s2_287 = __p2_287; \
47779  int32x2_t __rev0_287; __rev0_287 = __builtin_shufflevector(__s0_287, __s0_287, 1, 0); \
47780  int32x2_t __rev2_287; __rev2_287 = __builtin_shufflevector(__s2_287, __s2_287, 1, 0); \
47781  int32x2_t __ret_287; \
47782  __ret_287 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_287, __p3_287), __rev0_287, __p1_287); \
47783  __ret_287 = __builtin_shufflevector(__ret_287, __ret_287, 1, 0); \
47784  __ret_287; \
47785 })
47786 #endif
47787 
47788 #define vcopy_lane_s64(__p0_288, __p1_288, __p2_288, __p3_288) __extension__ ({ \
47789  int64x1_t __s0_288 = __p0_288; \
47790  int64x1_t __s2_288 = __p2_288; \
47791  int64x1_t __ret_288; \
47792  __ret_288 = vset_lane_s64(vget_lane_s64(__s2_288, __p3_288), __s0_288, __p1_288); \
47793  __ret_288; \
47794 })
47795 #ifdef __LITTLE_ENDIAN__
47796 #define vcopy_lane_s16(__p0_289, __p1_289, __p2_289, __p3_289) __extension__ ({ \
47797  int16x4_t __s0_289 = __p0_289; \
47798  int16x4_t __s2_289 = __p2_289; \
47799  int16x4_t __ret_289; \
47800  __ret_289 = vset_lane_s16(vget_lane_s16(__s2_289, __p3_289), __s0_289, __p1_289); \
47801  __ret_289; \
47802 })
47803 #else
47804 #define vcopy_lane_s16(__p0_290, __p1_290, __p2_290, __p3_290) __extension__ ({ \
47805  int16x4_t __s0_290 = __p0_290; \
47806  int16x4_t __s2_290 = __p2_290; \
47807  int16x4_t __rev0_290; __rev0_290 = __builtin_shufflevector(__s0_290, __s0_290, 3, 2, 1, 0); \
47808  int16x4_t __rev2_290; __rev2_290 = __builtin_shufflevector(__s2_290, __s2_290, 3, 2, 1, 0); \
47809  int16x4_t __ret_290; \
47810  __ret_290 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_290, __p3_290), __rev0_290, __p1_290); \
47811  __ret_290 = __builtin_shufflevector(__ret_290, __ret_290, 3, 2, 1, 0); \
47812  __ret_290; \
47813 })
47814 #endif
47815 
47816 #ifdef __LITTLE_ENDIAN__
47817 #define vcopyq_laneq_p8(__p0_291, __p1_291, __p2_291, __p3_291) __extension__ ({ \
47818  poly8x16_t __s0_291 = __p0_291; \
47819  poly8x16_t __s2_291 = __p2_291; \
47820  poly8x16_t __ret_291; \
47821  __ret_291 = vsetq_lane_p8(vgetq_lane_p8(__s2_291, __p3_291), __s0_291, __p1_291); \
47822  __ret_291; \
47823 })
47824 #else
47825 #define vcopyq_laneq_p8(__p0_292, __p1_292, __p2_292, __p3_292) __extension__ ({ \
47826  poly8x16_t __s0_292 = __p0_292; \
47827  poly8x16_t __s2_292 = __p2_292; \
47828  poly8x16_t __rev0_292; __rev0_292 = __builtin_shufflevector(__s0_292, __s0_292, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47829  poly8x16_t __rev2_292; __rev2_292 = __builtin_shufflevector(__s2_292, __s2_292, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47830  poly8x16_t __ret_292; \
47831  __ret_292 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_292, __p3_292), __rev0_292, __p1_292); \
47832  __ret_292 = __builtin_shufflevector(__ret_292, __ret_292, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47833  __ret_292; \
47834 })
47835 #endif
47836 
47837 #ifdef __LITTLE_ENDIAN__
47838 #define vcopyq_laneq_p16(__p0_293, __p1_293, __p2_293, __p3_293) __extension__ ({ \
47839  poly16x8_t __s0_293 = __p0_293; \
47840  poly16x8_t __s2_293 = __p2_293; \
47841  poly16x8_t __ret_293; \
47842  __ret_293 = vsetq_lane_p16(vgetq_lane_p16(__s2_293, __p3_293), __s0_293, __p1_293); \
47843  __ret_293; \
47844 })
47845 #else
47846 #define vcopyq_laneq_p16(__p0_294, __p1_294, __p2_294, __p3_294) __extension__ ({ \
47847  poly16x8_t __s0_294 = __p0_294; \
47848  poly16x8_t __s2_294 = __p2_294; \
47849  poly16x8_t __rev0_294; __rev0_294 = __builtin_shufflevector(__s0_294, __s0_294, 7, 6, 5, 4, 3, 2, 1, 0); \
47850  poly16x8_t __rev2_294; __rev2_294 = __builtin_shufflevector(__s2_294, __s2_294, 7, 6, 5, 4, 3, 2, 1, 0); \
47851  poly16x8_t __ret_294; \
47852  __ret_294 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_294, __p3_294), __rev0_294, __p1_294); \
47853  __ret_294 = __builtin_shufflevector(__ret_294, __ret_294, 7, 6, 5, 4, 3, 2, 1, 0); \
47854  __ret_294; \
47855 })
47856 #endif
47857 
47858 #ifdef __LITTLE_ENDIAN__
47859 #define vcopyq_laneq_u8(__p0_295, __p1_295, __p2_295, __p3_295) __extension__ ({ \
47860  uint8x16_t __s0_295 = __p0_295; \
47861  uint8x16_t __s2_295 = __p2_295; \
47862  uint8x16_t __ret_295; \
47863  __ret_295 = vsetq_lane_u8(vgetq_lane_u8(__s2_295, __p3_295), __s0_295, __p1_295); \
47864  __ret_295; \
47865 })
47866 #else
47867 #define vcopyq_laneq_u8(__p0_296, __p1_296, __p2_296, __p3_296) __extension__ ({ \
47868  uint8x16_t __s0_296 = __p0_296; \
47869  uint8x16_t __s2_296 = __p2_296; \
47870  uint8x16_t __rev0_296; __rev0_296 = __builtin_shufflevector(__s0_296, __s0_296, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47871  uint8x16_t __rev2_296; __rev2_296 = __builtin_shufflevector(__s2_296, __s2_296, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47872  uint8x16_t __ret_296; \
47873  __ret_296 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_296, __p3_296), __rev0_296, __p1_296); \
47874  __ret_296 = __builtin_shufflevector(__ret_296, __ret_296, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47875  __ret_296; \
47876 })
47877 #endif
47878 
47879 #ifdef __LITTLE_ENDIAN__
47880 #define vcopyq_laneq_u32(__p0_297, __p1_297, __p2_297, __p3_297) __extension__ ({ \
47881  uint32x4_t __s0_297 = __p0_297; \
47882  uint32x4_t __s2_297 = __p2_297; \
47883  uint32x4_t __ret_297; \
47884  __ret_297 = vsetq_lane_u32(vgetq_lane_u32(__s2_297, __p3_297), __s0_297, __p1_297); \
47885  __ret_297; \
47886 })
47887 #else
47888 #define vcopyq_laneq_u32(__p0_298, __p1_298, __p2_298, __p3_298) __extension__ ({ \
47889  uint32x4_t __s0_298 = __p0_298; \
47890  uint32x4_t __s2_298 = __p2_298; \
47891  uint32x4_t __rev0_298; __rev0_298 = __builtin_shufflevector(__s0_298, __s0_298, 3, 2, 1, 0); \
47892  uint32x4_t __rev2_298; __rev2_298 = __builtin_shufflevector(__s2_298, __s2_298, 3, 2, 1, 0); \
47893  uint32x4_t __ret_298; \
47894  __ret_298 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_298, __p3_298), __rev0_298, __p1_298); \
47895  __ret_298 = __builtin_shufflevector(__ret_298, __ret_298, 3, 2, 1, 0); \
47896  __ret_298; \
47897 })
47898 #endif
47899 
47900 #ifdef __LITTLE_ENDIAN__
47901 #define vcopyq_laneq_u64(__p0_299, __p1_299, __p2_299, __p3_299) __extension__ ({ \
47902  uint64x2_t __s0_299 = __p0_299; \
47903  uint64x2_t __s2_299 = __p2_299; \
47904  uint64x2_t __ret_299; \
47905  __ret_299 = vsetq_lane_u64(vgetq_lane_u64(__s2_299, __p3_299), __s0_299, __p1_299); \
47906  __ret_299; \
47907 })
47908 #else
47909 #define vcopyq_laneq_u64(__p0_300, __p1_300, __p2_300, __p3_300) __extension__ ({ \
47910  uint64x2_t __s0_300 = __p0_300; \
47911  uint64x2_t __s2_300 = __p2_300; \
47912  uint64x2_t __rev0_300; __rev0_300 = __builtin_shufflevector(__s0_300, __s0_300, 1, 0); \
47913  uint64x2_t __rev2_300; __rev2_300 = __builtin_shufflevector(__s2_300, __s2_300, 1, 0); \
47914  uint64x2_t __ret_300; \
47915  __ret_300 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_300, __p3_300), __rev0_300, __p1_300); \
47916  __ret_300 = __builtin_shufflevector(__ret_300, __ret_300, 1, 0); \
47917  __ret_300; \
47918 })
47919 #endif
47920 
47921 #ifdef __LITTLE_ENDIAN__
47922 #define vcopyq_laneq_u16(__p0_301, __p1_301, __p2_301, __p3_301) __extension__ ({ \
47923  uint16x8_t __s0_301 = __p0_301; \
47924  uint16x8_t __s2_301 = __p2_301; \
47925  uint16x8_t __ret_301; \
47926  __ret_301 = vsetq_lane_u16(vgetq_lane_u16(__s2_301, __p3_301), __s0_301, __p1_301); \
47927  __ret_301; \
47928 })
47929 #else
47930 #define vcopyq_laneq_u16(__p0_302, __p1_302, __p2_302, __p3_302) __extension__ ({ \
47931  uint16x8_t __s0_302 = __p0_302; \
47932  uint16x8_t __s2_302 = __p2_302; \
47933  uint16x8_t __rev0_302; __rev0_302 = __builtin_shufflevector(__s0_302, __s0_302, 7, 6, 5, 4, 3, 2, 1, 0); \
47934  uint16x8_t __rev2_302; __rev2_302 = __builtin_shufflevector(__s2_302, __s2_302, 7, 6, 5, 4, 3, 2, 1, 0); \
47935  uint16x8_t __ret_302; \
47936  __ret_302 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_302, __p3_302), __rev0_302, __p1_302); \
47937  __ret_302 = __builtin_shufflevector(__ret_302, __ret_302, 7, 6, 5, 4, 3, 2, 1, 0); \
47938  __ret_302; \
47939 })
47940 #endif
47941 
47942 #ifdef __LITTLE_ENDIAN__
47943 #define vcopyq_laneq_s8(__p0_303, __p1_303, __p2_303, __p3_303) __extension__ ({ \
47944  int8x16_t __s0_303 = __p0_303; \
47945  int8x16_t __s2_303 = __p2_303; \
47946  int8x16_t __ret_303; \
47947  __ret_303 = vsetq_lane_s8(vgetq_lane_s8(__s2_303, __p3_303), __s0_303, __p1_303); \
47948  __ret_303; \
47949 })
47950 #else
47951 #define vcopyq_laneq_s8(__p0_304, __p1_304, __p2_304, __p3_304) __extension__ ({ \
47952  int8x16_t __s0_304 = __p0_304; \
47953  int8x16_t __s2_304 = __p2_304; \
47954  int8x16_t __rev0_304; __rev0_304 = __builtin_shufflevector(__s0_304, __s0_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47955  int8x16_t __rev2_304; __rev2_304 = __builtin_shufflevector(__s2_304, __s2_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47956  int8x16_t __ret_304; \
47957  __ret_304 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_304, __p3_304), __rev0_304, __p1_304); \
47958  __ret_304 = __builtin_shufflevector(__ret_304, __ret_304, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47959  __ret_304; \
47960 })
47961 #endif
47962 
47963 #ifdef __LITTLE_ENDIAN__
47964 #define vcopyq_laneq_f32(__p0_305, __p1_305, __p2_305, __p3_305) __extension__ ({ \
47965  float32x4_t __s0_305 = __p0_305; \
47966  float32x4_t __s2_305 = __p2_305; \
47967  float32x4_t __ret_305; \
47968  __ret_305 = vsetq_lane_f32(vgetq_lane_f32(__s2_305, __p3_305), __s0_305, __p1_305); \
47969  __ret_305; \
47970 })
47971 #else
47972 #define vcopyq_laneq_f32(__p0_306, __p1_306, __p2_306, __p3_306) __extension__ ({ \
47973  float32x4_t __s0_306 = __p0_306; \
47974  float32x4_t __s2_306 = __p2_306; \
47975  float32x4_t __rev0_306; __rev0_306 = __builtin_shufflevector(__s0_306, __s0_306, 3, 2, 1, 0); \
47976  float32x4_t __rev2_306; __rev2_306 = __builtin_shufflevector(__s2_306, __s2_306, 3, 2, 1, 0); \
47977  float32x4_t __ret_306; \
47978  __ret_306 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_306, __p3_306), __rev0_306, __p1_306); \
47979  __ret_306 = __builtin_shufflevector(__ret_306, __ret_306, 3, 2, 1, 0); \
47980  __ret_306; \
47981 })
47982 #endif
47983 
47984 #ifdef __LITTLE_ENDIAN__
47985 #define vcopyq_laneq_s32(__p0_307, __p1_307, __p2_307, __p3_307) __extension__ ({ \
47986  int32x4_t __s0_307 = __p0_307; \
47987  int32x4_t __s2_307 = __p2_307; \
47988  int32x4_t __ret_307; \
47989  __ret_307 = vsetq_lane_s32(vgetq_lane_s32(__s2_307, __p3_307), __s0_307, __p1_307); \
47990  __ret_307; \
47991 })
47992 #else
47993 #define vcopyq_laneq_s32(__p0_308, __p1_308, __p2_308, __p3_308) __extension__ ({ \
47994  int32x4_t __s0_308 = __p0_308; \
47995  int32x4_t __s2_308 = __p2_308; \
47996  int32x4_t __rev0_308; __rev0_308 = __builtin_shufflevector(__s0_308, __s0_308, 3, 2, 1, 0); \
47997  int32x4_t __rev2_308; __rev2_308 = __builtin_shufflevector(__s2_308, __s2_308, 3, 2, 1, 0); \
47998  int32x4_t __ret_308; \
47999  __ret_308 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_308, __p3_308), __rev0_308, __p1_308); \
48000  __ret_308 = __builtin_shufflevector(__ret_308, __ret_308, 3, 2, 1, 0); \
48001  __ret_308; \
48002 })
48003 #endif
48004 
48005 #ifdef __LITTLE_ENDIAN__
48006 #define vcopyq_laneq_s64(__p0_309, __p1_309, __p2_309, __p3_309) __extension__ ({ \
48007  int64x2_t __s0_309 = __p0_309; \
48008  int64x2_t __s2_309 = __p2_309; \
48009  int64x2_t __ret_309; \
48010  __ret_309 = vsetq_lane_s64(vgetq_lane_s64(__s2_309, __p3_309), __s0_309, __p1_309); \
48011  __ret_309; \
48012 })
48013 #else
48014 #define vcopyq_laneq_s64(__p0_310, __p1_310, __p2_310, __p3_310) __extension__ ({ \
48015  int64x2_t __s0_310 = __p0_310; \
48016  int64x2_t __s2_310 = __p2_310; \
48017  int64x2_t __rev0_310; __rev0_310 = __builtin_shufflevector(__s0_310, __s0_310, 1, 0); \
48018  int64x2_t __rev2_310; __rev2_310 = __builtin_shufflevector(__s2_310, __s2_310, 1, 0); \
48019  int64x2_t __ret_310; \
48020  __ret_310 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_310, __p3_310), __rev0_310, __p1_310); \
48021  __ret_310 = __builtin_shufflevector(__ret_310, __ret_310, 1, 0); \
48022  __ret_310; \
48023 })
48024 #endif
48025 
48026 #ifdef __LITTLE_ENDIAN__
48027 #define vcopyq_laneq_s16(__p0_311, __p1_311, __p2_311, __p3_311) __extension__ ({ \
48028  int16x8_t __s0_311 = __p0_311; \
48029  int16x8_t __s2_311 = __p2_311; \
48030  int16x8_t __ret_311; \
48031  __ret_311 = vsetq_lane_s16(vgetq_lane_s16(__s2_311, __p3_311), __s0_311, __p1_311); \
48032  __ret_311; \
48033 })
48034 #else
48035 #define vcopyq_laneq_s16(__p0_312, __p1_312, __p2_312, __p3_312) __extension__ ({ \
48036  int16x8_t __s0_312 = __p0_312; \
48037  int16x8_t __s2_312 = __p2_312; \
48038  int16x8_t __rev0_312; __rev0_312 = __builtin_shufflevector(__s0_312, __s0_312, 7, 6, 5, 4, 3, 2, 1, 0); \
48039  int16x8_t __rev2_312; __rev2_312 = __builtin_shufflevector(__s2_312, __s2_312, 7, 6, 5, 4, 3, 2, 1, 0); \
48040  int16x8_t __ret_312; \
48041  __ret_312 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_312, __p3_312), __rev0_312, __p1_312); \
48042  __ret_312 = __builtin_shufflevector(__ret_312, __ret_312, 7, 6, 5, 4, 3, 2, 1, 0); \
48043  __ret_312; \
48044 })
48045 #endif
48046 
48047 #ifdef __LITTLE_ENDIAN__
48048 #define vcopy_laneq_p8(__p0_313, __p1_313, __p2_313, __p3_313) __extension__ ({ \
48049  poly8x8_t __s0_313 = __p0_313; \
48050  poly8x16_t __s2_313 = __p2_313; \
48051  poly8x8_t __ret_313; \
48052  __ret_313 = vset_lane_p8(vgetq_lane_p8(__s2_313, __p3_313), __s0_313, __p1_313); \
48053  __ret_313; \
48054 })
48055 #else
48056 #define vcopy_laneq_p8(__p0_314, __p1_314, __p2_314, __p3_314) __extension__ ({ \
48057  poly8x8_t __s0_314 = __p0_314; \
48058  poly8x16_t __s2_314 = __p2_314; \
48059  poly8x8_t __rev0_314; __rev0_314 = __builtin_shufflevector(__s0_314, __s0_314, 7, 6, 5, 4, 3, 2, 1, 0); \
48060  poly8x16_t __rev2_314; __rev2_314 = __builtin_shufflevector(__s2_314, __s2_314, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48061  poly8x8_t __ret_314; \
48062  __ret_314 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_314, __p3_314), __rev0_314, __p1_314); \
48063  __ret_314 = __builtin_shufflevector(__ret_314, __ret_314, 7, 6, 5, 4, 3, 2, 1, 0); \
48064  __ret_314; \
48065 })
48066 #endif
48067 
48068 #ifdef __LITTLE_ENDIAN__
48069 #define vcopy_laneq_p16(__p0_315, __p1_315, __p2_315, __p3_315) __extension__ ({ \
48070  poly16x4_t __s0_315 = __p0_315; \
48071  poly16x8_t __s2_315 = __p2_315; \
48072  poly16x4_t __ret_315; \
48073  __ret_315 = vset_lane_p16(vgetq_lane_p16(__s2_315, __p3_315), __s0_315, __p1_315); \
48074  __ret_315; \
48075 })
48076 #else
48077 #define vcopy_laneq_p16(__p0_316, __p1_316, __p2_316, __p3_316) __extension__ ({ \
48078  poly16x4_t __s0_316 = __p0_316; \
48079  poly16x8_t __s2_316 = __p2_316; \
48080  poly16x4_t __rev0_316; __rev0_316 = __builtin_shufflevector(__s0_316, __s0_316, 3, 2, 1, 0); \
48081  poly16x8_t __rev2_316; __rev2_316 = __builtin_shufflevector(__s2_316, __s2_316, 7, 6, 5, 4, 3, 2, 1, 0); \
48082  poly16x4_t __ret_316; \
48083  __ret_316 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_316, __p3_316), __rev0_316, __p1_316); \
48084  __ret_316 = __builtin_shufflevector(__ret_316, __ret_316, 3, 2, 1, 0); \
48085  __ret_316; \
48086 })
48087 #endif
48088 
48089 #ifdef __LITTLE_ENDIAN__
48090 #define vcopy_laneq_u8(__p0_317, __p1_317, __p2_317, __p3_317) __extension__ ({ \
48091  uint8x8_t __s0_317 = __p0_317; \
48092  uint8x16_t __s2_317 = __p2_317; \
48093  uint8x8_t __ret_317; \
48094  __ret_317 = vset_lane_u8(vgetq_lane_u8(__s2_317, __p3_317), __s0_317, __p1_317); \
48095  __ret_317; \
48096 })
48097 #else
48098 #define vcopy_laneq_u8(__p0_318, __p1_318, __p2_318, __p3_318) __extension__ ({ \
48099  uint8x8_t __s0_318 = __p0_318; \
48100  uint8x16_t __s2_318 = __p2_318; \
48101  uint8x8_t __rev0_318; __rev0_318 = __builtin_shufflevector(__s0_318, __s0_318, 7, 6, 5, 4, 3, 2, 1, 0); \
48102  uint8x16_t __rev2_318; __rev2_318 = __builtin_shufflevector(__s2_318, __s2_318, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48103  uint8x8_t __ret_318; \
48104  __ret_318 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_318, __p3_318), __rev0_318, __p1_318); \
48105  __ret_318 = __builtin_shufflevector(__ret_318, __ret_318, 7, 6, 5, 4, 3, 2, 1, 0); \
48106  __ret_318; \
48107 })
48108 #endif
48109 
48110 #ifdef __LITTLE_ENDIAN__
48111 #define vcopy_laneq_u32(__p0_319, __p1_319, __p2_319, __p3_319) __extension__ ({ \
48112  uint32x2_t __s0_319 = __p0_319; \
48113  uint32x4_t __s2_319 = __p2_319; \
48114  uint32x2_t __ret_319; \
48115  __ret_319 = vset_lane_u32(vgetq_lane_u32(__s2_319, __p3_319), __s0_319, __p1_319); \
48116  __ret_319; \
48117 })
48118 #else
48119 #define vcopy_laneq_u32(__p0_320, __p1_320, __p2_320, __p3_320) __extension__ ({ \
48120  uint32x2_t __s0_320 = __p0_320; \
48121  uint32x4_t __s2_320 = __p2_320; \
48122  uint32x2_t __rev0_320; __rev0_320 = __builtin_shufflevector(__s0_320, __s0_320, 1, 0); \
48123  uint32x4_t __rev2_320; __rev2_320 = __builtin_shufflevector(__s2_320, __s2_320, 3, 2, 1, 0); \
48124  uint32x2_t __ret_320; \
48125  __ret_320 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_320, __p3_320), __rev0_320, __p1_320); \
48126  __ret_320 = __builtin_shufflevector(__ret_320, __ret_320, 1, 0); \
48127  __ret_320; \
48128 })
48129 #endif
48130 
48131 #ifdef __LITTLE_ENDIAN__
48132 #define vcopy_laneq_u64(__p0_321, __p1_321, __p2_321, __p3_321) __extension__ ({ \
48133  uint64x1_t __s0_321 = __p0_321; \
48134  uint64x2_t __s2_321 = __p2_321; \
48135  uint64x1_t __ret_321; \
48136  __ret_321 = vset_lane_u64(vgetq_lane_u64(__s2_321, __p3_321), __s0_321, __p1_321); \
48137  __ret_321; \
48138 })
48139 #else
48140 #define vcopy_laneq_u64(__p0_322, __p1_322, __p2_322, __p3_322) __extension__ ({ \
48141  uint64x1_t __s0_322 = __p0_322; \
48142  uint64x2_t __s2_322 = __p2_322; \
48143  uint64x2_t __rev2_322; __rev2_322 = __builtin_shufflevector(__s2_322, __s2_322, 1, 0); \
48144  uint64x1_t __ret_322; \
48145  __ret_322 = vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_322, __p3_322), __s0_322, __p1_322); \
48146  __ret_322; \
48147 })
48148 #endif
48149 
48150 #ifdef __LITTLE_ENDIAN__
48151 #define vcopy_laneq_u16(__p0_323, __p1_323, __p2_323, __p3_323) __extension__ ({ \
48152  uint16x4_t __s0_323 = __p0_323; \
48153  uint16x8_t __s2_323 = __p2_323; \
48154  uint16x4_t __ret_323; \
48155  __ret_323 = vset_lane_u16(vgetq_lane_u16(__s2_323, __p3_323), __s0_323, __p1_323); \
48156  __ret_323; \
48157 })
48158 #else
48159 #define vcopy_laneq_u16(__p0_324, __p1_324, __p2_324, __p3_324) __extension__ ({ \
48160  uint16x4_t __s0_324 = __p0_324; \
48161  uint16x8_t __s2_324 = __p2_324; \
48162  uint16x4_t __rev0_324; __rev0_324 = __builtin_shufflevector(__s0_324, __s0_324, 3, 2, 1, 0); \
48163  uint16x8_t __rev2_324; __rev2_324 = __builtin_shufflevector(__s2_324, __s2_324, 7, 6, 5, 4, 3, 2, 1, 0); \
48164  uint16x4_t __ret_324; \
48165  __ret_324 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_324, __p3_324), __rev0_324, __p1_324); \
48166  __ret_324 = __builtin_shufflevector(__ret_324, __ret_324, 3, 2, 1, 0); \
48167  __ret_324; \
48168 })
48169 #endif
48170 
48171 #ifdef __LITTLE_ENDIAN__
48172 #define vcopy_laneq_s8(__p0_325, __p1_325, __p2_325, __p3_325) __extension__ ({ \
48173  int8x8_t __s0_325 = __p0_325; \
48174  int8x16_t __s2_325 = __p2_325; \
48175  int8x8_t __ret_325; \
48176  __ret_325 = vset_lane_s8(vgetq_lane_s8(__s2_325, __p3_325), __s0_325, __p1_325); \
48177  __ret_325; \
48178 })
48179 #else
48180 #define vcopy_laneq_s8(__p0_326, __p1_326, __p2_326, __p3_326) __extension__ ({ \
48181  int8x8_t __s0_326 = __p0_326; \
48182  int8x16_t __s2_326 = __p2_326; \
48183  int8x8_t __rev0_326; __rev0_326 = __builtin_shufflevector(__s0_326, __s0_326, 7, 6, 5, 4, 3, 2, 1, 0); \
48184  int8x16_t __rev2_326; __rev2_326 = __builtin_shufflevector(__s2_326, __s2_326, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48185  int8x8_t __ret_326; \
48186  __ret_326 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_326, __p3_326), __rev0_326, __p1_326); \
48187  __ret_326 = __builtin_shufflevector(__ret_326, __ret_326, 7, 6, 5, 4, 3, 2, 1, 0); \
48188  __ret_326; \
48189 })
48190 #endif
48191 
48192 #ifdef __LITTLE_ENDIAN__
48193 #define vcopy_laneq_f32(__p0_327, __p1_327, __p2_327, __p3_327) __extension__ ({ \
48194  float32x2_t __s0_327 = __p0_327; \
48195  float32x4_t __s2_327 = __p2_327; \
48196  float32x2_t __ret_327; \
48197  __ret_327 = vset_lane_f32(vgetq_lane_f32(__s2_327, __p3_327), __s0_327, __p1_327); \
48198  __ret_327; \
48199 })
48200 #else
48201 #define vcopy_laneq_f32(__p0_328, __p1_328, __p2_328, __p3_328) __extension__ ({ \
48202  float32x2_t __s0_328 = __p0_328; \
48203  float32x4_t __s2_328 = __p2_328; \
48204  float32x2_t __rev0_328; __rev0_328 = __builtin_shufflevector(__s0_328, __s0_328, 1, 0); \
48205  float32x4_t __rev2_328; __rev2_328 = __builtin_shufflevector(__s2_328, __s2_328, 3, 2, 1, 0); \
48206  float32x2_t __ret_328; \
48207  __ret_328 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_328, __p3_328), __rev0_328, __p1_328); \
48208  __ret_328 = __builtin_shufflevector(__ret_328, __ret_328, 1, 0); \
48209  __ret_328; \
48210 })
48211 #endif
48212 
48213 #ifdef __LITTLE_ENDIAN__
48214 #define vcopy_laneq_s32(__p0_329, __p1_329, __p2_329, __p3_329) __extension__ ({ \
48215  int32x2_t __s0_329 = __p0_329; \
48216  int32x4_t __s2_329 = __p2_329; \
48217  int32x2_t __ret_329; \
48218  __ret_329 = vset_lane_s32(vgetq_lane_s32(__s2_329, __p3_329), __s0_329, __p1_329); \
48219  __ret_329; \
48220 })
48221 #else
48222 #define vcopy_laneq_s32(__p0_330, __p1_330, __p2_330, __p3_330) __extension__ ({ \
48223  int32x2_t __s0_330 = __p0_330; \
48224  int32x4_t __s2_330 = __p2_330; \
48225  int32x2_t __rev0_330; __rev0_330 = __builtin_shufflevector(__s0_330, __s0_330, 1, 0); \
48226  int32x4_t __rev2_330; __rev2_330 = __builtin_shufflevector(__s2_330, __s2_330, 3, 2, 1, 0); \
48227  int32x2_t __ret_330; \
48228  __ret_330 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_330, __p3_330), __rev0_330, __p1_330); \
48229  __ret_330 = __builtin_shufflevector(__ret_330, __ret_330, 1, 0); \
48230  __ret_330; \
48231 })
48232 #endif
48233 
48234 #ifdef __LITTLE_ENDIAN__
48235 #define vcopy_laneq_s64(__p0_331, __p1_331, __p2_331, __p3_331) __extension__ ({ \
48236  int64x1_t __s0_331 = __p0_331; \
48237  int64x2_t __s2_331 = __p2_331; \
48238  int64x1_t __ret_331; \
48239  __ret_331 = vset_lane_s64(vgetq_lane_s64(__s2_331, __p3_331), __s0_331, __p1_331); \
48240  __ret_331; \
48241 })
48242 #else
48243 #define vcopy_laneq_s64(__p0_332, __p1_332, __p2_332, __p3_332) __extension__ ({ \
48244  int64x1_t __s0_332 = __p0_332; \
48245  int64x2_t __s2_332 = __p2_332; \
48246  int64x2_t __rev2_332; __rev2_332 = __builtin_shufflevector(__s2_332, __s2_332, 1, 0); \
48247  int64x1_t __ret_332; \
48248  __ret_332 = vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_332, __p3_332), __s0_332, __p1_332); \
48249  __ret_332; \
48250 })
48251 #endif
48252 
48253 #ifdef __LITTLE_ENDIAN__
48254 #define vcopy_laneq_s16(__p0_333, __p1_333, __p2_333, __p3_333) __extension__ ({ \
48255  int16x4_t __s0_333 = __p0_333; \
48256  int16x8_t __s2_333 = __p2_333; \
48257  int16x4_t __ret_333; \
48258  __ret_333 = vset_lane_s16(vgetq_lane_s16(__s2_333, __p3_333), __s0_333, __p1_333); \
48259  __ret_333; \
48260 })
48261 #else
48262 #define vcopy_laneq_s16(__p0_334, __p1_334, __p2_334, __p3_334) __extension__ ({ \
48263  int16x4_t __s0_334 = __p0_334; \
48264  int16x8_t __s2_334 = __p2_334; \
48265  int16x4_t __rev0_334; __rev0_334 = __builtin_shufflevector(__s0_334, __s0_334, 3, 2, 1, 0); \
48266  int16x8_t __rev2_334; __rev2_334 = __builtin_shufflevector(__s2_334, __s2_334, 7, 6, 5, 4, 3, 2, 1, 0); \
48267  int16x4_t __ret_334; \
48268  __ret_334 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_334, __p3_334), __rev0_334, __p1_334); \
48269  __ret_334 = __builtin_shufflevector(__ret_334, __ret_334, 3, 2, 1, 0); \
48270  __ret_334; \
48271 })
48272 #endif
48273 
48274 #define vcreate_p64(__p0) __extension__ ({ \
48275  poly64x1_t __ret; \
48276  uint64_t __promote = __p0; \
48277  __ret = (poly64x1_t)(__promote); \
48278  __ret; \
48279 })
48280 #define vcreate_f64(__p0) __extension__ ({ \
48281  float64x1_t __ret; \
48282  uint64_t __promote = __p0; \
48283  __ret = (float64x1_t)(__promote); \
48284  __ret; \
48285 })
48286 __ai float32_t vcvts_f32_s32(int32_t __p0) {
48287  float32_t __ret;
48288  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
48289  return __ret;
48290 }
48291 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
48292  float32_t __ret;
48293  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
48294  return __ret;
48295 }
48296 #ifdef __LITTLE_ENDIAN__
48297 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
48298  float32x2_t __ret;
48299  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
48300  return __ret;
48301 }
48302 #else
48303 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
48304  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48305  float32x2_t __ret;
48306  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
48307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48308  return __ret;
48309 }
48310 __ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
48311  float32x2_t __ret;
48312  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
48313  return __ret;
48314 }
48315 #endif
48316 
48317 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
48318  float64_t __ret;
48319  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
48320  return __ret;
48321 }
48322 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
48323  float64_t __ret;
48324  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
48325  return __ret;
48326 }
48327 #ifdef __LITTLE_ENDIAN__
48328 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
48329  float64x2_t __ret;
48330  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
48331  return __ret;
48332 }
48333 #else
48334 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
48335  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48336  float64x2_t __ret;
48337  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
48338  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48339  return __ret;
48340 }
48341 #endif
48342 
48343 #ifdef __LITTLE_ENDIAN__
48344 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
48345  float64x2_t __ret;
48346  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
48347  return __ret;
48348 }
48349 #else
48350 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
48351  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48352  float64x2_t __ret;
48353  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
48354  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48355  return __ret;
48356 }
48357 #endif
48358 
48359 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
48360  float64x1_t __ret;
48361  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
48362  return __ret;
48363 }
48364 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
48365  float64x1_t __ret;
48366  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
48367  return __ret;
48368 }
48369 #ifdef __LITTLE_ENDIAN__
48370 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
48371  float64x2_t __ret;
48372  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
48373  return __ret;
48374 }
48375 #else
48376 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
48377  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48378  float64x2_t __ret;
48379  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
48380  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48381  return __ret;
48382 }
48383 __ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
48384  float64x2_t __ret;
48385  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
48386  return __ret;
48387 }
48388 #endif
48389 
48390 #ifdef __LITTLE_ENDIAN__
48391 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
48392  float16x8_t __ret;
48393  __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
48394  return __ret;
48395 }
48396 #else
48397 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
48398  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48399  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
48400  float16x8_t __ret;
48401  __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
48402  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
48403  return __ret;
48404 }
48405 #endif
48406 
48407 #ifdef __LITTLE_ENDIAN__
48408 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
48409  float32x4_t __ret;
48410  __ret = vcvt_f32_f16(vget_high_f16(__p0));
48411  return __ret;
48412 }
48413 #else
48414 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
48415  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
48416  float32x4_t __ret;
48417  __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
48418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48419  return __ret;
48420 }
48421 #endif
48422 
48423 #ifdef __LITTLE_ENDIAN__
48424 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
48425  float32x4_t __ret;
48426  __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
48427  return __ret;
48428 }
48429 #else
48430 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
48431  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48432  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48433  float32x4_t __ret;
48434  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
48435  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48436  return __ret;
48437 }
48438 #endif
48439 
48440 #ifdef __LITTLE_ENDIAN__
48441 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
48442  float64x2_t __ret;
48443  __ret = vcvt_f64_f32(vget_high_f32(__p0));
48444  return __ret;
48445 }
48446 #else
48447 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
48448  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48449  float64x2_t __ret;
48450  __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
48451  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48452  return __ret;
48453 }
48454 #endif
48455 
48456 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
48457  uint32_t __s0 = __p0; \
48458  float32_t __ret; \
48459  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
48460  __ret; \
48461 })
48462 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
48463  int32_t __s0 = __p0; \
48464  float32_t __ret; \
48465  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
48466  __ret; \
48467 })
48468 #ifdef __LITTLE_ENDIAN__
48469 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
48470  uint64x2_t __s0 = __p0; \
48471  float64x2_t __ret; \
48472  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
48473  __ret; \
48474 })
48475 #else
48476 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
48477  uint64x2_t __s0 = __p0; \
48478  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48479  float64x2_t __ret; \
48480  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
48481  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48482  __ret; \
48483 })
48484 #endif
48485 
48486 #ifdef __LITTLE_ENDIAN__
48487 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
48488  int64x2_t __s0 = __p0; \
48489  float64x2_t __ret; \
48490  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
48491  __ret; \
48492 })
48493 #else
48494 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
48495  int64x2_t __s0 = __p0; \
48496  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48497  float64x2_t __ret; \
48498  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
48499  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48500  __ret; \
48501 })
48502 #endif
48503 
48504 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
48505  uint64x1_t __s0 = __p0; \
48506  float64x1_t __ret; \
48507  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
48508  __ret; \
48509 })
48510 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
48511  int64x1_t __s0 = __p0; \
48512  float64x1_t __ret; \
48513  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
48514  __ret; \
48515 })
48516 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
48517  uint64_t __s0 = __p0; \
48518  float64_t __ret; \
48519  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
48520  __ret; \
48521 })
48522 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
48523  int64_t __s0 = __p0; \
48524  float64_t __ret; \
48525  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
48526  __ret; \
48527 })
48528 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
48529  float32_t __s0 = __p0; \
48530  int32_t __ret; \
48531  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
48532  __ret; \
48533 })
48534 #ifdef __LITTLE_ENDIAN__
48535 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
48536  float64x2_t __s0 = __p0; \
48537  int64x2_t __ret; \
48538  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
48539  __ret; \
48540 })
48541 #else
48542 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
48543  float64x2_t __s0 = __p0; \
48544  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48545  int64x2_t __ret; \
48546  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
48547  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48548  __ret; \
48549 })
48550 #endif
48551 
48552 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
48553  float64x1_t __s0 = __p0; \
48554  int64x1_t __ret; \
48555  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
48556  __ret; \
48557 })
48558 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
48559  float64_t __s0 = __p0; \
48560  int64_t __ret; \
48561  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
48562  __ret; \
48563 })
48564 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
48565  float32_t __s0 = __p0; \
48566  uint32_t __ret; \
48567  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
48568  __ret; \
48569 })
48570 #ifdef __LITTLE_ENDIAN__
48571 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
48572  float64x2_t __s0 = __p0; \
48573  uint64x2_t __ret; \
48574  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
48575  __ret; \
48576 })
48577 #else
48578 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
48579  float64x2_t __s0 = __p0; \
48580  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48581  uint64x2_t __ret; \
48582  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
48583  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48584  __ret; \
48585 })
48586 #endif
48587 
48588 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
48589  float64x1_t __s0 = __p0; \
48590  uint64x1_t __ret; \
48591  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
48592  __ret; \
48593 })
48594 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
48595  float64_t __s0 = __p0; \
48596  uint64_t __ret; \
48597  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
48598  __ret; \
48599 })
48600 __ai int32_t vcvts_s32_f32(float32_t __p0) {
48601  int32_t __ret;
48602  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
48603  return __ret;
48604 }
48605 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
48606  int64_t __ret;
48607  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
48608  return __ret;
48609 }
48610 #ifdef __LITTLE_ENDIAN__
48611 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
48612  int64x2_t __ret;
48613  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
48614  return __ret;
48615 }
48616 #else
48617 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
48618  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48619  int64x2_t __ret;
48620  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
48621  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48622  return __ret;
48623 }
48624 #endif
48625 
48626 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
48627  int64x1_t __ret;
48628  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
48629  return __ret;
48630 }
48631 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
48632  uint32_t __ret;
48633  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
48634  return __ret;
48635 }
48636 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
48637  uint64_t __ret;
48638  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
48639  return __ret;
48640 }
48641 #ifdef __LITTLE_ENDIAN__
48642 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
48643  uint64x2_t __ret;
48644  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
48645  return __ret;
48646 }
48647 #else
48648 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
48649  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48650  uint64x2_t __ret;
48651  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
48652  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48653  return __ret;
48654 }
48655 #endif
48656 
48657 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
48658  uint64x1_t __ret;
48659  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
48660  return __ret;
48661 }
48662 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
48663  int32_t __ret;
48664  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
48665  return __ret;
48666 }
48667 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
48668  int64_t __ret;
48669  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
48670  return __ret;
48671 }
48672 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
48673  uint32_t __ret;
48674  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
48675  return __ret;
48676 }
48677 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
48678  uint64_t __ret;
48679  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
48680  return __ret;
48681 }
48682 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
48683  int32_t __ret;
48684  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
48685  return __ret;
48686 }
48687 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
48688  int64_t __ret;
48689  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
48690  return __ret;
48691 }
48692 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
48693  uint32_t __ret;
48694  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
48695  return __ret;
48696 }
48697 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
48698  uint64_t __ret;
48699  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
48700  return __ret;
48701 }
48702 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
48703  int32_t __ret;
48704  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
48705  return __ret;
48706 }
48707 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
48708  int64_t __ret;
48709  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
48710  return __ret;
48711 }
48712 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
48713  uint32_t __ret;
48714  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
48715  return __ret;
48716 }
48717 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
48718  uint64_t __ret;
48719  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
48720  return __ret;
48721 }
48722 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
48723  int32_t __ret;
48724  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
48725  return __ret;
48726 }
48727 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
48728  int64_t __ret;
48729  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
48730  return __ret;
48731 }
48732 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
48733  uint32_t __ret;
48734  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
48735  return __ret;
48736 }
48737 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
48738  uint64_t __ret;
48739  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
48740  return __ret;
48741 }
48742 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
48743  float32_t __ret;
48744  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
48745  return __ret;
48746 }
48747 #ifdef __LITTLE_ENDIAN__
48748 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
48749  float32x2_t __ret;
48750  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
48751  return __ret;
48752 }
48753 #else
48754 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
48755  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48756  float32x2_t __ret;
48757  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
48758  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48759  return __ret;
48760 }
48761 __ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
48762  float32x2_t __ret;
48763  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
48764  return __ret;
48765 }
48766 #endif
48767 
48768 #ifdef __LITTLE_ENDIAN__
48769 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
48770  float32x4_t __ret;
48771  __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
48772  return __ret;
48773 }
48774 #else
48775 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
48776  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48777  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48778  float32x4_t __ret;
48779  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
48780  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48781  return __ret;
48782 }
48783 #endif
48784 
48785 #ifdef __LITTLE_ENDIAN__
48786 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
48787  float64x2_t __ret;
48788  __ret = __p0 / __p1;
48789  return __ret;
48790 }
48791 #else
48792 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
48793  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48794  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48795  float64x2_t __ret;
48796  __ret = __rev0 / __rev1;
48797  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48798  return __ret;
48799 }
48800 #endif
48801 
48802 #ifdef __LITTLE_ENDIAN__
48803 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
48804  float32x4_t __ret;
48805  __ret = __p0 / __p1;
48806  return __ret;
48807 }
48808 #else
48809 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
48810  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48811  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
48812  float32x4_t __ret;
48813  __ret = __rev0 / __rev1;
48814  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48815  return __ret;
48816 }
48817 #endif
48818 
48819 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
48820  float64x1_t __ret;
48821  __ret = __p0 / __p1;
48822  return __ret;
48823 }
48824 #ifdef __LITTLE_ENDIAN__
48825 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
48826  float32x2_t __ret;
48827  __ret = __p0 / __p1;
48828  return __ret;
48829 }
48830 #else
48831 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
48832  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48833  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48834  float32x2_t __ret;
48835  __ret = __rev0 / __rev1;
48836  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48837  return __ret;
48838 }
48839 #endif
48840 
48841 #ifdef __LITTLE_ENDIAN__
48842 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
48843  poly8x8_t __s0 = __p0; \
48844  poly8_t __ret; \
48845  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((poly8x8_t)__s0, __p1); \
48846  __ret; \
48847 })
48848 #else
48849 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
48850  poly8x8_t __s0 = __p0; \
48851  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
48852  poly8_t __ret; \
48853  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((poly8x8_t)__rev0, __p1); \
48854  __ret; \
48855 })
48856 #endif
48857 
48858 #ifdef __LITTLE_ENDIAN__
48859 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
48860  poly16x4_t __s0 = __p0; \
48861  poly16_t __ret; \
48862  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((poly16x4_t)__s0, __p1); \
48863  __ret; \
48864 })
48865 #else
48866 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
48867  poly16x4_t __s0 = __p0; \
48868  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
48869  poly16_t __ret; \
48870  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((poly16x4_t)__rev0, __p1); \
48871  __ret; \
48872 })
48873 #endif
48874 
48875 #ifdef __LITTLE_ENDIAN__
48876 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
48877  uint8x8_t __s0 = __p0; \
48878  uint8_t __ret; \
48879  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
48880  __ret; \
48881 })
48882 #else
48883 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
48884  uint8x8_t __s0 = __p0; \
48885  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
48886  uint8_t __ret; \
48887  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
48888  __ret; \
48889 })
48890 #endif
48891 
48892 #ifdef __LITTLE_ENDIAN__
48893 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
48894  uint32x2_t __s0 = __p0; \
48895  uint32_t __ret; \
48896  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__s0, __p1); \
48897  __ret; \
48898 })
48899 #else
48900 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
48901  uint32x2_t __s0 = __p0; \
48902  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48903  uint32_t __ret; \
48904  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__rev0, __p1); \
48905  __ret; \
48906 })
48907 #endif
48908 
48909 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
48910  uint64x1_t __s0 = __p0; \
48911  uint64_t __ret; \
48912  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int64x1_t)__s0, __p1); \
48913  __ret; \
48914 })
48915 #ifdef __LITTLE_ENDIAN__
48916 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
48917  uint16x4_t __s0 = __p0; \
48918  uint16_t __ret; \
48919  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__s0, __p1); \
48920  __ret; \
48921 })
48922 #else
48923 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
48924  uint16x4_t __s0 = __p0; \
48925  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
48926  uint16_t __ret; \
48927  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__rev0, __p1); \
48928  __ret; \
48929 })
48930 #endif
48931 
48932 #ifdef __LITTLE_ENDIAN__
48933 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
48934  int8x8_t __s0 = __p0; \
48935  int8_t __ret; \
48936  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
48937  __ret; \
48938 })
48939 #else
48940 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
48941  int8x8_t __s0 = __p0; \
48942  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
48943  int8_t __ret; \
48944  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
48945  __ret; \
48946 })
48947 #endif
48948 
48949 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
48950  float64x1_t __s0 = __p0; \
48951  float64_t __ret; \
48952  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((float64x1_t)__s0, __p1); \
48953  __ret; \
48954 })
48955 #ifdef __LITTLE_ENDIAN__
48956 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
48957  float32x2_t __s0 = __p0; \
48958  float32_t __ret; \
48959  __ret = (float32_t) __builtin_neon_vdups_lane_f32((float32x2_t)__s0, __p1); \
48960  __ret; \
48961 })
48962 #else
48963 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
48964  float32x2_t __s0 = __p0; \
48965  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48966  float32_t __ret; \
48967  __ret = (float32_t) __builtin_neon_vdups_lane_f32((float32x2_t)__rev0, __p1); \
48968  __ret; \
48969 })
48970 #endif
48971 
48972 #ifdef __LITTLE_ENDIAN__
48973 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
48974  int32x2_t __s0 = __p0; \
48975  int32_t __ret; \
48976  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__s0, __p1); \
48977  __ret; \
48978 })
48979 #else
48980 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
48981  int32x2_t __s0 = __p0; \
48982  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48983  int32_t __ret; \
48984  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int32x2_t)__rev0, __p1); \
48985  __ret; \
48986 })
48987 #endif
48988 
48989 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
48990  int64x1_t __s0 = __p0; \
48991  int64_t __ret; \
48992  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int64x1_t)__s0, __p1); \
48993  __ret; \
48994 })
48995 #ifdef __LITTLE_ENDIAN__
48996 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
48997  int16x4_t __s0 = __p0; \
48998  int16_t __ret; \
48999  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__s0, __p1); \
49000  __ret; \
49001 })
49002 #else
49003 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
49004  int16x4_t __s0 = __p0; \
49005  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
49006  int16_t __ret; \
49007  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int16x4_t)__rev0, __p1); \
49008  __ret; \
49009 })
49010 #endif
49011 
49012 #define vdup_lane_p64(__p0_335, __p1_335) __extension__ ({ \
49013  poly64x1_t __s0_335 = __p0_335; \
49014  poly64x1_t __ret_335; \
49015  __ret_335 = splat_lane_p64(__s0_335, __p1_335); \
49016  __ret_335; \
49017 })
49018 #ifdef __LITTLE_ENDIAN__
49019 #define vdupq_lane_p64(__p0_336, __p1_336) __extension__ ({ \
49020  poly64x1_t __s0_336 = __p0_336; \
49021  poly64x2_t __ret_336; \
49022  __ret_336 = splatq_lane_p64(__s0_336, __p1_336); \
49023  __ret_336; \
49024 })
49025 #else
49026 #define vdupq_lane_p64(__p0_337, __p1_337) __extension__ ({ \
49027  poly64x1_t __s0_337 = __p0_337; \
49028  poly64x2_t __ret_337; \
49029  __ret_337 = __noswap_splatq_lane_p64(__s0_337, __p1_337); \
49030  __ret_337 = __builtin_shufflevector(__ret_337, __ret_337, 1, 0); \
49031  __ret_337; \
49032 })
49033 #endif
49034 
49035 #ifdef __LITTLE_ENDIAN__
49036 #define vdupq_lane_f64(__p0_338, __p1_338) __extension__ ({ \
49037  float64x1_t __s0_338 = __p0_338; \
49038  float64x2_t __ret_338; \
49039  __ret_338 = splatq_lane_f64(__s0_338, __p1_338); \
49040  __ret_338; \
49041 })
49042 #else
49043 #define vdupq_lane_f64(__p0_339, __p1_339) __extension__ ({ \
49044  float64x1_t __s0_339 = __p0_339; \
49045  float64x2_t __ret_339; \
49046  __ret_339 = __noswap_splatq_lane_f64(__s0_339, __p1_339); \
49047  __ret_339 = __builtin_shufflevector(__ret_339, __ret_339, 1, 0); \
49048  __ret_339; \
49049 })
49050 #endif
49051 
49052 #ifdef __LITTLE_ENDIAN__
49053 #define vdupq_lane_f16(__p0_340, __p1_340) __extension__ ({ \
49054  float16x4_t __s0_340 = __p0_340; \
49055  float16x8_t __ret_340; \
49056  __ret_340 = splatq_lane_f16(__s0_340, __p1_340); \
49057  __ret_340; \
49058 })
49059 #else
49060 #define vdupq_lane_f16(__p0_341, __p1_341) __extension__ ({ \
49061  float16x4_t __s0_341 = __p0_341; \
49062  float16x4_t __rev0_341; __rev0_341 = __builtin_shufflevector(__s0_341, __s0_341, 3, 2, 1, 0); \
49063  float16x8_t __ret_341; \
49064  __ret_341 = __noswap_splatq_lane_f16(__rev0_341, __p1_341); \
49065  __ret_341 = __builtin_shufflevector(__ret_341, __ret_341, 7, 6, 5, 4, 3, 2, 1, 0); \
49066  __ret_341; \
49067 })
49068 #endif
49069 
49070 #define vdup_lane_f64(__p0_342, __p1_342) __extension__ ({ \
49071  float64x1_t __s0_342 = __p0_342; \
49072  float64x1_t __ret_342; \
49073  __ret_342 = splat_lane_f64(__s0_342, __p1_342); \
49074  __ret_342; \
49075 })
49076 #ifdef __LITTLE_ENDIAN__
49077 #define vdup_lane_f16(__p0_343, __p1_343) __extension__ ({ \
49078  float16x4_t __s0_343 = __p0_343; \
49079  float16x4_t __ret_343; \
49080  __ret_343 = splat_lane_f16(__s0_343, __p1_343); \
49081  __ret_343; \
49082 })
49083 #else
49084 #define vdup_lane_f16(__p0_344, __p1_344) __extension__ ({ \
49085  float16x4_t __s0_344 = __p0_344; \
49086  float16x4_t __rev0_344; __rev0_344 = __builtin_shufflevector(__s0_344, __s0_344, 3, 2, 1, 0); \
49087  float16x4_t __ret_344; \
49088  __ret_344 = __noswap_splat_lane_f16(__rev0_344, __p1_344); \
49089  __ret_344 = __builtin_shufflevector(__ret_344, __ret_344, 3, 2, 1, 0); \
49090  __ret_344; \
49091 })
49092 #endif
49093 
49094 #ifdef __LITTLE_ENDIAN__
49095 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
49096  poly8x16_t __s0 = __p0; \
49097  poly8_t __ret; \
49098  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((poly8x16_t)__s0, __p1); \
49099  __ret; \
49100 })
49101 #else
49102 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
49103  poly8x16_t __s0 = __p0; \
49104  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49105  poly8_t __ret; \
49106  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((poly8x16_t)__rev0, __p1); \
49107  __ret; \
49108 })
49109 #endif
49110 
49111 #ifdef __LITTLE_ENDIAN__
49112 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
49113  poly16x8_t __s0 = __p0; \
49114  poly16_t __ret; \
49115  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((poly16x8_t)__s0, __p1); \
49116  __ret; \
49117 })
49118 #else
49119 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
49120  poly16x8_t __s0 = __p0; \
49121  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
49122  poly16_t __ret; \
49123  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((poly16x8_t)__rev0, __p1); \
49124  __ret; \
49125 })
49126 #endif
49127 
49128 #ifdef __LITTLE_ENDIAN__
49129 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
49130  uint8x16_t __s0 = __p0; \
49131  uint8_t __ret; \
49132  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
49133  __ret; \
49134 })
49135 #else
49136 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
49137  uint8x16_t __s0 = __p0; \
49138  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49139  uint8_t __ret; \
49140  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
49141  __ret; \
49142 })
49143 #endif
49144 
49145 #ifdef __LITTLE_ENDIAN__
49146 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
49147  uint32x4_t __s0 = __p0; \
49148  uint32_t __ret; \
49149  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__s0, __p1); \
49150  __ret; \
49151 })
49152 #else
49153 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
49154  uint32x4_t __s0 = __p0; \
49155  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
49156  uint32_t __ret; \
49157  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__rev0, __p1); \
49158  __ret; \
49159 })
49160 #endif
49161 
49162 #ifdef __LITTLE_ENDIAN__
49163 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
49164  uint64x2_t __s0 = __p0; \
49165  uint64_t __ret; \
49166  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__s0, __p1); \
49167  __ret; \
49168 })
49169 #else
49170 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
49171  uint64x2_t __s0 = __p0; \
49172  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49173  uint64_t __ret; \
49174  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__rev0, __p1); \
49175  __ret; \
49176 })
49177 #endif
49178 
49179 #ifdef __LITTLE_ENDIAN__
49180 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
49181  uint16x8_t __s0 = __p0; \
49182  uint16_t __ret; \
49183  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__s0, __p1); \
49184  __ret; \
49185 })
49186 #else
49187 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
49188  uint16x8_t __s0 = __p0; \
49189  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
49190  uint16_t __ret; \
49191  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__rev0, __p1); \
49192  __ret; \
49193 })
49194 #endif
49195 
49196 #ifdef __LITTLE_ENDIAN__
49197 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
49198  int8x16_t __s0 = __p0; \
49199  int8_t __ret; \
49200  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
49201  __ret; \
49202 })
49203 #else
49204 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
49205  int8x16_t __s0 = __p0; \
49206  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49207  int8_t __ret; \
49208  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
49209  __ret; \
49210 })
49211 #endif
49212 
49213 #ifdef __LITTLE_ENDIAN__
49214 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
49215  float64x2_t __s0 = __p0; \
49216  float64_t __ret; \
49217  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((float64x2_t)__s0, __p1); \
49218  __ret; \
49219 })
49220 #else
49221 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
49222  float64x2_t __s0 = __p0; \
49223  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49224  float64_t __ret; \
49225  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((float64x2_t)__rev0, __p1); \
49226  __ret; \
49227 })
49228 #endif
49229 
49230 #ifdef __LITTLE_ENDIAN__
49231 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
49232  float32x4_t __s0 = __p0; \
49233  float32_t __ret; \
49234  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((float32x4_t)__s0, __p1); \
49235  __ret; \
49236 })
49237 #else
49238 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
49239  float32x4_t __s0 = __p0; \
49240  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
49241  float32_t __ret; \
49242  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((float32x4_t)__rev0, __p1); \
49243  __ret; \
49244 })
49245 #endif
49246 
49247 #ifdef __LITTLE_ENDIAN__
49248 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
49249  int32x4_t __s0 = __p0; \
49250  int32_t __ret; \
49251  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__s0, __p1); \
49252  __ret; \
49253 })
49254 #else
49255 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
49256  int32x4_t __s0 = __p0; \
49257  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
49258  int32_t __ret; \
49259  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int32x4_t)__rev0, __p1); \
49260  __ret; \
49261 })
49262 #endif
49263 
49264 #ifdef __LITTLE_ENDIAN__
49265 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
49266  int64x2_t __s0 = __p0; \
49267  int64_t __ret; \
49268  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__s0, __p1); \
49269  __ret; \
49270 })
49271 #else
49272 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
49273  int64x2_t __s0 = __p0; \
49274  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49275  int64_t __ret; \
49276  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int64x2_t)__rev0, __p1); \
49277  __ret; \
49278 })
49279 #endif
49280 
49281 #ifdef __LITTLE_ENDIAN__
49282 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
49283  int16x8_t __s0 = __p0; \
49284  int16_t __ret; \
49285  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__s0, __p1); \
49286  __ret; \
49287 })
49288 #else
49289 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
49290  int16x8_t __s0 = __p0; \
49291  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
49292  int16_t __ret; \
49293  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int16x8_t)__rev0, __p1); \
49294  __ret; \
49295 })
49296 #endif
49297 
49298 #ifdef __LITTLE_ENDIAN__
49299 #define vdup_laneq_p8(__p0_345, __p1_345) __extension__ ({ \
49300  poly8x16_t __s0_345 = __p0_345; \
49301  poly8x8_t __ret_345; \
49302  __ret_345 = splat_laneq_p8(__s0_345, __p1_345); \
49303  __ret_345; \
49304 })
49305 #else
49306 #define vdup_laneq_p8(__p0_346, __p1_346) __extension__ ({ \
49307  poly8x16_t __s0_346 = __p0_346; \
49308  poly8x16_t __rev0_346; __rev0_346 = __builtin_shufflevector(__s0_346, __s0_346, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49309  poly8x8_t __ret_346; \
49310  __ret_346 = __noswap_splat_laneq_p8(__rev0_346, __p1_346); \
49311  __ret_346 = __builtin_shufflevector(__ret_346, __ret_346, 7, 6, 5, 4, 3, 2, 1, 0); \
49312  __ret_346; \
49313 })
49314 #endif
49315 
49316 #ifdef __LITTLE_ENDIAN__
49317 #define vdup_laneq_p64(__p0_347, __p1_347) __extension__ ({ \
49318  poly64x2_t __s0_347 = __p0_347; \
49319  poly64x1_t __ret_347; \
49320  __ret_347 = splat_laneq_p64(__s0_347, __p1_347); \
49321  __ret_347; \
49322 })
49323 #else
49324 #define vdup_laneq_p64(__p0_348, __p1_348) __extension__ ({ \
49325  poly64x2_t __s0_348 = __p0_348; \
49326  poly64x2_t __rev0_348; __rev0_348 = __builtin_shufflevector(__s0_348, __s0_348, 1, 0); \
49327  poly64x1_t __ret_348; \
49328  __ret_348 = __noswap_splat_laneq_p64(__rev0_348, __p1_348); \
49329  __ret_348; \
49330 })
49331 #endif
49332 
49333 #ifdef __LITTLE_ENDIAN__
49334 #define vdup_laneq_p16(__p0_349, __p1_349) __extension__ ({ \
49335  poly16x8_t __s0_349 = __p0_349; \
49336  poly16x4_t __ret_349; \
49337  __ret_349 = splat_laneq_p16(__s0_349, __p1_349); \
49338  __ret_349; \
49339 })
49340 #else
49341 #define vdup_laneq_p16(__p0_350, __p1_350) __extension__ ({ \
49342  poly16x8_t __s0_350 = __p0_350; \
49343  poly16x8_t __rev0_350; __rev0_350 = __builtin_shufflevector(__s0_350, __s0_350, 7, 6, 5, 4, 3, 2, 1, 0); \
49344  poly16x4_t __ret_350; \
49345  __ret_350 = __noswap_splat_laneq_p16(__rev0_350, __p1_350); \
49346  __ret_350 = __builtin_shufflevector(__ret_350, __ret_350, 3, 2, 1, 0); \
49347  __ret_350; \
49348 })
49349 #endif
49350 
49351 #ifdef __LITTLE_ENDIAN__
49352 #define vdupq_laneq_p8(__p0_351, __p1_351) __extension__ ({ \
49353  poly8x16_t __s0_351 = __p0_351; \
49354  poly8x16_t __ret_351; \
49355  __ret_351 = splatq_laneq_p8(__s0_351, __p1_351); \
49356  __ret_351; \
49357 })
49358 #else
49359 #define vdupq_laneq_p8(__p0_352, __p1_352) __extension__ ({ \
49360  poly8x16_t __s0_352 = __p0_352; \
49361  poly8x16_t __rev0_352; __rev0_352 = __builtin_shufflevector(__s0_352, __s0_352, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49362  poly8x16_t __ret_352; \
49363  __ret_352 = __noswap_splatq_laneq_p8(__rev0_352, __p1_352); \
49364  __ret_352 = __builtin_shufflevector(__ret_352, __ret_352, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49365  __ret_352; \
49366 })
49367 #endif
49368 
49369 #ifdef __LITTLE_ENDIAN__
49370 #define vdupq_laneq_p64(__p0_353, __p1_353) __extension__ ({ \
49371  poly64x2_t __s0_353 = __p0_353; \
49372  poly64x2_t __ret_353; \
49373  __ret_353 = splatq_laneq_p64(__s0_353, __p1_353); \
49374  __ret_353; \
49375 })
49376 #else
49377 #define vdupq_laneq_p64(__p0_354, __p1_354) __extension__ ({ \
49378  poly64x2_t __s0_354 = __p0_354; \
49379  poly64x2_t __rev0_354; __rev0_354 = __builtin_shufflevector(__s0_354, __s0_354, 1, 0); \
49380  poly64x2_t __ret_354; \
49381  __ret_354 = __noswap_splatq_laneq_p64(__rev0_354, __p1_354); \
49382  __ret_354 = __builtin_shufflevector(__ret_354, __ret_354, 1, 0); \
49383  __ret_354; \
49384 })
49385 #endif
49386 
49387 #ifdef __LITTLE_ENDIAN__
49388 #define vdupq_laneq_p16(__p0_355, __p1_355) __extension__ ({ \
49389  poly16x8_t __s0_355 = __p0_355; \
49390  poly16x8_t __ret_355; \
49391  __ret_355 = splatq_laneq_p16(__s0_355, __p1_355); \
49392  __ret_355; \
49393 })
49394 #else
49395 #define vdupq_laneq_p16(__p0_356, __p1_356) __extension__ ({ \
49396  poly16x8_t __s0_356 = __p0_356; \
49397  poly16x8_t __rev0_356; __rev0_356 = __builtin_shufflevector(__s0_356, __s0_356, 7, 6, 5, 4, 3, 2, 1, 0); \
49398  poly16x8_t __ret_356; \
49399  __ret_356 = __noswap_splatq_laneq_p16(__rev0_356, __p1_356); \
49400  __ret_356 = __builtin_shufflevector(__ret_356, __ret_356, 7, 6, 5, 4, 3, 2, 1, 0); \
49401  __ret_356; \
49402 })
49403 #endif
49404 
49405 #ifdef __LITTLE_ENDIAN__
49406 #define vdupq_laneq_u8(__p0_357, __p1_357) __extension__ ({ \
49407  uint8x16_t __s0_357 = __p0_357; \
49408  uint8x16_t __ret_357; \
49409  __ret_357 = splatq_laneq_u8(__s0_357, __p1_357); \
49410  __ret_357; \
49411 })
49412 #else
49413 #define vdupq_laneq_u8(__p0_358, __p1_358) __extension__ ({ \
49414  uint8x16_t __s0_358 = __p0_358; \
49415  uint8x16_t __rev0_358; __rev0_358 = __builtin_shufflevector(__s0_358, __s0_358, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49416  uint8x16_t __ret_358; \
49417  __ret_358 = __noswap_splatq_laneq_u8(__rev0_358, __p1_358); \
49418  __ret_358 = __builtin_shufflevector(__ret_358, __ret_358, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49419  __ret_358; \
49420 })
49421 #endif
49422 
49423 #ifdef __LITTLE_ENDIAN__
49424 #define vdupq_laneq_u32(__p0_359, __p1_359) __extension__ ({ \
49425  uint32x4_t __s0_359 = __p0_359; \
49426  uint32x4_t __ret_359; \
49427  __ret_359 = splatq_laneq_u32(__s0_359, __p1_359); \
49428  __ret_359; \
49429 })
49430 #else
49431 #define vdupq_laneq_u32(__p0_360, __p1_360) __extension__ ({ \
49432  uint32x4_t __s0_360 = __p0_360; \
49433  uint32x4_t __rev0_360; __rev0_360 = __builtin_shufflevector(__s0_360, __s0_360, 3, 2, 1, 0); \
49434  uint32x4_t __ret_360; \
49435  __ret_360 = __noswap_splatq_laneq_u32(__rev0_360, __p1_360); \
49436  __ret_360 = __builtin_shufflevector(__ret_360, __ret_360, 3, 2, 1, 0); \
49437  __ret_360; \
49438 })
49439 #endif
49440 
49441 #ifdef __LITTLE_ENDIAN__
49442 #define vdupq_laneq_u64(__p0_361, __p1_361) __extension__ ({ \
49443  uint64x2_t __s0_361 = __p0_361; \
49444  uint64x2_t __ret_361; \
49445  __ret_361 = splatq_laneq_u64(__s0_361, __p1_361); \
49446  __ret_361; \
49447 })
49448 #else
49449 #define vdupq_laneq_u64(__p0_362, __p1_362) __extension__ ({ \
49450  uint64x2_t __s0_362 = __p0_362; \
49451  uint64x2_t __rev0_362; __rev0_362 = __builtin_shufflevector(__s0_362, __s0_362, 1, 0); \
49452  uint64x2_t __ret_362; \
49453  __ret_362 = __noswap_splatq_laneq_u64(__rev0_362, __p1_362); \
49454  __ret_362 = __builtin_shufflevector(__ret_362, __ret_362, 1, 0); \
49455  __ret_362; \
49456 })
49457 #endif
49458 
49459 #ifdef __LITTLE_ENDIAN__
49460 #define vdupq_laneq_u16(__p0_363, __p1_363) __extension__ ({ \
49461  uint16x8_t __s0_363 = __p0_363; \
49462  uint16x8_t __ret_363; \
49463  __ret_363 = splatq_laneq_u16(__s0_363, __p1_363); \
49464  __ret_363; \
49465 })
49466 #else
49467 #define vdupq_laneq_u16(__p0_364, __p1_364) __extension__ ({ \
49468  uint16x8_t __s0_364 = __p0_364; \
49469  uint16x8_t __rev0_364; __rev0_364 = __builtin_shufflevector(__s0_364, __s0_364, 7, 6, 5, 4, 3, 2, 1, 0); \
49470  uint16x8_t __ret_364; \
49471  __ret_364 = __noswap_splatq_laneq_u16(__rev0_364, __p1_364); \
49472  __ret_364 = __builtin_shufflevector(__ret_364, __ret_364, 7, 6, 5, 4, 3, 2, 1, 0); \
49473  __ret_364; \
49474 })
49475 #endif
49476 
49477 #ifdef __LITTLE_ENDIAN__
49478 #define vdupq_laneq_s8(__p0_365, __p1_365) __extension__ ({ \
49479  int8x16_t __s0_365 = __p0_365; \
49480  int8x16_t __ret_365; \
49481  __ret_365 = splatq_laneq_s8(__s0_365, __p1_365); \
49482  __ret_365; \
49483 })
49484 #else
49485 #define vdupq_laneq_s8(__p0_366, __p1_366) __extension__ ({ \
49486  int8x16_t __s0_366 = __p0_366; \
49487  int8x16_t __rev0_366; __rev0_366 = __builtin_shufflevector(__s0_366, __s0_366, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49488  int8x16_t __ret_366; \
49489  __ret_366 = __noswap_splatq_laneq_s8(__rev0_366, __p1_366); \
49490  __ret_366 = __builtin_shufflevector(__ret_366, __ret_366, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49491  __ret_366; \
49492 })
49493 #endif
49494 
49495 #ifdef __LITTLE_ENDIAN__
49496 #define vdupq_laneq_f64(__p0_367, __p1_367) __extension__ ({ \
49497  float64x2_t __s0_367 = __p0_367; \
49498  float64x2_t __ret_367; \
49499  __ret_367 = splatq_laneq_f64(__s0_367, __p1_367); \
49500  __ret_367; \
49501 })
49502 #else
49503 #define vdupq_laneq_f64(__p0_368, __p1_368) __extension__ ({ \
49504  float64x2_t __s0_368 = __p0_368; \
49505  float64x2_t __rev0_368; __rev0_368 = __builtin_shufflevector(__s0_368, __s0_368, 1, 0); \
49506  float64x2_t __ret_368; \
49507  __ret_368 = __noswap_splatq_laneq_f64(__rev0_368, __p1_368); \
49508  __ret_368 = __builtin_shufflevector(__ret_368, __ret_368, 1, 0); \
49509  __ret_368; \
49510 })
49511 #endif
49512 
49513 #ifdef __LITTLE_ENDIAN__
49514 #define vdupq_laneq_f32(__p0_369, __p1_369) __extension__ ({ \
49515  float32x4_t __s0_369 = __p0_369; \
49516  float32x4_t __ret_369; \
49517  __ret_369 = splatq_laneq_f32(__s0_369, __p1_369); \
49518  __ret_369; \
49519 })
49520 #else
49521 #define vdupq_laneq_f32(__p0_370, __p1_370) __extension__ ({ \
49522  float32x4_t __s0_370 = __p0_370; \
49523  float32x4_t __rev0_370; __rev0_370 = __builtin_shufflevector(__s0_370, __s0_370, 3, 2, 1, 0); \
49524  float32x4_t __ret_370; \
49525  __ret_370 = __noswap_splatq_laneq_f32(__rev0_370, __p1_370); \
49526  __ret_370 = __builtin_shufflevector(__ret_370, __ret_370, 3, 2, 1, 0); \
49527  __ret_370; \
49528 })
49529 #endif
49530 
49531 #ifdef __LITTLE_ENDIAN__
49532 #define vdupq_laneq_f16(__p0_371, __p1_371) __extension__ ({ \
49533  float16x8_t __s0_371 = __p0_371; \
49534  float16x8_t __ret_371; \
49535  __ret_371 = splatq_laneq_f16(__s0_371, __p1_371); \
49536  __ret_371; \
49537 })
49538 #else
49539 #define vdupq_laneq_f16(__p0_372, __p1_372) __extension__ ({ \
49540  float16x8_t __s0_372 = __p0_372; \
49541  float16x8_t __rev0_372; __rev0_372 = __builtin_shufflevector(__s0_372, __s0_372, 7, 6, 5, 4, 3, 2, 1, 0); \
49542  float16x8_t __ret_372; \
49543  __ret_372 = __noswap_splatq_laneq_f16(__rev0_372, __p1_372); \
49544  __ret_372 = __builtin_shufflevector(__ret_372, __ret_372, 7, 6, 5, 4, 3, 2, 1, 0); \
49545  __ret_372; \
49546 })
49547 #endif
49548 
49549 #ifdef __LITTLE_ENDIAN__
49550 #define vdupq_laneq_s32(__p0_373, __p1_373) __extension__ ({ \
49551  int32x4_t __s0_373 = __p0_373; \
49552  int32x4_t __ret_373; \
49553  __ret_373 = splatq_laneq_s32(__s0_373, __p1_373); \
49554  __ret_373; \
49555 })
49556 #else
49557 #define vdupq_laneq_s32(__p0_374, __p1_374) __extension__ ({ \
49558  int32x4_t __s0_374 = __p0_374; \
49559  int32x4_t __rev0_374; __rev0_374 = __builtin_shufflevector(__s0_374, __s0_374, 3, 2, 1, 0); \
49560  int32x4_t __ret_374; \
49561  __ret_374 = __noswap_splatq_laneq_s32(__rev0_374, __p1_374); \
49562  __ret_374 = __builtin_shufflevector(__ret_374, __ret_374, 3, 2, 1, 0); \
49563  __ret_374; \
49564 })
49565 #endif
49566 
49567 #ifdef __LITTLE_ENDIAN__
49568 #define vdupq_laneq_s64(__p0_375, __p1_375) __extension__ ({ \
49569  int64x2_t __s0_375 = __p0_375; \
49570  int64x2_t __ret_375; \
49571  __ret_375 = splatq_laneq_s64(__s0_375, __p1_375); \
49572  __ret_375; \
49573 })
49574 #else
49575 #define vdupq_laneq_s64(__p0_376, __p1_376) __extension__ ({ \
49576  int64x2_t __s0_376 = __p0_376; \
49577  int64x2_t __rev0_376; __rev0_376 = __builtin_shufflevector(__s0_376, __s0_376, 1, 0); \
49578  int64x2_t __ret_376; \
49579  __ret_376 = __noswap_splatq_laneq_s64(__rev0_376, __p1_376); \
49580  __ret_376 = __builtin_shufflevector(__ret_376, __ret_376, 1, 0); \
49581  __ret_376; \
49582 })
49583 #endif
49584 
49585 #ifdef __LITTLE_ENDIAN__
49586 #define vdupq_laneq_s16(__p0_377, __p1_377) __extension__ ({ \
49587  int16x8_t __s0_377 = __p0_377; \
49588  int16x8_t __ret_377; \
49589  __ret_377 = splatq_laneq_s16(__s0_377, __p1_377); \
49590  __ret_377; \
49591 })
49592 #else
49593 #define vdupq_laneq_s16(__p0_378, __p1_378) __extension__ ({ \
49594  int16x8_t __s0_378 = __p0_378; \
49595  int16x8_t __rev0_378; __rev0_378 = __builtin_shufflevector(__s0_378, __s0_378, 7, 6, 5, 4, 3, 2, 1, 0); \
49596  int16x8_t __ret_378; \
49597  __ret_378 = __noswap_splatq_laneq_s16(__rev0_378, __p1_378); \
49598  __ret_378 = __builtin_shufflevector(__ret_378, __ret_378, 7, 6, 5, 4, 3, 2, 1, 0); \
49599  __ret_378; \
49600 })
49601 #endif
49602 
49603 #ifdef __LITTLE_ENDIAN__
49604 #define vdup_laneq_u8(__p0_379, __p1_379) __extension__ ({ \
49605  uint8x16_t __s0_379 = __p0_379; \
49606  uint8x8_t __ret_379; \
49607  __ret_379 = splat_laneq_u8(__s0_379, __p1_379); \
49608  __ret_379; \
49609 })
49610 #else
49611 #define vdup_laneq_u8(__p0_380, __p1_380) __extension__ ({ \
49612  uint8x16_t __s0_380 = __p0_380; \
49613  uint8x16_t __rev0_380; __rev0_380 = __builtin_shufflevector(__s0_380, __s0_380, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49614  uint8x8_t __ret_380; \
49615  __ret_380 = __noswap_splat_laneq_u8(__rev0_380, __p1_380); \
49616  __ret_380 = __builtin_shufflevector(__ret_380, __ret_380, 7, 6, 5, 4, 3, 2, 1, 0); \
49617  __ret_380; \
49618 })
49619 #endif
49620 
49621 #ifdef __LITTLE_ENDIAN__
49622 #define vdup_laneq_u32(__p0_381, __p1_381) __extension__ ({ \
49623  uint32x4_t __s0_381 = __p0_381; \
49624  uint32x2_t __ret_381; \
49625  __ret_381 = splat_laneq_u32(__s0_381, __p1_381); \
49626  __ret_381; \
49627 })
49628 #else
49629 #define vdup_laneq_u32(__p0_382, __p1_382) __extension__ ({ \
49630  uint32x4_t __s0_382 = __p0_382; \
49631  uint32x4_t __rev0_382; __rev0_382 = __builtin_shufflevector(__s0_382, __s0_382, 3, 2, 1, 0); \
49632  uint32x2_t __ret_382; \
49633  __ret_382 = __noswap_splat_laneq_u32(__rev0_382, __p1_382); \
49634  __ret_382 = __builtin_shufflevector(__ret_382, __ret_382, 1, 0); \
49635  __ret_382; \
49636 })
49637 #endif
49638 
49639 #ifdef __LITTLE_ENDIAN__
49640 #define vdup_laneq_u64(__p0_383, __p1_383) __extension__ ({ \
49641  uint64x2_t __s0_383 = __p0_383; \
49642  uint64x1_t __ret_383; \
49643  __ret_383 = splat_laneq_u64(__s0_383, __p1_383); \
49644  __ret_383; \
49645 })
49646 #else
49647 #define vdup_laneq_u64(__p0_384, __p1_384) __extension__ ({ \
49648  uint64x2_t __s0_384 = __p0_384; \
49649  uint64x2_t __rev0_384; __rev0_384 = __builtin_shufflevector(__s0_384, __s0_384, 1, 0); \
49650  uint64x1_t __ret_384; \
49651  __ret_384 = __noswap_splat_laneq_u64(__rev0_384, __p1_384); \
49652  __ret_384; \
49653 })
49654 #endif
49655 
49656 #ifdef __LITTLE_ENDIAN__
49657 #define vdup_laneq_u16(__p0_385, __p1_385) __extension__ ({ \
49658  uint16x8_t __s0_385 = __p0_385; \
49659  uint16x4_t __ret_385; \
49660  __ret_385 = splat_laneq_u16(__s0_385, __p1_385); \
49661  __ret_385; \
49662 })
49663 #else
49664 #define vdup_laneq_u16(__p0_386, __p1_386) __extension__ ({ \
49665  uint16x8_t __s0_386 = __p0_386; \
49666  uint16x8_t __rev0_386; __rev0_386 = __builtin_shufflevector(__s0_386, __s0_386, 7, 6, 5, 4, 3, 2, 1, 0); \
49667  uint16x4_t __ret_386; \
49668  __ret_386 = __noswap_splat_laneq_u16(__rev0_386, __p1_386); \
49669  __ret_386 = __builtin_shufflevector(__ret_386, __ret_386, 3, 2, 1, 0); \
49670  __ret_386; \
49671 })
49672 #endif
49673 
49674 #ifdef __LITTLE_ENDIAN__
49675 #define vdup_laneq_s8(__p0_387, __p1_387) __extension__ ({ \
49676  int8x16_t __s0_387 = __p0_387; \
49677  int8x8_t __ret_387; \
49678  __ret_387 = splat_laneq_s8(__s0_387, __p1_387); \
49679  __ret_387; \
49680 })
49681 #else
49682 #define vdup_laneq_s8(__p0_388, __p1_388) __extension__ ({ \
49683  int8x16_t __s0_388 = __p0_388; \
49684  int8x16_t __rev0_388; __rev0_388 = __builtin_shufflevector(__s0_388, __s0_388, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49685  int8x8_t __ret_388; \
49686  __ret_388 = __noswap_splat_laneq_s8(__rev0_388, __p1_388); \
49687  __ret_388 = __builtin_shufflevector(__ret_388, __ret_388, 7, 6, 5, 4, 3, 2, 1, 0); \
49688  __ret_388; \
49689 })
49690 #endif
49691 
49692 #ifdef __LITTLE_ENDIAN__
49693 #define vdup_laneq_f64(__p0_389, __p1_389) __extension__ ({ \
49694  float64x2_t __s0_389 = __p0_389; \
49695  float64x1_t __ret_389; \
49696  __ret_389 = splat_laneq_f64(__s0_389, __p1_389); \
49697  __ret_389; \
49698 })
49699 #else
49700 #define vdup_laneq_f64(__p0_390, __p1_390) __extension__ ({ \
49701  float64x2_t __s0_390 = __p0_390; \
49702  float64x2_t __rev0_390; __rev0_390 = __builtin_shufflevector(__s0_390, __s0_390, 1, 0); \
49703  float64x1_t __ret_390; \
49704  __ret_390 = __noswap_splat_laneq_f64(__rev0_390, __p1_390); \
49705  __ret_390; \
49706 })
49707 #endif
49708 
49709 #ifdef __LITTLE_ENDIAN__
49710 #define vdup_laneq_f32(__p0_391, __p1_391) __extension__ ({ \
49711  float32x4_t __s0_391 = __p0_391; \
49712  float32x2_t __ret_391; \
49713  __ret_391 = splat_laneq_f32(__s0_391, __p1_391); \
49714  __ret_391; \
49715 })
49716 #else
49717 #define vdup_laneq_f32(__p0_392, __p1_392) __extension__ ({ \
49718  float32x4_t __s0_392 = __p0_392; \
49719  float32x4_t __rev0_392; __rev0_392 = __builtin_shufflevector(__s0_392, __s0_392, 3, 2, 1, 0); \
49720  float32x2_t __ret_392; \
49721  __ret_392 = __noswap_splat_laneq_f32(__rev0_392, __p1_392); \
49722  __ret_392 = __builtin_shufflevector(__ret_392, __ret_392, 1, 0); \
49723  __ret_392; \
49724 })
49725 #endif
49726 
49727 #ifdef __LITTLE_ENDIAN__
49728 #define vdup_laneq_f16(__p0_393, __p1_393) __extension__ ({ \
49729  float16x8_t __s0_393 = __p0_393; \
49730  float16x4_t __ret_393; \
49731  __ret_393 = splat_laneq_f16(__s0_393, __p1_393); \
49732  __ret_393; \
49733 })
49734 #else
49735 #define vdup_laneq_f16(__p0_394, __p1_394) __extension__ ({ \
49736  float16x8_t __s0_394 = __p0_394; \
49737  float16x8_t __rev0_394; __rev0_394 = __builtin_shufflevector(__s0_394, __s0_394, 7, 6, 5, 4, 3, 2, 1, 0); \
49738  float16x4_t __ret_394; \
49739  __ret_394 = __noswap_splat_laneq_f16(__rev0_394, __p1_394); \
49740  __ret_394 = __builtin_shufflevector(__ret_394, __ret_394, 3, 2, 1, 0); \
49741  __ret_394; \
49742 })
49743 #endif
49744 
49745 #ifdef __LITTLE_ENDIAN__
49746 #define vdup_laneq_s32(__p0_395, __p1_395) __extension__ ({ \
49747  int32x4_t __s0_395 = __p0_395; \
49748  int32x2_t __ret_395; \
49749  __ret_395 = splat_laneq_s32(__s0_395, __p1_395); \
49750  __ret_395; \
49751 })
49752 #else
49753 #define vdup_laneq_s32(__p0_396, __p1_396) __extension__ ({ \
49754  int32x4_t __s0_396 = __p0_396; \
49755  int32x4_t __rev0_396; __rev0_396 = __builtin_shufflevector(__s0_396, __s0_396, 3, 2, 1, 0); \
49756  int32x2_t __ret_396; \
49757  __ret_396 = __noswap_splat_laneq_s32(__rev0_396, __p1_396); \
49758  __ret_396 = __builtin_shufflevector(__ret_396, __ret_396, 1, 0); \
49759  __ret_396; \
49760 })
49761 #endif
49762 
49763 #ifdef __LITTLE_ENDIAN__
49764 #define vdup_laneq_s64(__p0_397, __p1_397) __extension__ ({ \
49765  int64x2_t __s0_397 = __p0_397; \
49766  int64x1_t __ret_397; \
49767  __ret_397 = splat_laneq_s64(__s0_397, __p1_397); \
49768  __ret_397; \
49769 })
49770 #else
49771 #define vdup_laneq_s64(__p0_398, __p1_398) __extension__ ({ \
49772  int64x2_t __s0_398 = __p0_398; \
49773  int64x2_t __rev0_398; __rev0_398 = __builtin_shufflevector(__s0_398, __s0_398, 1, 0); \
49774  int64x1_t __ret_398; \
49775  __ret_398 = __noswap_splat_laneq_s64(__rev0_398, __p1_398); \
49776  __ret_398; \
49777 })
49778 #endif
49779 
49780 #ifdef __LITTLE_ENDIAN__
49781 #define vdup_laneq_s16(__p0_399, __p1_399) __extension__ ({ \
49782  int16x8_t __s0_399 = __p0_399; \
49783  int16x4_t __ret_399; \
49784  __ret_399 = splat_laneq_s16(__s0_399, __p1_399); \
49785  __ret_399; \
49786 })
49787 #else
49788 #define vdup_laneq_s16(__p0_400, __p1_400) __extension__ ({ \
49789  int16x8_t __s0_400 = __p0_400; \
49790  int16x8_t __rev0_400; __rev0_400 = __builtin_shufflevector(__s0_400, __s0_400, 7, 6, 5, 4, 3, 2, 1, 0); \
49791  int16x4_t __ret_400; \
49792  __ret_400 = __noswap_splat_laneq_s16(__rev0_400, __p1_400); \
49793  __ret_400 = __builtin_shufflevector(__ret_400, __ret_400, 3, 2, 1, 0); \
49794  __ret_400; \
49795 })
49796 #endif
49797 
49798 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
49799  poly64x1_t __ret;
49800  __ret = (poly64x1_t) {__p0};
49801  return __ret;
49802 }
49803 #ifdef __LITTLE_ENDIAN__
49804 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
49805  poly64x2_t __ret;
49806  __ret = (poly64x2_t) {__p0, __p0};
49807  return __ret;
49808 }
49809 #else
49810 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
49811  poly64x2_t __ret;
49812  __ret = (poly64x2_t) {__p0, __p0};
49813  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
49814  return __ret;
49815 }
49816 #endif
49817 
49818 #ifdef __LITTLE_ENDIAN__
49819 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
49820  float64x2_t __ret;
49821  __ret = (float64x2_t) {__p0, __p0};
49822  return __ret;
49823 }
49824 #else
49825 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
49826  float64x2_t __ret;
49827  __ret = (float64x2_t) {__p0, __p0};
49828  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
49829  return __ret;
49830 }
49831 #endif
49832 
49833 __ai float64x1_t vdup_n_f64(float64_t __p0) {
49834  float64x1_t __ret;
49835  __ret = (float64x1_t) {__p0};
49836  return __ret;
49837 }
49838 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
49839  poly64x1_t __s0 = __p0; \
49840  poly64x1_t __s1 = __p1; \
49841  poly64x1_t __ret; \
49842  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
49843  __ret; \
49844 })
49845 #ifdef __LITTLE_ENDIAN__
49846 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
49847  poly64x2_t __s0 = __p0; \
49848  poly64x2_t __s1 = __p1; \
49849  poly64x2_t __ret; \
49850  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
49851  __ret; \
49852 })
49853 #else
49854 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
49855  poly64x2_t __s0 = __p0; \
49856  poly64x2_t __s1 = __p1; \
49857  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49858  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
49859  poly64x2_t __ret; \
49860  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
49861  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
49862  __ret; \
49863 })
49864 #endif
49865 
49866 #ifdef __LITTLE_ENDIAN__
49867 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
49868  float64x2_t __s0 = __p0; \
49869  float64x2_t __s1 = __p1; \
49870  float64x2_t __ret; \
49871  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
49872  __ret; \
49873 })
49874 #else
49875 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
49876  float64x2_t __s0 = __p0; \
49877  float64x2_t __s1 = __p1; \
49878  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49879  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
49880  float64x2_t __ret; \
49881  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
49882  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
49883  __ret; \
49884 })
49885 #endif
49886 
49887 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
49888  float64x1_t __s0 = __p0; \
49889  float64x1_t __s1 = __p1; \
49890  float64x1_t __ret; \
49891  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
49892  __ret; \
49893 })
49894 #ifdef __LITTLE_ENDIAN__
49895 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49896  float64x2_t __ret;
49897  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
49898  return __ret;
49899 }
49900 #else
49901 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49902  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
49903  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
49904  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
49905  float64x2_t __ret;
49906  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
49907  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
49908  return __ret;
49909 }
49910 __ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
49911  float64x2_t __ret;
49912  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
49913  return __ret;
49914 }
49915 #endif
49916 
49917 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
49918  float64x1_t __ret;
49919  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
49920  return __ret;
49921 }
49922 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
49923  float64_t __s0 = __p0; \
49924  float64_t __s1 = __p1; \
49925  float64x1_t __s2 = __p2; \
49926  float64_t __ret; \
49927  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (float64x1_t)__s2, __p3); \
49928  __ret; \
49929 })
49930 #ifdef __LITTLE_ENDIAN__
49931 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
49932  float32_t __s0 = __p0; \
49933  float32_t __s1 = __p1; \
49934  float32x2_t __s2 = __p2; \
49935  float32_t __ret; \
49936  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__s2, __p3); \
49937  __ret; \
49938 })
49939 #else
49940 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
49941  float32_t __s0 = __p0; \
49942  float32_t __s1 = __p1; \
49943  float32x2_t __s2 = __p2; \
49944  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
49945  float32_t __ret; \
49946  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__rev2, __p3); \
49947  __ret; \
49948 })
49949 #define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
49950  float32_t __s0 = __p0; \
49951  float32_t __s1 = __p1; \
49952  float32x2_t __s2 = __p2; \
49953  float32_t __ret; \
49954  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (float32x2_t)__s2, __p3); \
49955  __ret; \
49956 })
49957 #endif
49958 
49959 #ifdef __LITTLE_ENDIAN__
49960 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
49961  float64x2_t __s0 = __p0; \
49962  float64x2_t __s1 = __p1; \
49963  float64x1_t __s2 = __p2; \
49964  float64x2_t __ret; \
49965  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
49966  __ret; \
49967 })
49968 #else
49969 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
49970  float64x2_t __s0 = __p0; \
49971  float64x2_t __s1 = __p1; \
49972  float64x1_t __s2 = __p2; \
49973  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
49974  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
49975  float64x2_t __ret; \
49976  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
49977  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
49978  __ret; \
49979 })
49980 #define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
49981  float64x2_t __s0 = __p0; \
49982  float64x2_t __s1 = __p1; \
49983  float64x1_t __s2 = __p2; \
49984  float64x2_t __ret; \
49985  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
49986  __ret; \
49987 })
49988 #endif
49989 
49990 #ifdef __LITTLE_ENDIAN__
49991 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
49992  float32x4_t __s0 = __p0; \
49993  float32x4_t __s1 = __p1; \
49994  float32x2_t __s2 = __p2; \
49995  float32x4_t __ret; \
49996  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
49997  __ret; \
49998 })
49999 #else
50000 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50001  float32x4_t __s0 = __p0; \
50002  float32x4_t __s1 = __p1; \
50003  float32x2_t __s2 = __p2; \
50004  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
50005  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
50006  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
50007  float32x4_t __ret; \
50008  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
50009  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
50010  __ret; \
50011 })
50012 #define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50013  float32x4_t __s0 = __p0; \
50014  float32x4_t __s1 = __p1; \
50015  float32x2_t __s2 = __p2; \
50016  float32x4_t __ret; \
50017  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
50018  __ret; \
50019 })
50020 #endif
50021 
50022 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50023  float64x1_t __s0 = __p0; \
50024  float64x1_t __s1 = __p1; \
50025  float64x1_t __s2 = __p2; \
50026  float64x1_t __ret; \
50027  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
50028  __ret; \
50029 })
50030 #ifdef __LITTLE_ENDIAN__
50031 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50032  float32x2_t __s0 = __p0; \
50033  float32x2_t __s1 = __p1; \
50034  float32x2_t __s2 = __p2; \
50035  float32x2_t __ret; \
50036  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
50037  __ret; \
50038 })
50039 #else
50040 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50041  float32x2_t __s0 = __p0; \
50042  float32x2_t __s1 = __p1; \
50043  float32x2_t __s2 = __p2; \
50044  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
50045  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50046  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
50047  float32x2_t __ret; \
50048  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
50049  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50050  __ret; \
50051 })
50052 #define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50053  float32x2_t __s0 = __p0; \
50054  float32x2_t __s1 = __p1; \
50055  float32x2_t __s2 = __p2; \
50056  float32x2_t __ret; \
50057  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
50058  __ret; \
50059 })
50060 #endif
50061 
50062 #ifdef __LITTLE_ENDIAN__
50063 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50064  float64_t __s0 = __p0; \
50065  float64_t __s1 = __p1; \
50066  float64x2_t __s2 = __p2; \
50067  float64_t __ret; \
50068  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__s2, __p3); \
50069  __ret; \
50070 })
50071 #else
50072 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50073  float64_t __s0 = __p0; \
50074  float64_t __s1 = __p1; \
50075  float64x2_t __s2 = __p2; \
50076  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
50077  float64_t __ret; \
50078  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__rev2, __p3); \
50079  __ret; \
50080 })
50081 #define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50082  float64_t __s0 = __p0; \
50083  float64_t __s1 = __p1; \
50084  float64x2_t __s2 = __p2; \
50085  float64_t __ret; \
50086  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (float64x2_t)__s2, __p3); \
50087  __ret; \
50088 })
50089 #endif
50090 
50091 #ifdef __LITTLE_ENDIAN__
50092 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50093  float32_t __s0 = __p0; \
50094  float32_t __s1 = __p1; \
50095  float32x4_t __s2 = __p2; \
50096  float32_t __ret; \
50097  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__s2, __p3); \
50098  __ret; \
50099 })
50100 #else
50101 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50102  float32_t __s0 = __p0; \
50103  float32_t __s1 = __p1; \
50104  float32x4_t __s2 = __p2; \
50105  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
50106  float32_t __ret; \
50107  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__rev2, __p3); \
50108  __ret; \
50109 })
50110 #define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50111  float32_t __s0 = __p0; \
50112  float32_t __s1 = __p1; \
50113  float32x4_t __s2 = __p2; \
50114  float32_t __ret; \
50115  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (float32x4_t)__s2, __p3); \
50116  __ret; \
50117 })
50118 #endif
50119 
50120 #ifdef __LITTLE_ENDIAN__
50121 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50122  float64x2_t __s0 = __p0; \
50123  float64x2_t __s1 = __p1; \
50124  float64x2_t __s2 = __p2; \
50125  float64x2_t __ret; \
50126  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
50127  __ret; \
50128 })
50129 #else
50130 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50131  float64x2_t __s0 = __p0; \
50132  float64x2_t __s1 = __p1; \
50133  float64x2_t __s2 = __p2; \
50134  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
50135  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50136  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
50137  float64x2_t __ret; \
50138  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
50139  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50140  __ret; \
50141 })
50142 #define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50143  float64x2_t __s0 = __p0; \
50144  float64x2_t __s1 = __p1; \
50145  float64x2_t __s2 = __p2; \
50146  float64x2_t __ret; \
50147  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
50148  __ret; \
50149 })
50150 #endif
50151 
50152 #ifdef __LITTLE_ENDIAN__
50153 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50154  float32x4_t __s0 = __p0; \
50155  float32x4_t __s1 = __p1; \
50156  float32x4_t __s2 = __p2; \
50157  float32x4_t __ret; \
50158  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
50159  __ret; \
50160 })
50161 #else
50162 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50163  float32x4_t __s0 = __p0; \
50164  float32x4_t __s1 = __p1; \
50165  float32x4_t __s2 = __p2; \
50166  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
50167  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
50168  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
50169  float32x4_t __ret; \
50170  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
50171  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
50172  __ret; \
50173 })
50174 #define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50175  float32x4_t __s0 = __p0; \
50176  float32x4_t __s1 = __p1; \
50177  float32x4_t __s2 = __p2; \
50178  float32x4_t __ret; \
50179  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
50180  __ret; \
50181 })
50182 #endif
50183 
50184 #ifdef __LITTLE_ENDIAN__
50185 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50186  float64x1_t __s0 = __p0; \
50187  float64x1_t __s1 = __p1; \
50188  float64x2_t __s2 = __p2; \
50189  float64x1_t __ret; \
50190  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
50191  __ret; \
50192 })
50193 #else
50194 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50195  float64x1_t __s0 = __p0; \
50196  float64x1_t __s1 = __p1; \
50197  float64x2_t __s2 = __p2; \
50198  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
50199  float64x1_t __ret; \
50200  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
50201  __ret; \
50202 })
50203 #define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
50204  float64x1_t __s0 = __p0; \
50205  float64x1_t __s1 = __p1; \
50206  float64x2_t __s2 = __p2; \
50207  float64x1_t __ret; \
50208  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
50209  __ret; \
50210 })
50211 #endif
50212 
50213 #ifdef __LITTLE_ENDIAN__
50214 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50215  float32x2_t __s0 = __p0; \
50216  float32x2_t __s1 = __p1; \
50217  float32x4_t __s2 = __p2; \
50218  float32x2_t __ret; \
50219  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
50220  __ret; \
50221 })
50222 #else
50223 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50224  float32x2_t __s0 = __p0; \
50225  float32x2_t __s1 = __p1; \
50226  float32x4_t __s2 = __p2; \
50227  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
50228  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50229  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
50230  float32x2_t __ret; \
50231  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
50232  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50233  __ret; \
50234 })
50235 #define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
50236  float32x2_t __s0 = __p0; \
50237  float32x2_t __s1 = __p1; \
50238  float32x4_t __s2 = __p2; \
50239  float32x2_t __ret; \
50240  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
50241  __ret; \
50242 })
50243 #endif
50244 
50245 #ifdef __LITTLE_ENDIAN__
50246 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
50247  float64x2_t __ret;
50248  __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
50249  return __ret;
50250 }
50251 #else
50252 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
50253  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50254  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50255  float64x2_t __ret;
50256  __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
50257  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50258  return __ret;
50259 }
50260 #endif
50261 
50262 __ai float64x1_t vfma_n_f64(float64x1_t __p0, float64x1_t __p1, float64_t __p2) {
50263  float64x1_t __ret;
50264  __ret = vfma_f64(__p0, __p1, (float64x1_t) {__p2});
50265  return __ret;
50266 }
50267 #ifdef __LITTLE_ENDIAN__
50268 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
50269  float64x2_t __ret;
50270  __ret = vfmaq_f64(__p0, -__p1, __p2);
50271  return __ret;
50272 }
50273 #else
50274 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
50275  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50276  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50277  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
50278  float64x2_t __ret;
50279  __ret = __noswap_vfmaq_f64(__rev0, -__rev1, __rev2);
50280  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50281  return __ret;
50282 }
50283 #endif
50284 
50285 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
50286  float64x1_t __ret;
50287  __ret = vfma_f64(__p0, -__p1, __p2);
50288  return __ret;
50289 }
50290 #define vfmsd_lane_f64(__p0_401, __p1_401, __p2_401, __p3_401) __extension__ ({ \
50291  float64_t __s0_401 = __p0_401; \
50292  float64_t __s1_401 = __p1_401; \
50293  float64x1_t __s2_401 = __p2_401; \
50294  float64_t __ret_401; \
50295  __ret_401 = vfmad_lane_f64(__s0_401, -__s1_401, __s2_401, __p3_401); \
50296  __ret_401; \
50297 })
50298 #ifdef __LITTLE_ENDIAN__
50299 #define vfmss_lane_f32(__p0_402, __p1_402, __p2_402, __p3_402) __extension__ ({ \
50300  float32_t __s0_402 = __p0_402; \
50301  float32_t __s1_402 = __p1_402; \
50302  float32x2_t __s2_402 = __p2_402; \
50303  float32_t __ret_402; \
50304  __ret_402 = vfmas_lane_f32(__s0_402, -__s1_402, __s2_402, __p3_402); \
50305  __ret_402; \
50306 })
50307 #else
50308 #define vfmss_lane_f32(__p0_403, __p1_403, __p2_403, __p3_403) __extension__ ({ \
50309  float32_t __s0_403 = __p0_403; \
50310  float32_t __s1_403 = __p1_403; \
50311  float32x2_t __s2_403 = __p2_403; \
50312  float32x2_t __rev2_403; __rev2_403 = __builtin_shufflevector(__s2_403, __s2_403, 1, 0); \
50313  float32_t __ret_403; \
50314  __ret_403 = __noswap_vfmas_lane_f32(__s0_403, -__s1_403, __rev2_403, __p3_403); \
50315  __ret_403; \
50316 })
50317 #endif
50318 
50319 #ifdef __LITTLE_ENDIAN__
50320 #define vfmsq_lane_f64(__p0_404, __p1_404, __p2_404, __p3_404) __extension__ ({ \
50321  float64x2_t __s0_404 = __p0_404; \
50322  float64x2_t __s1_404 = __p1_404; \
50323  float64x1_t __s2_404 = __p2_404; \
50324  float64x2_t __ret_404; \
50325  __ret_404 = vfmaq_lane_f64(__s0_404, -__s1_404, __s2_404, __p3_404); \
50326  __ret_404; \
50327 })
50328 #else
50329 #define vfmsq_lane_f64(__p0_405, __p1_405, __p2_405, __p3_405) __extension__ ({ \
50330  float64x2_t __s0_405 = __p0_405; \
50331  float64x2_t __s1_405 = __p1_405; \
50332  float64x1_t __s2_405 = __p2_405; \
50333  float64x2_t __rev0_405; __rev0_405 = __builtin_shufflevector(__s0_405, __s0_405, 1, 0); \
50334  float64x2_t __rev1_405; __rev1_405 = __builtin_shufflevector(__s1_405, __s1_405, 1, 0); \
50335  float64x2_t __ret_405; \
50336  __ret_405 = __noswap_vfmaq_lane_f64(__rev0_405, -__rev1_405, __s2_405, __p3_405); \
50337  __ret_405 = __builtin_shufflevector(__ret_405, __ret_405, 1, 0); \
50338  __ret_405; \
50339 })
50340 #endif
50341 
50342 #ifdef __LITTLE_ENDIAN__
50343 #define vfmsq_lane_f32(__p0_406, __p1_406, __p2_406, __p3_406) __extension__ ({ \
50344  float32x4_t __s0_406 = __p0_406; \
50345  float32x4_t __s1_406 = __p1_406; \
50346  float32x2_t __s2_406 = __p2_406; \
50347  float32x4_t __ret_406; \
50348  __ret_406 = vfmaq_lane_f32(__s0_406, -__s1_406, __s2_406, __p3_406); \
50349  __ret_406; \
50350 })
50351 #else
50352 #define vfmsq_lane_f32(__p0_407, __p1_407, __p2_407, __p3_407) __extension__ ({ \
50353  float32x4_t __s0_407 = __p0_407; \
50354  float32x4_t __s1_407 = __p1_407; \
50355  float32x2_t __s2_407 = __p2_407; \
50356  float32x4_t __rev0_407; __rev0_407 = __builtin_shufflevector(__s0_407, __s0_407, 3, 2, 1, 0); \
50357  float32x4_t __rev1_407; __rev1_407 = __builtin_shufflevector(__s1_407, __s1_407, 3, 2, 1, 0); \
50358  float32x2_t __rev2_407; __rev2_407 = __builtin_shufflevector(__s2_407, __s2_407, 1, 0); \
50359  float32x4_t __ret_407; \
50360  __ret_407 = __noswap_vfmaq_lane_f32(__rev0_407, -__rev1_407, __rev2_407, __p3_407); \
50361  __ret_407 = __builtin_shufflevector(__ret_407, __ret_407, 3, 2, 1, 0); \
50362  __ret_407; \
50363 })
50364 #endif
50365 
50366 #define vfms_lane_f64(__p0_408, __p1_408, __p2_408, __p3_408) __extension__ ({ \
50367  float64x1_t __s0_408 = __p0_408; \
50368  float64x1_t __s1_408 = __p1_408; \
50369  float64x1_t __s2_408 = __p2_408; \
50370  float64x1_t __ret_408; \
50371  __ret_408 = vfma_lane_f64(__s0_408, -__s1_408, __s2_408, __p3_408); \
50372  __ret_408; \
50373 })
50374 #ifdef __LITTLE_ENDIAN__
50375 #define vfms_lane_f32(__p0_409, __p1_409, __p2_409, __p3_409) __extension__ ({ \
50376  float32x2_t __s0_409 = __p0_409; \
50377  float32x2_t __s1_409 = __p1_409; \
50378  float32x2_t __s2_409 = __p2_409; \
50379  float32x2_t __ret_409; \
50380  __ret_409 = vfma_lane_f32(__s0_409, -__s1_409, __s2_409, __p3_409); \
50381  __ret_409; \
50382 })
50383 #else
50384 #define vfms_lane_f32(__p0_410, __p1_410, __p2_410, __p3_410) __extension__ ({ \
50385  float32x2_t __s0_410 = __p0_410; \
50386  float32x2_t __s1_410 = __p1_410; \
50387  float32x2_t __s2_410 = __p2_410; \
50388  float32x2_t __rev0_410; __rev0_410 = __builtin_shufflevector(__s0_410, __s0_410, 1, 0); \
50389  float32x2_t __rev1_410; __rev1_410 = __builtin_shufflevector(__s1_410, __s1_410, 1, 0); \
50390  float32x2_t __rev2_410; __rev2_410 = __builtin_shufflevector(__s2_410, __s2_410, 1, 0); \
50391  float32x2_t __ret_410; \
50392  __ret_410 = __noswap_vfma_lane_f32(__rev0_410, -__rev1_410, __rev2_410, __p3_410); \
50393  __ret_410 = __builtin_shufflevector(__ret_410, __ret_410, 1, 0); \
50394  __ret_410; \
50395 })
50396 #endif
50397 
50398 #ifdef __LITTLE_ENDIAN__
50399 #define vfmsd_laneq_f64(__p0_411, __p1_411, __p2_411, __p3_411) __extension__ ({ \
50400  float64_t __s0_411 = __p0_411; \
50401  float64_t __s1_411 = __p1_411; \
50402  float64x2_t __s2_411 = __p2_411; \
50403  float64_t __ret_411; \
50404  __ret_411 = vfmad_laneq_f64(__s0_411, -__s1_411, __s2_411, __p3_411); \
50405  __ret_411; \
50406 })
50407 #else
50408 #define vfmsd_laneq_f64(__p0_412, __p1_412, __p2_412, __p3_412) __extension__ ({ \
50409  float64_t __s0_412 = __p0_412; \
50410  float64_t __s1_412 = __p1_412; \
50411  float64x2_t __s2_412 = __p2_412; \
50412  float64x2_t __rev2_412; __rev2_412 = __builtin_shufflevector(__s2_412, __s2_412, 1, 0); \
50413  float64_t __ret_412; \
50414  __ret_412 = __noswap_vfmad_laneq_f64(__s0_412, -__s1_412, __rev2_412, __p3_412); \
50415  __ret_412; \
50416 })
50417 #endif
50418 
50419 #ifdef __LITTLE_ENDIAN__
50420 #define vfmss_laneq_f32(__p0_413, __p1_413, __p2_413, __p3_413) __extension__ ({ \
50421  float32_t __s0_413 = __p0_413; \
50422  float32_t __s1_413 = __p1_413; \
50423  float32x4_t __s2_413 = __p2_413; \
50424  float32_t __ret_413; \
50425  __ret_413 = vfmas_laneq_f32(__s0_413, -__s1_413, __s2_413, __p3_413); \
50426  __ret_413; \
50427 })
50428 #else
50429 #define vfmss_laneq_f32(__p0_414, __p1_414, __p2_414, __p3_414) __extension__ ({ \
50430  float32_t __s0_414 = __p0_414; \
50431  float32_t __s1_414 = __p1_414; \
50432  float32x4_t __s2_414 = __p2_414; \
50433  float32x4_t __rev2_414; __rev2_414 = __builtin_shufflevector(__s2_414, __s2_414, 3, 2, 1, 0); \
50434  float32_t __ret_414; \
50435  __ret_414 = __noswap_vfmas_laneq_f32(__s0_414, -__s1_414, __rev2_414, __p3_414); \
50436  __ret_414; \
50437 })
50438 #endif
50439 
50440 #ifdef __LITTLE_ENDIAN__
50441 #define vfmsq_laneq_f64(__p0_415, __p1_415, __p2_415, __p3_415) __extension__ ({ \
50442  float64x2_t __s0_415 = __p0_415; \
50443  float64x2_t __s1_415 = __p1_415; \
50444  float64x2_t __s2_415 = __p2_415; \
50445  float64x2_t __ret_415; \
50446  __ret_415 = vfmaq_laneq_f64(__s0_415, -__s1_415, __s2_415, __p3_415); \
50447  __ret_415; \
50448 })
50449 #else
50450 #define vfmsq_laneq_f64(__p0_416, __p1_416, __p2_416, __p3_416) __extension__ ({ \
50451  float64x2_t __s0_416 = __p0_416; \
50452  float64x2_t __s1_416 = __p1_416; \
50453  float64x2_t __s2_416 = __p2_416; \
50454  float64x2_t __rev0_416; __rev0_416 = __builtin_shufflevector(__s0_416, __s0_416, 1, 0); \
50455  float64x2_t __rev1_416; __rev1_416 = __builtin_shufflevector(__s1_416, __s1_416, 1, 0); \
50456  float64x2_t __rev2_416; __rev2_416 = __builtin_shufflevector(__s2_416, __s2_416, 1, 0); \
50457  float64x2_t __ret_416; \
50458  __ret_416 = __noswap_vfmaq_laneq_f64(__rev0_416, -__rev1_416, __rev2_416, __p3_416); \
50459  __ret_416 = __builtin_shufflevector(__ret_416, __ret_416, 1, 0); \
50460  __ret_416; \
50461 })
50462 #endif
50463 
50464 #ifdef __LITTLE_ENDIAN__
50465 #define vfmsq_laneq_f32(__p0_417, __p1_417, __p2_417, __p3_417) __extension__ ({ \
50466  float32x4_t __s0_417 = __p0_417; \
50467  float32x4_t __s1_417 = __p1_417; \
50468  float32x4_t __s2_417 = __p2_417; \
50469  float32x4_t __ret_417; \
50470  __ret_417 = vfmaq_laneq_f32(__s0_417, -__s1_417, __s2_417, __p3_417); \
50471  __ret_417; \
50472 })
50473 #else
50474 #define vfmsq_laneq_f32(__p0_418, __p1_418, __p2_418, __p3_418) __extension__ ({ \
50475  float32x4_t __s0_418 = __p0_418; \
50476  float32x4_t __s1_418 = __p1_418; \
50477  float32x4_t __s2_418 = __p2_418; \
50478  float32x4_t __rev0_418; __rev0_418 = __builtin_shufflevector(__s0_418, __s0_418, 3, 2, 1, 0); \
50479  float32x4_t __rev1_418; __rev1_418 = __builtin_shufflevector(__s1_418, __s1_418, 3, 2, 1, 0); \
50480  float32x4_t __rev2_418; __rev2_418 = __builtin_shufflevector(__s2_418, __s2_418, 3, 2, 1, 0); \
50481  float32x4_t __ret_418; \
50482  __ret_418 = __noswap_vfmaq_laneq_f32(__rev0_418, -__rev1_418, __rev2_418, __p3_418); \
50483  __ret_418 = __builtin_shufflevector(__ret_418, __ret_418, 3, 2, 1, 0); \
50484  __ret_418; \
50485 })
50486 #endif
50487 
50488 #ifdef __LITTLE_ENDIAN__
50489 #define vfms_laneq_f64(__p0_419, __p1_419, __p2_419, __p3_419) __extension__ ({ \
50490  float64x1_t __s0_419 = __p0_419; \
50491  float64x1_t __s1_419 = __p1_419; \
50492  float64x2_t __s2_419 = __p2_419; \
50493  float64x1_t __ret_419; \
50494  __ret_419 = vfma_laneq_f64(__s0_419, -__s1_419, __s2_419, __p3_419); \
50495  __ret_419; \
50496 })
50497 #else
50498 #define vfms_laneq_f64(__p0_420, __p1_420, __p2_420, __p3_420) __extension__ ({ \
50499  float64x1_t __s0_420 = __p0_420; \
50500  float64x1_t __s1_420 = __p1_420; \
50501  float64x2_t __s2_420 = __p2_420; \
50502  float64x2_t __rev2_420; __rev2_420 = __builtin_shufflevector(__s2_420, __s2_420, 1, 0); \
50503  float64x1_t __ret_420; \
50504  __ret_420 = __noswap_vfma_laneq_f64(__s0_420, -__s1_420, __rev2_420, __p3_420); \
50505  __ret_420; \
50506 })
50507 #endif
50508 
50509 #ifdef __LITTLE_ENDIAN__
50510 #define vfms_laneq_f32(__p0_421, __p1_421, __p2_421, __p3_421) __extension__ ({ \
50511  float32x2_t __s0_421 = __p0_421; \
50512  float32x2_t __s1_421 = __p1_421; \
50513  float32x4_t __s2_421 = __p2_421; \
50514  float32x2_t __ret_421; \
50515  __ret_421 = vfma_laneq_f32(__s0_421, -__s1_421, __s2_421, __p3_421); \
50516  __ret_421; \
50517 })
50518 #else
50519 #define vfms_laneq_f32(__p0_422, __p1_422, __p2_422, __p3_422) __extension__ ({ \
50520  float32x2_t __s0_422 = __p0_422; \
50521  float32x2_t __s1_422 = __p1_422; \
50522  float32x4_t __s2_422 = __p2_422; \
50523  float32x2_t __rev0_422; __rev0_422 = __builtin_shufflevector(__s0_422, __s0_422, 1, 0); \
50524  float32x2_t __rev1_422; __rev1_422 = __builtin_shufflevector(__s1_422, __s1_422, 1, 0); \
50525  float32x4_t __rev2_422; __rev2_422 = __builtin_shufflevector(__s2_422, __s2_422, 3, 2, 1, 0); \
50526  float32x2_t __ret_422; \
50527  __ret_422 = __noswap_vfma_laneq_f32(__rev0_422, -__rev1_422, __rev2_422, __p3_422); \
50528  __ret_422 = __builtin_shufflevector(__ret_422, __ret_422, 1, 0); \
50529  __ret_422; \
50530 })
50531 #endif
50532 
50533 #ifdef __LITTLE_ENDIAN__
50534 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
50535  float64x2_t __ret;
50536  __ret = vfmaq_f64(__p0, -__p1, (float64x2_t) {__p2, __p2});
50537  return __ret;
50538 }
50539 #else
50540 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
50541  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50542  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50543  float64x2_t __ret;
50544  __ret = __noswap_vfmaq_f64(__rev0, -__rev1, (float64x2_t) {__p2, __p2});
50545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50546  return __ret;
50547 }
50548 #endif
50549 
50550 #ifdef __LITTLE_ENDIAN__
50551 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
50552  float32x4_t __ret;
50553  __ret = vfmaq_f32(__p0, -__p1, (float32x4_t) {__p2, __p2, __p2, __p2});
50554  return __ret;
50555 }
50556 #else
50557 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
50558  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50559  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50560  float32x4_t __ret;
50561  __ret = __noswap_vfmaq_f32(__rev0, -__rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
50562  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50563  return __ret;
50564 }
50565 #endif
50566 
50567 __ai float64x1_t vfms_n_f64(float64x1_t __p0, float64x1_t __p1, float64_t __p2) {
50568  float64x1_t __ret;
50569  __ret = vfma_f64(__p0, -__p1, (float64x1_t) {__p2});
50570  return __ret;
50571 }
50572 #ifdef __LITTLE_ENDIAN__
50573 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
50574  float32x2_t __ret;
50575  __ret = vfma_f32(__p0, -__p1, (float32x2_t) {__p2, __p2});
50576  return __ret;
50577 }
50578 #else
50579 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
50580  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50581  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50582  float32x2_t __ret;
50583  __ret = __noswap_vfma_f32(__rev0, -__rev1, (float32x2_t) {__p2, __p2});
50584  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50585  return __ret;
50586 }
50587 #endif
50588 
50589 #ifdef __LITTLE_ENDIAN__
50590 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
50591  poly64x1_t __ret;
50592  __ret = __builtin_shufflevector(__p0, __p0, 1);
50593  return __ret;
50594 }
50595 #else
50596 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
50597  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50598  poly64x1_t __ret;
50599  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
50600  return __ret;
50601 }
50602 __ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
50603  poly64x1_t __ret;
50604  __ret = __builtin_shufflevector(__p0, __p0, 1);
50605  return __ret;
50606 }
50607 #endif
50608 
50609 #ifdef __LITTLE_ENDIAN__
50610 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
50611  float64x1_t __ret;
50612  __ret = __builtin_shufflevector(__p0, __p0, 1);
50613  return __ret;
50614 }
50615 #else
50616 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
50617  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50618  float64x1_t __ret;
50619  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
50620  return __ret;
50621 }
50622 #endif
50623 
50624 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
50625  poly64x1_t __s0 = __p0; \
50626  poly64_t __ret; \
50627  __ret = (poly64_t) __builtin_neon_vget_lane_i64((poly64x1_t)__s0, __p1); \
50628  __ret; \
50629 })
50630 #ifdef __LITTLE_ENDIAN__
50631 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
50632  poly64x2_t __s0 = __p0; \
50633  poly64_t __ret; \
50634  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__s0, __p1); \
50635  __ret; \
50636 })
50637 #else
50638 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
50639  poly64x2_t __s0 = __p0; \
50640  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
50641  poly64_t __ret; \
50642  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__rev0, __p1); \
50643  __ret; \
50644 })
50645 #define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
50646  poly64x2_t __s0 = __p0; \
50647  poly64_t __ret; \
50648  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((poly64x2_t)__s0, __p1); \
50649  __ret; \
50650 })
50651 #endif
50652 
50653 #ifdef __LITTLE_ENDIAN__
50654 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
50655  float64x2_t __s0 = __p0; \
50656  float64_t __ret; \
50657  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__s0, __p1); \
50658  __ret; \
50659 })
50660 #else
50661 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
50662  float64x2_t __s0 = __p0; \
50663  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
50664  float64_t __ret; \
50665  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__rev0, __p1); \
50666  __ret; \
50667 })
50668 #define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
50669  float64x2_t __s0 = __p0; \
50670  float64_t __ret; \
50671  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((float64x2_t)__s0, __p1); \
50672  __ret; \
50673 })
50674 #endif
50675 
50676 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
50677  float64x1_t __s0 = __p0; \
50678  float64_t __ret; \
50679  __ret = (float64_t) __builtin_neon_vget_lane_f64((float64x1_t)__s0, __p1); \
50680  __ret; \
50681 })
50682 #ifdef __LITTLE_ENDIAN__
50683 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
50684  poly64x1_t __ret;
50685  __ret = __builtin_shufflevector(__p0, __p0, 0);
50686  return __ret;
50687 }
50688 #else
50689 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
50690  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50691  poly64x1_t __ret;
50692  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
50693  return __ret;
50694 }
50695 #endif
50696 
50697 #ifdef __LITTLE_ENDIAN__
50698 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
50699  float64x1_t __ret;
50700  __ret = __builtin_shufflevector(__p0, __p0, 0);
50701  return __ret;
50702 }
50703 #else
50704 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
50705  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50706  float64x1_t __ret;
50707  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
50708  return __ret;
50709 }
50710 #endif
50711 
50712 #define vld1_p64(__p0) __extension__ ({ \
50713  poly64x1_t __ret; \
50714  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
50715  __ret; \
50716 })
50717 #ifdef __LITTLE_ENDIAN__
50718 #define vld1q_p64(__p0) __extension__ ({ \
50719  poly64x2_t __ret; \
50720  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
50721  __ret; \
50722 })
50723 #else
50724 #define vld1q_p64(__p0) __extension__ ({ \
50725  poly64x2_t __ret; \
50726  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
50727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50728  __ret; \
50729 })
50730 #endif
50731 
50732 #ifdef __LITTLE_ENDIAN__
50733 #define vld1q_f64(__p0) __extension__ ({ \
50734  float64x2_t __ret; \
50735  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
50736  __ret; \
50737 })
50738 #else
50739 #define vld1q_f64(__p0) __extension__ ({ \
50740  float64x2_t __ret; \
50741  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
50742  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50743  __ret; \
50744 })
50745 #endif
50746 
50747 #define vld1_f64(__p0) __extension__ ({ \
50748  float64x1_t __ret; \
50749  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
50750  __ret; \
50751 })
50752 #define vld1_dup_p64(__p0) __extension__ ({ \
50753  poly64x1_t __ret; \
50754  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
50755  __ret; \
50756 })
50757 #ifdef __LITTLE_ENDIAN__
50758 #define vld1q_dup_p64(__p0) __extension__ ({ \
50759  poly64x2_t __ret; \
50760  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
50761  __ret; \
50762 })
50763 #else
50764 #define vld1q_dup_p64(__p0) __extension__ ({ \
50765  poly64x2_t __ret; \
50766  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
50767  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50768  __ret; \
50769 })
50770 #endif
50771 
50772 #ifdef __LITTLE_ENDIAN__
50773 #define vld1q_dup_f64(__p0) __extension__ ({ \
50774  float64x2_t __ret; \
50775  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
50776  __ret; \
50777 })
50778 #else
50779 #define vld1q_dup_f64(__p0) __extension__ ({ \
50780  float64x2_t __ret; \
50781  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
50782  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50783  __ret; \
50784 })
50785 #endif
50786 
50787 #define vld1_dup_f64(__p0) __extension__ ({ \
50788  float64x1_t __ret; \
50789  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
50790  __ret; \
50791 })
50792 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50793  poly64x1_t __s1 = __p1; \
50794  poly64x1_t __ret; \
50795  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
50796  __ret; \
50797 })
50798 #ifdef __LITTLE_ENDIAN__
50799 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50800  poly64x2_t __s1 = __p1; \
50801  poly64x2_t __ret; \
50802  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
50803  __ret; \
50804 })
50805 #else
50806 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50807  poly64x2_t __s1 = __p1; \
50808  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50809  poly64x2_t __ret; \
50810  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
50811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50812  __ret; \
50813 })
50814 #endif
50815 
50816 #ifdef __LITTLE_ENDIAN__
50817 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50818  float64x2_t __s1 = __p1; \
50819  float64x2_t __ret; \
50820  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
50821  __ret; \
50822 })
50823 #else
50824 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50825  float64x2_t __s1 = __p1; \
50826  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
50827  float64x2_t __ret; \
50828  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
50829  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
50830  __ret; \
50831 })
50832 #endif
50833 
50834 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50835  float64x1_t __s1 = __p1; \
50836  float64x1_t __ret; \
50837  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
50838  __ret; \
50839 })
50840 #define vld1_p64_x2(__p0) __extension__ ({ \
50841  poly64x1x2_t __ret; \
50842  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
50843  __ret; \
50844 })
50845 #ifdef __LITTLE_ENDIAN__
50846 #define vld1q_p64_x2(__p0) __extension__ ({ \
50847  poly64x2x2_t __ret; \
50848  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
50849  __ret; \
50850 })
50851 #else
50852 #define vld1q_p64_x2(__p0) __extension__ ({ \
50853  poly64x2x2_t __ret; \
50854  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
50855  \
50856  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50857  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50858  __ret; \
50859 })
50860 #endif
50861 
50862 #ifdef __LITTLE_ENDIAN__
50863 #define vld1q_f64_x2(__p0) __extension__ ({ \
50864  float64x2x2_t __ret; \
50865  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
50866  __ret; \
50867 })
50868 #else
50869 #define vld1q_f64_x2(__p0) __extension__ ({ \
50870  float64x2x2_t __ret; \
50871  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
50872  \
50873  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50874  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50875  __ret; \
50876 })
50877 #endif
50878 
50879 #define vld1_f64_x2(__p0) __extension__ ({ \
50880  float64x1x2_t __ret; \
50881  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
50882  __ret; \
50883 })
50884 #define vld1_p64_x3(__p0) __extension__ ({ \
50885  poly64x1x3_t __ret; \
50886  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
50887  __ret; \
50888 })
50889 #ifdef __LITTLE_ENDIAN__
50890 #define vld1q_p64_x3(__p0) __extension__ ({ \
50891  poly64x2x3_t __ret; \
50892  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
50893  __ret; \
50894 })
50895 #else
50896 #define vld1q_p64_x3(__p0) __extension__ ({ \
50897  poly64x2x3_t __ret; \
50898  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
50899  \
50900  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50901  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50902  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50903  __ret; \
50904 })
50905 #endif
50906 
50907 #ifdef __LITTLE_ENDIAN__
50908 #define vld1q_f64_x3(__p0) __extension__ ({ \
50909  float64x2x3_t __ret; \
50910  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
50911  __ret; \
50912 })
50913 #else
50914 #define vld1q_f64_x3(__p0) __extension__ ({ \
50915  float64x2x3_t __ret; \
50916  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
50917  \
50918  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50919  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50920  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50921  __ret; \
50922 })
50923 #endif
50924 
50925 #define vld1_f64_x3(__p0) __extension__ ({ \
50926  float64x1x3_t __ret; \
50927  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
50928  __ret; \
50929 })
50930 #define vld1_p64_x4(__p0) __extension__ ({ \
50931  poly64x1x4_t __ret; \
50932  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
50933  __ret; \
50934 })
50935 #ifdef __LITTLE_ENDIAN__
50936 #define vld1q_p64_x4(__p0) __extension__ ({ \
50937  poly64x2x4_t __ret; \
50938  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
50939  __ret; \
50940 })
50941 #else
50942 #define vld1q_p64_x4(__p0) __extension__ ({ \
50943  poly64x2x4_t __ret; \
50944  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
50945  \
50946  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50947  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50948  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50949  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50950  __ret; \
50951 })
50952 #endif
50953 
50954 #ifdef __LITTLE_ENDIAN__
50955 #define vld1q_f64_x4(__p0) __extension__ ({ \
50956  float64x2x4_t __ret; \
50957  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
50958  __ret; \
50959 })
50960 #else
50961 #define vld1q_f64_x4(__p0) __extension__ ({ \
50962  float64x2x4_t __ret; \
50963  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
50964  \
50965  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50966  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50967  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50968  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50969  __ret; \
50970 })
50971 #endif
50972 
50973 #define vld1_f64_x4(__p0) __extension__ ({ \
50974  float64x1x4_t __ret; \
50975  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
50976  __ret; \
50977 })
50978 #define vld2_p64(__p0) __extension__ ({ \
50979  poly64x1x2_t __ret; \
50980  __builtin_neon_vld2_v(&__ret, __p0, 6); \
50981  __ret; \
50982 })
50983 #ifdef __LITTLE_ENDIAN__
50984 #define vld2q_p64(__p0) __extension__ ({ \
50985  poly64x2x2_t __ret; \
50986  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
50987  __ret; \
50988 })
50989 #else
50990 #define vld2q_p64(__p0) __extension__ ({ \
50991  poly64x2x2_t __ret; \
50992  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
50993  \
50994  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50995  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50996  __ret; \
50997 })
50998 #endif
50999 
51000 #ifdef __LITTLE_ENDIAN__
51001 #define vld2q_u64(__p0) __extension__ ({ \
51002  uint64x2x2_t __ret; \
51003  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
51004  __ret; \
51005 })
51006 #else
51007 #define vld2q_u64(__p0) __extension__ ({ \
51008  uint64x2x2_t __ret; \
51009  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
51010  \
51011  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51012  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51013  __ret; \
51014 })
51015 #endif
51016 
51017 #ifdef __LITTLE_ENDIAN__
51018 #define vld2q_f64(__p0) __extension__ ({ \
51019  float64x2x2_t __ret; \
51020  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
51021  __ret; \
51022 })
51023 #else
51024 #define vld2q_f64(__p0) __extension__ ({ \
51025  float64x2x2_t __ret; \
51026  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
51027  \
51028  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51029  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51030  __ret; \
51031 })
51032 #endif
51033 
51034 #ifdef __LITTLE_ENDIAN__
51035 #define vld2q_s64(__p0) __extension__ ({ \
51036  int64x2x2_t __ret; \
51037  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
51038  __ret; \
51039 })
51040 #else
51041 #define vld2q_s64(__p0) __extension__ ({ \
51042  int64x2x2_t __ret; \
51043  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
51044  \
51045  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51046  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51047  __ret; \
51048 })
51049 #endif
51050 
51051 #define vld2_f64(__p0) __extension__ ({ \
51052  float64x1x2_t __ret; \
51053  __builtin_neon_vld2_v(&__ret, __p0, 10); \
51054  __ret; \
51055 })
51056 #define vld2_dup_p64(__p0) __extension__ ({ \
51057  poly64x1x2_t __ret; \
51058  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
51059  __ret; \
51060 })
51061 #ifdef __LITTLE_ENDIAN__
51062 #define vld2q_dup_p64(__p0) __extension__ ({ \
51063  poly64x2x2_t __ret; \
51064  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
51065  __ret; \
51066 })
51067 #else
51068 #define vld2q_dup_p64(__p0) __extension__ ({ \
51069  poly64x2x2_t __ret; \
51070  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
51071  \
51072  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51073  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51074  __ret; \
51075 })
51076 #endif
51077 
51078 #ifdef __LITTLE_ENDIAN__
51079 #define vld2q_dup_f64(__p0) __extension__ ({ \
51080  float64x2x2_t __ret; \
51081  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
51082  __ret; \
51083 })
51084 #else
51085 #define vld2q_dup_f64(__p0) __extension__ ({ \
51086  float64x2x2_t __ret; \
51087  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
51088  \
51089  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51090  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51091  __ret; \
51092 })
51093 #endif
51094 
51095 #define vld2_dup_f64(__p0) __extension__ ({ \
51096  float64x1x2_t __ret; \
51097  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
51098  __ret; \
51099 })
51100 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51101  poly64x1x2_t __s1 = __p1; \
51102  poly64x1x2_t __ret; \
51103  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
51104  __ret; \
51105 })
51106 #ifdef __LITTLE_ENDIAN__
51107 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51108  poly8x16x2_t __s1 = __p1; \
51109  poly8x16x2_t __ret; \
51110  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
51111  __ret; \
51112 })
51113 #else
51114 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51115  poly8x16x2_t __s1 = __p1; \
51116  poly8x16x2_t __rev1; \
51117  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51118  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51119  poly8x16x2_t __ret; \
51120  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
51121  \
51122  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51123  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51124  __ret; \
51125 })
51126 #endif
51127 
51128 #ifdef __LITTLE_ENDIAN__
51129 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51130  poly64x2x2_t __s1 = __p1; \
51131  poly64x2x2_t __ret; \
51132  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
51133  __ret; \
51134 })
51135 #else
51136 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51137  poly64x2x2_t __s1 = __p1; \
51138  poly64x2x2_t __rev1; \
51139  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51140  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51141  poly64x2x2_t __ret; \
51142  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
51143  \
51144  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51145  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51146  __ret; \
51147 })
51148 #endif
51149 
51150 #ifdef __LITTLE_ENDIAN__
51151 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51152  uint8x16x2_t __s1 = __p1; \
51153  uint8x16x2_t __ret; \
51154  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
51155  __ret; \
51156 })
51157 #else
51158 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51159  uint8x16x2_t __s1 = __p1; \
51160  uint8x16x2_t __rev1; \
51161  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51162  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51163  uint8x16x2_t __ret; \
51164  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
51165  \
51166  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51167  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51168  __ret; \
51169 })
51170 #endif
51171 
51172 #ifdef __LITTLE_ENDIAN__
51173 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51174  uint64x2x2_t __s1 = __p1; \
51175  uint64x2x2_t __ret; \
51176  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
51177  __ret; \
51178 })
51179 #else
51180 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51181  uint64x2x2_t __s1 = __p1; \
51182  uint64x2x2_t __rev1; \
51183  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51184  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51185  uint64x2x2_t __ret; \
51186  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
51187  \
51188  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51189  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51190  __ret; \
51191 })
51192 #endif
51193 
51194 #ifdef __LITTLE_ENDIAN__
51195 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51196  int8x16x2_t __s1 = __p1; \
51197  int8x16x2_t __ret; \
51198  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
51199  __ret; \
51200 })
51201 #else
51202 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51203  int8x16x2_t __s1 = __p1; \
51204  int8x16x2_t __rev1; \
51205  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51206  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51207  int8x16x2_t __ret; \
51208  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
51209  \
51210  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51211  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51212  __ret; \
51213 })
51214 #endif
51215 
51216 #ifdef __LITTLE_ENDIAN__
51217 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51218  float64x2x2_t __s1 = __p1; \
51219  float64x2x2_t __ret; \
51220  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 42); \
51221  __ret; \
51222 })
51223 #else
51224 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51225  float64x2x2_t __s1 = __p1; \
51226  float64x2x2_t __rev1; \
51227  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51228  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51229  float64x2x2_t __ret; \
51230  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 42); \
51231  \
51232  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51233  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51234  __ret; \
51235 })
51236 #endif
51237 
51238 #ifdef __LITTLE_ENDIAN__
51239 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51240  int64x2x2_t __s1 = __p1; \
51241  int64x2x2_t __ret; \
51242  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 35); \
51243  __ret; \
51244 })
51245 #else
51246 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51247  int64x2x2_t __s1 = __p1; \
51248  int64x2x2_t __rev1; \
51249  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51250  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51251  int64x2x2_t __ret; \
51252  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 35); \
51253  \
51254  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51255  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51256  __ret; \
51257 })
51258 #endif
51259 
51260 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51261  uint64x1x2_t __s1 = __p1; \
51262  uint64x1x2_t __ret; \
51263  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
51264  __ret; \
51265 })
51266 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51267  float64x1x2_t __s1 = __p1; \
51268  float64x1x2_t __ret; \
51269  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 10); \
51270  __ret; \
51271 })
51272 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51273  int64x1x2_t __s1 = __p1; \
51274  int64x1x2_t __ret; \
51275  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 3); \
51276  __ret; \
51277 })
51278 #define vld3_p64(__p0) __extension__ ({ \
51279  poly64x1x3_t __ret; \
51280  __builtin_neon_vld3_v(&__ret, __p0, 6); \
51281  __ret; \
51282 })
51283 #ifdef __LITTLE_ENDIAN__
51284 #define vld3q_p64(__p0) __extension__ ({ \
51285  poly64x2x3_t __ret; \
51286  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
51287  __ret; \
51288 })
51289 #else
51290 #define vld3q_p64(__p0) __extension__ ({ \
51291  poly64x2x3_t __ret; \
51292  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
51293  \
51294  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51295  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51296  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51297  __ret; \
51298 })
51299 #endif
51300 
51301 #ifdef __LITTLE_ENDIAN__
51302 #define vld3q_u64(__p0) __extension__ ({ \
51303  uint64x2x3_t __ret; \
51304  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
51305  __ret; \
51306 })
51307 #else
51308 #define vld3q_u64(__p0) __extension__ ({ \
51309  uint64x2x3_t __ret; \
51310  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
51311  \
51312  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51313  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51314  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51315  __ret; \
51316 })
51317 #endif
51318 
51319 #ifdef __LITTLE_ENDIAN__
51320 #define vld3q_f64(__p0) __extension__ ({ \
51321  float64x2x3_t __ret; \
51322  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
51323  __ret; \
51324 })
51325 #else
51326 #define vld3q_f64(__p0) __extension__ ({ \
51327  float64x2x3_t __ret; \
51328  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
51329  \
51330  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51331  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51332  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51333  __ret; \
51334 })
51335 #endif
51336 
51337 #ifdef __LITTLE_ENDIAN__
51338 #define vld3q_s64(__p0) __extension__ ({ \
51339  int64x2x3_t __ret; \
51340  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
51341  __ret; \
51342 })
51343 #else
51344 #define vld3q_s64(__p0) __extension__ ({ \
51345  int64x2x3_t __ret; \
51346  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
51347  \
51348  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51349  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51350  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51351  __ret; \
51352 })
51353 #endif
51354 
51355 #define vld3_f64(__p0) __extension__ ({ \
51356  float64x1x3_t __ret; \
51357  __builtin_neon_vld3_v(&__ret, __p0, 10); \
51358  __ret; \
51359 })
51360 #define vld3_dup_p64(__p0) __extension__ ({ \
51361  poly64x1x3_t __ret; \
51362  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
51363  __ret; \
51364 })
51365 #ifdef __LITTLE_ENDIAN__
51366 #define vld3q_dup_p64(__p0) __extension__ ({ \
51367  poly64x2x3_t __ret; \
51368  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
51369  __ret; \
51370 })
51371 #else
51372 #define vld3q_dup_p64(__p0) __extension__ ({ \
51373  poly64x2x3_t __ret; \
51374  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
51375  \
51376  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51377  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51378  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51379  __ret; \
51380 })
51381 #endif
51382 
51383 #ifdef __LITTLE_ENDIAN__
51384 #define vld3q_dup_f64(__p0) __extension__ ({ \
51385  float64x2x3_t __ret; \
51386  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
51387  __ret; \
51388 })
51389 #else
51390 #define vld3q_dup_f64(__p0) __extension__ ({ \
51391  float64x2x3_t __ret; \
51392  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
51393  \
51394  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51395  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51396  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51397  __ret; \
51398 })
51399 #endif
51400 
51401 #define vld3_dup_f64(__p0) __extension__ ({ \
51402  float64x1x3_t __ret; \
51403  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
51404  __ret; \
51405 })
51406 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51407  poly64x1x3_t __s1 = __p1; \
51408  poly64x1x3_t __ret; \
51409  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
51410  __ret; \
51411 })
51412 #ifdef __LITTLE_ENDIAN__
51413 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51414  poly8x16x3_t __s1 = __p1; \
51415  poly8x16x3_t __ret; \
51416  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
51417  __ret; \
51418 })
51419 #else
51420 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51421  poly8x16x3_t __s1 = __p1; \
51422  poly8x16x3_t __rev1; \
51423  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51424  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51425  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51426  poly8x16x3_t __ret; \
51427  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
51428  \
51429  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51430  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51431  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51432  __ret; \
51433 })
51434 #endif
51435 
51436 #ifdef __LITTLE_ENDIAN__
51437 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51438  poly64x2x3_t __s1 = __p1; \
51439  poly64x2x3_t __ret; \
51440  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
51441  __ret; \
51442 })
51443 #else
51444 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51445  poly64x2x3_t __s1 = __p1; \
51446  poly64x2x3_t __rev1; \
51447  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51448  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51449  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51450  poly64x2x3_t __ret; \
51451  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
51452  \
51453  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51454  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51455  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51456  __ret; \
51457 })
51458 #endif
51459 
51460 #ifdef __LITTLE_ENDIAN__
51461 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51462  uint8x16x3_t __s1 = __p1; \
51463  uint8x16x3_t __ret; \
51464  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
51465  __ret; \
51466 })
51467 #else
51468 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51469  uint8x16x3_t __s1 = __p1; \
51470  uint8x16x3_t __rev1; \
51471  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51472  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51473  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51474  uint8x16x3_t __ret; \
51475  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
51476  \
51477  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51478  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51479  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51480  __ret; \
51481 })
51482 #endif
51483 
51484 #ifdef __LITTLE_ENDIAN__
51485 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51486  uint64x2x3_t __s1 = __p1; \
51487  uint64x2x3_t __ret; \
51488  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
51489  __ret; \
51490 })
51491 #else
51492 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51493  uint64x2x3_t __s1 = __p1; \
51494  uint64x2x3_t __rev1; \
51495  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51496  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51497  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51498  uint64x2x3_t __ret; \
51499  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
51500  \
51501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51503  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51504  __ret; \
51505 })
51506 #endif
51507 
51508 #ifdef __LITTLE_ENDIAN__
51509 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51510  int8x16x3_t __s1 = __p1; \
51511  int8x16x3_t __ret; \
51512  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
51513  __ret; \
51514 })
51515 #else
51516 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51517  int8x16x3_t __s1 = __p1; \
51518  int8x16x3_t __rev1; \
51519  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51520  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51521  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51522  int8x16x3_t __ret; \
51523  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
51524  \
51525  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51526  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51527  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51528  __ret; \
51529 })
51530 #endif
51531 
51532 #ifdef __LITTLE_ENDIAN__
51533 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51534  float64x2x3_t __s1 = __p1; \
51535  float64x2x3_t __ret; \
51536  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 42); \
51537  __ret; \
51538 })
51539 #else
51540 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51541  float64x2x3_t __s1 = __p1; \
51542  float64x2x3_t __rev1; \
51543  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51544  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51545  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51546  float64x2x3_t __ret; \
51547  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 42); \
51548  \
51549  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51550  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51551  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51552  __ret; \
51553 })
51554 #endif
51555 
51556 #ifdef __LITTLE_ENDIAN__
51557 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51558  int64x2x3_t __s1 = __p1; \
51559  int64x2x3_t __ret; \
51560  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 35); \
51561  __ret; \
51562 })
51563 #else
51564 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51565  int64x2x3_t __s1 = __p1; \
51566  int64x2x3_t __rev1; \
51567  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51568  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51569  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51570  int64x2x3_t __ret; \
51571  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 35); \
51572  \
51573  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51574  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51575  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51576  __ret; \
51577 })
51578 #endif
51579 
51580 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51581  uint64x1x3_t __s1 = __p1; \
51582  uint64x1x3_t __ret; \
51583  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51584  __ret; \
51585 })
51586 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51587  float64x1x3_t __s1 = __p1; \
51588  float64x1x3_t __ret; \
51589  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 10); \
51590  __ret; \
51591 })
51592 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51593  int64x1x3_t __s1 = __p1; \
51594  int64x1x3_t __ret; \
51595  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 3); \
51596  __ret; \
51597 })
51598 #define vld4_p64(__p0) __extension__ ({ \
51599  poly64x1x4_t __ret; \
51600  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51601  __ret; \
51602 })
51603 #ifdef __LITTLE_ENDIAN__
51604 #define vld4q_p64(__p0) __extension__ ({ \
51605  poly64x2x4_t __ret; \
51606  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51607  __ret; \
51608 })
51609 #else
51610 #define vld4q_p64(__p0) __extension__ ({ \
51611  poly64x2x4_t __ret; \
51612  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51613  \
51614  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51615  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51616  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51617  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51618  __ret; \
51619 })
51620 #endif
51621 
51622 #ifdef __LITTLE_ENDIAN__
51623 #define vld4q_u64(__p0) __extension__ ({ \
51624  uint64x2x4_t __ret; \
51625  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51626  __ret; \
51627 })
51628 #else
51629 #define vld4q_u64(__p0) __extension__ ({ \
51630  uint64x2x4_t __ret; \
51631  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51632  \
51633  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51634  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51635  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51636  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51637  __ret; \
51638 })
51639 #endif
51640 
51641 #ifdef __LITTLE_ENDIAN__
51642 #define vld4q_f64(__p0) __extension__ ({ \
51643  float64x2x4_t __ret; \
51644  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51645  __ret; \
51646 })
51647 #else
51648 #define vld4q_f64(__p0) __extension__ ({ \
51649  float64x2x4_t __ret; \
51650  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51651  \
51652  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51653  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51654  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51655  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51656  __ret; \
51657 })
51658 #endif
51659 
51660 #ifdef __LITTLE_ENDIAN__
51661 #define vld4q_s64(__p0) __extension__ ({ \
51662  int64x2x4_t __ret; \
51663  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51664  __ret; \
51665 })
51666 #else
51667 #define vld4q_s64(__p0) __extension__ ({ \
51668  int64x2x4_t __ret; \
51669  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51670  \
51671  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51672  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51673  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51674  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51675  __ret; \
51676 })
51677 #endif
51678 
51679 #define vld4_f64(__p0) __extension__ ({ \
51680  float64x1x4_t __ret; \
51681  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51682  __ret; \
51683 })
51684 #define vld4_dup_p64(__p0) __extension__ ({ \
51685  poly64x1x4_t __ret; \
51686  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51687  __ret; \
51688 })
51689 #ifdef __LITTLE_ENDIAN__
51690 #define vld4q_dup_p64(__p0) __extension__ ({ \
51691  poly64x2x4_t __ret; \
51692  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51693  __ret; \
51694 })
51695 #else
51696 #define vld4q_dup_p64(__p0) __extension__ ({ \
51697  poly64x2x4_t __ret; \
51698  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51699  \
51700  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51701  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51702  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51703  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51704  __ret; \
51705 })
51706 #endif
51707 
51708 #ifdef __LITTLE_ENDIAN__
51709 #define vld4q_dup_f64(__p0) __extension__ ({ \
51710  float64x2x4_t __ret; \
51711  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51712  __ret; \
51713 })
51714 #else
51715 #define vld4q_dup_f64(__p0) __extension__ ({ \
51716  float64x2x4_t __ret; \
51717  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51718  \
51719  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51720  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51721  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51722  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51723  __ret; \
51724 })
51725 #endif
51726 
51727 #define vld4_dup_f64(__p0) __extension__ ({ \
51728  float64x1x4_t __ret; \
51729  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51730  __ret; \
51731 })
51732 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51733  poly64x1x4_t __s1 = __p1; \
51734  poly64x1x4_t __ret; \
51735  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51736  __ret; \
51737 })
51738 #ifdef __LITTLE_ENDIAN__
51739 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51740  poly8x16x4_t __s1 = __p1; \
51741  poly8x16x4_t __ret; \
51742  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
51743  __ret; \
51744 })
51745 #else
51746 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51747  poly8x16x4_t __s1 = __p1; \
51748  poly8x16x4_t __rev1; \
51749  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51750  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51751  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51752  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51753  poly8x16x4_t __ret; \
51754  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
51755  \
51756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51758  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51759  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51760  __ret; \
51761 })
51762 #endif
51763 
51764 #ifdef __LITTLE_ENDIAN__
51765 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51766  poly64x2x4_t __s1 = __p1; \
51767  poly64x2x4_t __ret; \
51768  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
51769  __ret; \
51770 })
51771 #else
51772 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51773  poly64x2x4_t __s1 = __p1; \
51774  poly64x2x4_t __rev1; \
51775  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51776  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51777  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51778  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51779  poly64x2x4_t __ret; \
51780  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
51781  \
51782  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51783  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51784  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51785  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51786  __ret; \
51787 })
51788 #endif
51789 
51790 #ifdef __LITTLE_ENDIAN__
51791 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51792  uint8x16x4_t __s1 = __p1; \
51793  uint8x16x4_t __ret; \
51794  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
51795  __ret; \
51796 })
51797 #else
51798 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51799  uint8x16x4_t __s1 = __p1; \
51800  uint8x16x4_t __rev1; \
51801  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51802  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51803  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51804  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51805  uint8x16x4_t __ret; \
51806  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
51807  \
51808  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51809  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51810  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51811  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51812  __ret; \
51813 })
51814 #endif
51815 
51816 #ifdef __LITTLE_ENDIAN__
51817 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51818  uint64x2x4_t __s1 = __p1; \
51819  uint64x2x4_t __ret; \
51820  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
51821  __ret; \
51822 })
51823 #else
51824 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51825  uint64x2x4_t __s1 = __p1; \
51826  uint64x2x4_t __rev1; \
51827  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51828  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51829  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51830  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51831  uint64x2x4_t __ret; \
51832  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
51833  \
51834  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51835  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51836  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51837  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51838  __ret; \
51839 })
51840 #endif
51841 
51842 #ifdef __LITTLE_ENDIAN__
51843 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51844  int8x16x4_t __s1 = __p1; \
51845  int8x16x4_t __ret; \
51846  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
51847  __ret; \
51848 })
51849 #else
51850 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51851  int8x16x4_t __s1 = __p1; \
51852  int8x16x4_t __rev1; \
51853  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51854  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51855  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51856  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51857  int8x16x4_t __ret; \
51858  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
51859  \
51860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51863  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51864  __ret; \
51865 })
51866 #endif
51867 
51868 #ifdef __LITTLE_ENDIAN__
51869 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51870  float64x2x4_t __s1 = __p1; \
51871  float64x2x4_t __ret; \
51872  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 42); \
51873  __ret; \
51874 })
51875 #else
51876 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51877  float64x2x4_t __s1 = __p1; \
51878  float64x2x4_t __rev1; \
51879  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51880  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51881  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51882  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51883  float64x2x4_t __ret; \
51884  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 42); \
51885  \
51886  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51887  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51888  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51889  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51890  __ret; \
51891 })
51892 #endif
51893 
51894 #ifdef __LITTLE_ENDIAN__
51895 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51896  int64x2x4_t __s1 = __p1; \
51897  int64x2x4_t __ret; \
51898  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 35); \
51899  __ret; \
51900 })
51901 #else
51902 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51903  int64x2x4_t __s1 = __p1; \
51904  int64x2x4_t __rev1; \
51905  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51906  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51907  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51908  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51909  int64x2x4_t __ret; \
51910  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 35); \
51911  \
51912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51915  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51916  __ret; \
51917 })
51918 #endif
51919 
51920 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51921  uint64x1x4_t __s1 = __p1; \
51922  uint64x1x4_t __ret; \
51923  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51924  __ret; \
51925 })
51926 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51927  float64x1x4_t __s1 = __p1; \
51928  float64x1x4_t __ret; \
51929  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 10); \
51930  __ret; \
51931 })
51932 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51933  int64x1x4_t __s1 = __p1; \
51934  int64x1x4_t __ret; \
51935  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 3); \
51936  __ret; \
51937 })
51938 #define vldrq_p128(__p0) __extension__ ({ \
51939  poly128_t __ret; \
51940  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51941  __ret; \
51942 })
51943 #ifdef __LITTLE_ENDIAN__
51944 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51945  float64x2_t __ret;
51946  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51947  return __ret;
51948 }
51949 #else
51950 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51951  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51952  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51953  float64x2_t __ret;
51954  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51955  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51956  return __ret;
51957 }
51958 #endif
51959 
51960 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51961  float64x1_t __ret;
51962  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51963  return __ret;
51964 }
51965 #ifdef __LITTLE_ENDIAN__
51966 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51967  float64_t __ret;
51968  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64(__p0);
51969  return __ret;
51970 }
51971 #else
51972 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51973  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51974  float64_t __ret;
51975  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64(__rev0);
51976  return __ret;
51977 }
51978 #endif
51979 
51980 #ifdef __LITTLE_ENDIAN__
51981 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51982  float32_t __ret;
51983  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32(__p0);
51984  return __ret;
51985 }
51986 #else
51987 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51988  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51989  float32_t __ret;
51990  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32(__rev0);
51991  return __ret;
51992 }
51993 #endif
51994 
51995 #ifdef __LITTLE_ENDIAN__
51996 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51997  float32_t __ret;
51998  __ret = (float32_t) __builtin_neon_vmaxnmv_f32(__p0);
51999  return __ret;
52000 }
52001 #else
52002 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
52003  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52004  float32_t __ret;
52005  __ret = (float32_t) __builtin_neon_vmaxnmv_f32(__rev0);
52006  return __ret;
52007 }
52008 #endif
52009 
52010 #ifdef __LITTLE_ENDIAN__
52011 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
52012  uint8_t __ret;
52013  __ret = (uint8_t) __builtin_neon_vmaxvq_u8(__p0);
52014  return __ret;
52015 }
52016 #else
52017 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
52018  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52019  uint8_t __ret;
52020  __ret = (uint8_t) __builtin_neon_vmaxvq_u8(__rev0);
52021  return __ret;
52022 }
52023 #endif
52024 
52025 #ifdef __LITTLE_ENDIAN__
52026 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
52027  uint32_t __ret;
52028  __ret = (uint32_t) __builtin_neon_vmaxvq_u32(__p0);
52029  return __ret;
52030 }
52031 #else
52032 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
52033  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52034  uint32_t __ret;
52035  __ret = (uint32_t) __builtin_neon_vmaxvq_u32(__rev0);
52036  return __ret;
52037 }
52038 #endif
52039 
52040 #ifdef __LITTLE_ENDIAN__
52041 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
52042  uint16_t __ret;
52043  __ret = (uint16_t) __builtin_neon_vmaxvq_u16(__p0);
52044  return __ret;
52045 }
52046 #else
52047 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
52048  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52049  uint16_t __ret;
52050  __ret = (uint16_t) __builtin_neon_vmaxvq_u16(__rev0);
52051  return __ret;
52052 }
52053 #endif
52054 
52055 #ifdef __LITTLE_ENDIAN__
52056 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
52057  int8_t __ret;
52058  __ret = (int8_t) __builtin_neon_vmaxvq_s8(__p0);
52059  return __ret;
52060 }
52061 #else
52062 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
52063  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52064  int8_t __ret;
52065  __ret = (int8_t) __builtin_neon_vmaxvq_s8(__rev0);
52066  return __ret;
52067 }
52068 #endif
52069 
52070 #ifdef __LITTLE_ENDIAN__
52071 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
52072  float64_t __ret;
52073  __ret = (float64_t) __builtin_neon_vmaxvq_f64(__p0);
52074  return __ret;
52075 }
52076 #else
52077 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
52078  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52079  float64_t __ret;
52080  __ret = (float64_t) __builtin_neon_vmaxvq_f64(__rev0);
52081  return __ret;
52082 }
52083 #endif
52084 
52085 #ifdef __LITTLE_ENDIAN__
52086 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
52087  float32_t __ret;
52088  __ret = (float32_t) __builtin_neon_vmaxvq_f32(__p0);
52089  return __ret;
52090 }
52091 #else
52092 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
52093  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52094  float32_t __ret;
52095  __ret = (float32_t) __builtin_neon_vmaxvq_f32(__rev0);
52096  return __ret;
52097 }
52098 #endif
52099 
52100 #ifdef __LITTLE_ENDIAN__
52101 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
52102  int32_t __ret;
52103  __ret = (int32_t) __builtin_neon_vmaxvq_s32(__p0);
52104  return __ret;
52105 }
52106 #else
52107 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
52108  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52109  int32_t __ret;
52110  __ret = (int32_t) __builtin_neon_vmaxvq_s32(__rev0);
52111  return __ret;
52112 }
52113 #endif
52114 
52115 #ifdef __LITTLE_ENDIAN__
52116 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
52117  int16_t __ret;
52118  __ret = (int16_t) __builtin_neon_vmaxvq_s16(__p0);
52119  return __ret;
52120 }
52121 #else
52122 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
52123  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52124  int16_t __ret;
52125  __ret = (int16_t) __builtin_neon_vmaxvq_s16(__rev0);
52126  return __ret;
52127 }
52128 #endif
52129 
52130 #ifdef __LITTLE_ENDIAN__
52131 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
52132  uint8_t __ret;
52133  __ret = (uint8_t) __builtin_neon_vmaxv_u8(__p0);
52134  return __ret;
52135 }
52136 #else
52137 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
52138  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52139  uint8_t __ret;
52140  __ret = (uint8_t) __builtin_neon_vmaxv_u8(__rev0);
52141  return __ret;
52142 }
52143 #endif
52144 
52145 #ifdef __LITTLE_ENDIAN__
52146 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
52147  uint32_t __ret;
52148  __ret = (uint32_t) __builtin_neon_vmaxv_u32(__p0);
52149  return __ret;
52150 }
52151 #else
52152 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
52153  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52154  uint32_t __ret;
52155  __ret = (uint32_t) __builtin_neon_vmaxv_u32(__rev0);
52156  return __ret;
52157 }
52158 #endif
52159 
52160 #ifdef __LITTLE_ENDIAN__
52161 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
52162  uint16_t __ret;
52163  __ret = (uint16_t) __builtin_neon_vmaxv_u16(__p0);
52164  return __ret;
52165 }
52166 #else
52167 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
52168  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52169  uint16_t __ret;
52170  __ret = (uint16_t) __builtin_neon_vmaxv_u16(__rev0);
52171  return __ret;
52172 }
52173 #endif
52174 
52175 #ifdef __LITTLE_ENDIAN__
52176 __ai int8_t vmaxv_s8(int8x8_t __p0) {
52177  int8_t __ret;
52178  __ret = (int8_t) __builtin_neon_vmaxv_s8(__p0);
52179  return __ret;
52180 }
52181 #else
52182 __ai int8_t vmaxv_s8(int8x8_t __p0) {
52183  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52184  int8_t __ret;
52185  __ret = (int8_t) __builtin_neon_vmaxv_s8(__rev0);
52186  return __ret;
52187 }
52188 #endif
52189 
52190 #ifdef __LITTLE_ENDIAN__
52191 __ai float32_t vmaxv_f32(float32x2_t __p0) {
52192  float32_t __ret;
52193  __ret = (float32_t) __builtin_neon_vmaxv_f32(__p0);
52194  return __ret;
52195 }
52196 #else
52197 __ai float32_t vmaxv_f32(float32x2_t __p0) {
52198  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52199  float32_t __ret;
52200  __ret = (float32_t) __builtin_neon_vmaxv_f32(__rev0);
52201  return __ret;
52202 }
52203 #endif
52204 
52205 #ifdef __LITTLE_ENDIAN__
52206 __ai int32_t vmaxv_s32(int32x2_t __p0) {
52207  int32_t __ret;
52208  __ret = (int32_t) __builtin_neon_vmaxv_s32(__p0);
52209  return __ret;
52210 }
52211 #else
52212 __ai int32_t vmaxv_s32(int32x2_t __p0) {
52213  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52214  int32_t __ret;
52215  __ret = (int32_t) __builtin_neon_vmaxv_s32(__rev0);
52216  return __ret;
52217 }
52218 #endif
52219 
52220 #ifdef __LITTLE_ENDIAN__
52221 __ai int16_t vmaxv_s16(int16x4_t __p0) {
52222  int16_t __ret;
52223  __ret = (int16_t) __builtin_neon_vmaxv_s16(__p0);
52224  return __ret;
52225 }
52226 #else
52227 __ai int16_t vmaxv_s16(int16x4_t __p0) {
52228  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52229  int16_t __ret;
52230  __ret = (int16_t) __builtin_neon_vmaxv_s16(__rev0);
52231  return __ret;
52232 }
52233 #endif
52234 
52235 #ifdef __LITTLE_ENDIAN__
52236 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52237  float64x2_t __ret;
52238  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52239  return __ret;
52240 }
52241 #else
52242 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52243  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52244  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52245  float64x2_t __ret;
52246  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52247  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52248  return __ret;
52249 }
52250 #endif
52251 
52252 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52253  float64x1_t __ret;
52254  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52255  return __ret;
52256 }
52257 #ifdef __LITTLE_ENDIAN__
52258 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
52259  float64_t __ret;
52260  __ret = (float64_t) __builtin_neon_vminnmvq_f64(__p0);
52261  return __ret;
52262 }
52263 #else
52264 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
52265  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52266  float64_t __ret;
52267  __ret = (float64_t) __builtin_neon_vminnmvq_f64(__rev0);
52268  return __ret;
52269 }
52270 #endif
52271 
52272 #ifdef __LITTLE_ENDIAN__
52273 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
52274  float32_t __ret;
52275  __ret = (float32_t) __builtin_neon_vminnmvq_f32(__p0);
52276  return __ret;
52277 }
52278 #else
52279 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
52280  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52281  float32_t __ret;
52282  __ret = (float32_t) __builtin_neon_vminnmvq_f32(__rev0);
52283  return __ret;
52284 }
52285 #endif
52286 
52287 #ifdef __LITTLE_ENDIAN__
52288 __ai float32_t vminnmv_f32(float32x2_t __p0) {
52289  float32_t __ret;
52290  __ret = (float32_t) __builtin_neon_vminnmv_f32(__p0);
52291  return __ret;
52292 }
52293 #else
52294 __ai float32_t vminnmv_f32(float32x2_t __p0) {
52295  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52296  float32_t __ret;
52297  __ret = (float32_t) __builtin_neon_vminnmv_f32(__rev0);
52298  return __ret;
52299 }
52300 #endif
52301 
52302 #ifdef __LITTLE_ENDIAN__
52303 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
52304  uint8_t __ret;
52305  __ret = (uint8_t) __builtin_neon_vminvq_u8(__p0);
52306  return __ret;
52307 }
52308 #else
52309 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
52310  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52311  uint8_t __ret;
52312  __ret = (uint8_t) __builtin_neon_vminvq_u8(__rev0);
52313  return __ret;
52314 }
52315 #endif
52316 
52317 #ifdef __LITTLE_ENDIAN__
52318 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
52319  uint32_t __ret;
52320  __ret = (uint32_t) __builtin_neon_vminvq_u32(__p0);
52321  return __ret;
52322 }
52323 #else
52324 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
52325  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52326  uint32_t __ret;
52327  __ret = (uint32_t) __builtin_neon_vminvq_u32(__rev0);
52328  return __ret;
52329 }
52330 #endif
52331 
52332 #ifdef __LITTLE_ENDIAN__
52333 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
52334  uint16_t __ret;
52335  __ret = (uint16_t) __builtin_neon_vminvq_u16(__p0);
52336  return __ret;
52337 }
52338 #else
52339 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
52340  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52341  uint16_t __ret;
52342  __ret = (uint16_t) __builtin_neon_vminvq_u16(__rev0);
52343  return __ret;
52344 }
52345 #endif
52346 
52347 #ifdef __LITTLE_ENDIAN__
52348 __ai int8_t vminvq_s8(int8x16_t __p0) {
52349  int8_t __ret;
52350  __ret = (int8_t) __builtin_neon_vminvq_s8(__p0);
52351  return __ret;
52352 }
52353 #else
52354 __ai int8_t vminvq_s8(int8x16_t __p0) {
52355  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52356  int8_t __ret;
52357  __ret = (int8_t) __builtin_neon_vminvq_s8(__rev0);
52358  return __ret;
52359 }
52360 #endif
52361 
52362 #ifdef __LITTLE_ENDIAN__
52363 __ai float64_t vminvq_f64(float64x2_t __p0) {
52364  float64_t __ret;
52365  __ret = (float64_t) __builtin_neon_vminvq_f64(__p0);
52366  return __ret;
52367 }
52368 #else
52369 __ai float64_t vminvq_f64(float64x2_t __p0) {
52370  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52371  float64_t __ret;
52372  __ret = (float64_t) __builtin_neon_vminvq_f64(__rev0);
52373  return __ret;
52374 }
52375 #endif
52376 
52377 #ifdef __LITTLE_ENDIAN__
52378 __ai float32_t vminvq_f32(float32x4_t __p0) {
52379  float32_t __ret;
52380  __ret = (float32_t) __builtin_neon_vminvq_f32(__p0);
52381  return __ret;
52382 }
52383 #else
52384 __ai float32_t vminvq_f32(float32x4_t __p0) {
52385  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52386  float32_t __ret;
52387  __ret = (float32_t) __builtin_neon_vminvq_f32(__rev0);
52388  return __ret;
52389 }
52390 #endif
52391 
52392 #ifdef __LITTLE_ENDIAN__
52393 __ai int32_t vminvq_s32(int32x4_t __p0) {
52394  int32_t __ret;
52395  __ret = (int32_t) __builtin_neon_vminvq_s32(__p0);
52396  return __ret;
52397 }
52398 #else
52399 __ai int32_t vminvq_s32(int32x4_t __p0) {
52400  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52401  int32_t __ret;
52402  __ret = (int32_t) __builtin_neon_vminvq_s32(__rev0);
52403  return __ret;
52404 }
52405 #endif
52406 
52407 #ifdef __LITTLE_ENDIAN__
52408 __ai int16_t vminvq_s16(int16x8_t __p0) {
52409  int16_t __ret;
52410  __ret = (int16_t) __builtin_neon_vminvq_s16(__p0);
52411  return __ret;
52412 }
52413 #else
52414 __ai int16_t vminvq_s16(int16x8_t __p0) {
52415  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52416  int16_t __ret;
52417  __ret = (int16_t) __builtin_neon_vminvq_s16(__rev0);
52418  return __ret;
52419 }
52420 #endif
52421 
52422 #ifdef __LITTLE_ENDIAN__
52423 __ai uint8_t vminv_u8(uint8x8_t __p0) {
52424  uint8_t __ret;
52425  __ret = (uint8_t) __builtin_neon_vminv_u8(__p0);
52426  return __ret;
52427 }
52428 #else
52429 __ai uint8_t vminv_u8(uint8x8_t __p0) {
52430  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52431  uint8_t __ret;
52432  __ret = (uint8_t) __builtin_neon_vminv_u8(__rev0);
52433  return __ret;
52434 }
52435 #endif
52436 
52437 #ifdef __LITTLE_ENDIAN__
52438 __ai uint32_t vminv_u32(uint32x2_t __p0) {
52439  uint32_t __ret;
52440  __ret = (uint32_t) __builtin_neon_vminv_u32(__p0);
52441  return __ret;
52442 }
52443 #else
52444 __ai uint32_t vminv_u32(uint32x2_t __p0) {
52445  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52446  uint32_t __ret;
52447  __ret = (uint32_t) __builtin_neon_vminv_u32(__rev0);
52448  return __ret;
52449 }
52450 #endif
52451 
52452 #ifdef __LITTLE_ENDIAN__
52453 __ai uint16_t vminv_u16(uint16x4_t __p0) {
52454  uint16_t __ret;
52455  __ret = (uint16_t) __builtin_neon_vminv_u16(__p0);
52456  return __ret;
52457 }
52458 #else
52459 __ai uint16_t vminv_u16(uint16x4_t __p0) {
52460  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52461  uint16_t __ret;
52462  __ret = (uint16_t) __builtin_neon_vminv_u16(__rev0);
52463  return __ret;
52464 }
52465 #endif
52466 
52467 #ifdef __LITTLE_ENDIAN__
52468 __ai int8_t vminv_s8(int8x8_t __p0) {
52469  int8_t __ret;
52470  __ret = (int8_t) __builtin_neon_vminv_s8(__p0);
52471  return __ret;
52472 }
52473 #else
52474 __ai int8_t vminv_s8(int8x8_t __p0) {
52475  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52476  int8_t __ret;
52477  __ret = (int8_t) __builtin_neon_vminv_s8(__rev0);
52478  return __ret;
52479 }
52480 #endif
52481 
52482 #ifdef __LITTLE_ENDIAN__
52483 __ai float32_t vminv_f32(float32x2_t __p0) {
52484  float32_t __ret;
52485  __ret = (float32_t) __builtin_neon_vminv_f32(__p0);
52486  return __ret;
52487 }
52488 #else
52489 __ai float32_t vminv_f32(float32x2_t __p0) {
52490  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52491  float32_t __ret;
52492  __ret = (float32_t) __builtin_neon_vminv_f32(__rev0);
52493  return __ret;
52494 }
52495 #endif
52496 
52497 #ifdef __LITTLE_ENDIAN__
52498 __ai int32_t vminv_s32(int32x2_t __p0) {
52499  int32_t __ret;
52500  __ret = (int32_t) __builtin_neon_vminv_s32(__p0);
52501  return __ret;
52502 }
52503 #else
52504 __ai int32_t vminv_s32(int32x2_t __p0) {
52505  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52506  int32_t __ret;
52507  __ret = (int32_t) __builtin_neon_vminv_s32(__rev0);
52508  return __ret;
52509 }
52510 #endif
52511 
52512 #ifdef __LITTLE_ENDIAN__
52513 __ai int16_t vminv_s16(int16x4_t __p0) {
52514  int16_t __ret;
52515  __ret = (int16_t) __builtin_neon_vminv_s16(__p0);
52516  return __ret;
52517 }
52518 #else
52519 __ai int16_t vminv_s16(int16x4_t __p0) {
52520  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52521  int16_t __ret;
52522  __ret = (int16_t) __builtin_neon_vminv_s16(__rev0);
52523  return __ret;
52524 }
52525 #endif
52526 
52527 #ifdef __LITTLE_ENDIAN__
52528 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52529  float64x2_t __ret;
52530  __ret = __p0 + __p1 * __p2;
52531  return __ret;
52532 }
52533 #else
52534 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52535  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52536  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52537  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52538  float64x2_t __ret;
52539  __ret = __rev0 + __rev1 * __rev2;
52540  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52541  return __ret;
52542 }
52543 #endif
52544 
52545 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52546  float64x1_t __ret;
52547  __ret = __p0 + __p1 * __p2;
52548  return __ret;
52549 }
52550 #ifdef __LITTLE_ENDIAN__
52551 #define vmlaq_laneq_u32(__p0_423, __p1_423, __p2_423, __p3_423) __extension__ ({ \
52552  uint32x4_t __s0_423 = __p0_423; \
52553  uint32x4_t __s1_423 = __p1_423; \
52554  uint32x4_t __s2_423 = __p2_423; \
52555  uint32x4_t __ret_423; \
52556  __ret_423 = __s0_423 + __s1_423 * splatq_laneq_u32(__s2_423, __p3_423); \
52557  __ret_423; \
52558 })
52559 #else
52560 #define vmlaq_laneq_u32(__p0_424, __p1_424, __p2_424, __p3_424) __extension__ ({ \
52561  uint32x4_t __s0_424 = __p0_424; \
52562  uint32x4_t __s1_424 = __p1_424; \
52563  uint32x4_t __s2_424 = __p2_424; \
52564  uint32x4_t __rev0_424; __rev0_424 = __builtin_shufflevector(__s0_424, __s0_424, 3, 2, 1, 0); \
52565  uint32x4_t __rev1_424; __rev1_424 = __builtin_shufflevector(__s1_424, __s1_424, 3, 2, 1, 0); \
52566  uint32x4_t __rev2_424; __rev2_424 = __builtin_shufflevector(__s2_424, __s2_424, 3, 2, 1, 0); \
52567  uint32x4_t __ret_424; \
52568  __ret_424 = __rev0_424 + __rev1_424 * __noswap_splatq_laneq_u32(__rev2_424, __p3_424); \
52569  __ret_424 = __builtin_shufflevector(__ret_424, __ret_424, 3, 2, 1, 0); \
52570  __ret_424; \
52571 })
52572 #endif
52573 
52574 #ifdef __LITTLE_ENDIAN__
52575 #define vmlaq_laneq_u16(__p0_425, __p1_425, __p2_425, __p3_425) __extension__ ({ \
52576  uint16x8_t __s0_425 = __p0_425; \
52577  uint16x8_t __s1_425 = __p1_425; \
52578  uint16x8_t __s2_425 = __p2_425; \
52579  uint16x8_t __ret_425; \
52580  __ret_425 = __s0_425 + __s1_425 * splatq_laneq_u16(__s2_425, __p3_425); \
52581  __ret_425; \
52582 })
52583 #else
52584 #define vmlaq_laneq_u16(__p0_426, __p1_426, __p2_426, __p3_426) __extension__ ({ \
52585  uint16x8_t __s0_426 = __p0_426; \
52586  uint16x8_t __s1_426 = __p1_426; \
52587  uint16x8_t __s2_426 = __p2_426; \
52588  uint16x8_t __rev0_426; __rev0_426 = __builtin_shufflevector(__s0_426, __s0_426, 7, 6, 5, 4, 3, 2, 1, 0); \
52589  uint16x8_t __rev1_426; __rev1_426 = __builtin_shufflevector(__s1_426, __s1_426, 7, 6, 5, 4, 3, 2, 1, 0); \
52590  uint16x8_t __rev2_426; __rev2_426 = __builtin_shufflevector(__s2_426, __s2_426, 7, 6, 5, 4, 3, 2, 1, 0); \
52591  uint16x8_t __ret_426; \
52592  __ret_426 = __rev0_426 + __rev1_426 * __noswap_splatq_laneq_u16(__rev2_426, __p3_426); \
52593  __ret_426 = __builtin_shufflevector(__ret_426, __ret_426, 7, 6, 5, 4, 3, 2, 1, 0); \
52594  __ret_426; \
52595 })
52596 #endif
52597 
52598 #ifdef __LITTLE_ENDIAN__
52599 #define vmlaq_laneq_f32(__p0_427, __p1_427, __p2_427, __p3_427) __extension__ ({ \
52600  float32x4_t __s0_427 = __p0_427; \
52601  float32x4_t __s1_427 = __p1_427; \
52602  float32x4_t __s2_427 = __p2_427; \
52603  float32x4_t __ret_427; \
52604  __ret_427 = __s0_427 + __s1_427 * splatq_laneq_f32(__s2_427, __p3_427); \
52605  __ret_427; \
52606 })
52607 #else
52608 #define vmlaq_laneq_f32(__p0_428, __p1_428, __p2_428, __p3_428) __extension__ ({ \
52609  float32x4_t __s0_428 = __p0_428; \
52610  float32x4_t __s1_428 = __p1_428; \
52611  float32x4_t __s2_428 = __p2_428; \
52612  float32x4_t __rev0_428; __rev0_428 = __builtin_shufflevector(__s0_428, __s0_428, 3, 2, 1, 0); \
52613  float32x4_t __rev1_428; __rev1_428 = __builtin_shufflevector(__s1_428, __s1_428, 3, 2, 1, 0); \
52614  float32x4_t __rev2_428; __rev2_428 = __builtin_shufflevector(__s2_428, __s2_428, 3, 2, 1, 0); \
52615  float32x4_t __ret_428; \
52616  __ret_428 = __rev0_428 + __rev1_428 * __noswap_splatq_laneq_f32(__rev2_428, __p3_428); \
52617  __ret_428 = __builtin_shufflevector(__ret_428, __ret_428, 3, 2, 1, 0); \
52618  __ret_428; \
52619 })
52620 #endif
52621 
52622 #ifdef __LITTLE_ENDIAN__
52623 #define vmlaq_laneq_s32(__p0_429, __p1_429, __p2_429, __p3_429) __extension__ ({ \
52624  int32x4_t __s0_429 = __p0_429; \
52625  int32x4_t __s1_429 = __p1_429; \
52626  int32x4_t __s2_429 = __p2_429; \
52627  int32x4_t __ret_429; \
52628  __ret_429 = __s0_429 + __s1_429 * splatq_laneq_s32(__s2_429, __p3_429); \
52629  __ret_429; \
52630 })
52631 #else
52632 #define vmlaq_laneq_s32(__p0_430, __p1_430, __p2_430, __p3_430) __extension__ ({ \
52633  int32x4_t __s0_430 = __p0_430; \
52634  int32x4_t __s1_430 = __p1_430; \
52635  int32x4_t __s2_430 = __p2_430; \
52636  int32x4_t __rev0_430; __rev0_430 = __builtin_shufflevector(__s0_430, __s0_430, 3, 2, 1, 0); \
52637  int32x4_t __rev1_430; __rev1_430 = __builtin_shufflevector(__s1_430, __s1_430, 3, 2, 1, 0); \
52638  int32x4_t __rev2_430; __rev2_430 = __builtin_shufflevector(__s2_430, __s2_430, 3, 2, 1, 0); \
52639  int32x4_t __ret_430; \
52640  __ret_430 = __rev0_430 + __rev1_430 * __noswap_splatq_laneq_s32(__rev2_430, __p3_430); \
52641  __ret_430 = __builtin_shufflevector(__ret_430, __ret_430, 3, 2, 1, 0); \
52642  __ret_430; \
52643 })
52644 #endif
52645 
52646 #ifdef __LITTLE_ENDIAN__
52647 #define vmlaq_laneq_s16(__p0_431, __p1_431, __p2_431, __p3_431) __extension__ ({ \
52648  int16x8_t __s0_431 = __p0_431; \
52649  int16x8_t __s1_431 = __p1_431; \
52650  int16x8_t __s2_431 = __p2_431; \
52651  int16x8_t __ret_431; \
52652  __ret_431 = __s0_431 + __s1_431 * splatq_laneq_s16(__s2_431, __p3_431); \
52653  __ret_431; \
52654 })
52655 #else
52656 #define vmlaq_laneq_s16(__p0_432, __p1_432, __p2_432, __p3_432) __extension__ ({ \
52657  int16x8_t __s0_432 = __p0_432; \
52658  int16x8_t __s1_432 = __p1_432; \
52659  int16x8_t __s2_432 = __p2_432; \
52660  int16x8_t __rev0_432; __rev0_432 = __builtin_shufflevector(__s0_432, __s0_432, 7, 6, 5, 4, 3, 2, 1, 0); \
52661  int16x8_t __rev1_432; __rev1_432 = __builtin_shufflevector(__s1_432, __s1_432, 7, 6, 5, 4, 3, 2, 1, 0); \
52662  int16x8_t __rev2_432; __rev2_432 = __builtin_shufflevector(__s2_432, __s2_432, 7, 6, 5, 4, 3, 2, 1, 0); \
52663  int16x8_t __ret_432; \
52664  __ret_432 = __rev0_432 + __rev1_432 * __noswap_splatq_laneq_s16(__rev2_432, __p3_432); \
52665  __ret_432 = __builtin_shufflevector(__ret_432, __ret_432, 7, 6, 5, 4, 3, 2, 1, 0); \
52666  __ret_432; \
52667 })
52668 #endif
52669 
52670 #ifdef __LITTLE_ENDIAN__
52671 #define vmla_laneq_u32(__p0_433, __p1_433, __p2_433, __p3_433) __extension__ ({ \
52672  uint32x2_t __s0_433 = __p0_433; \
52673  uint32x2_t __s1_433 = __p1_433; \
52674  uint32x4_t __s2_433 = __p2_433; \
52675  uint32x2_t __ret_433; \
52676  __ret_433 = __s0_433 + __s1_433 * splat_laneq_u32(__s2_433, __p3_433); \
52677  __ret_433; \
52678 })
52679 #else
52680 #define vmla_laneq_u32(__p0_434, __p1_434, __p2_434, __p3_434) __extension__ ({ \
52681  uint32x2_t __s0_434 = __p0_434; \
52682  uint32x2_t __s1_434 = __p1_434; \
52683  uint32x4_t __s2_434 = __p2_434; \
52684  uint32x2_t __rev0_434; __rev0_434 = __builtin_shufflevector(__s0_434, __s0_434, 1, 0); \
52685  uint32x2_t __rev1_434; __rev1_434 = __builtin_shufflevector(__s1_434, __s1_434, 1, 0); \
52686  uint32x4_t __rev2_434; __rev2_434 = __builtin_shufflevector(__s2_434, __s2_434, 3, 2, 1, 0); \
52687  uint32x2_t __ret_434; \
52688  __ret_434 = __rev0_434 + __rev1_434 * __noswap_splat_laneq_u32(__rev2_434, __p3_434); \
52689  __ret_434 = __builtin_shufflevector(__ret_434, __ret_434, 1, 0); \
52690  __ret_434; \
52691 })
52692 #endif
52693 
52694 #ifdef __LITTLE_ENDIAN__
52695 #define vmla_laneq_u16(__p0_435, __p1_435, __p2_435, __p3_435) __extension__ ({ \
52696  uint16x4_t __s0_435 = __p0_435; \
52697  uint16x4_t __s1_435 = __p1_435; \
52698  uint16x8_t __s2_435 = __p2_435; \
52699  uint16x4_t __ret_435; \
52700  __ret_435 = __s0_435 + __s1_435 * splat_laneq_u16(__s2_435, __p3_435); \
52701  __ret_435; \
52702 })
52703 #else
52704 #define vmla_laneq_u16(__p0_436, __p1_436, __p2_436, __p3_436) __extension__ ({ \
52705  uint16x4_t __s0_436 = __p0_436; \
52706  uint16x4_t __s1_436 = __p1_436; \
52707  uint16x8_t __s2_436 = __p2_436; \
52708  uint16x4_t __rev0_436; __rev0_436 = __builtin_shufflevector(__s0_436, __s0_436, 3, 2, 1, 0); \
52709  uint16x4_t __rev1_436; __rev1_436 = __builtin_shufflevector(__s1_436, __s1_436, 3, 2, 1, 0); \
52710  uint16x8_t __rev2_436; __rev2_436 = __builtin_shufflevector(__s2_436, __s2_436, 7, 6, 5, 4, 3, 2, 1, 0); \
52711  uint16x4_t __ret_436; \
52712  __ret_436 = __rev0_436 + __rev1_436 * __noswap_splat_laneq_u16(__rev2_436, __p3_436); \
52713  __ret_436 = __builtin_shufflevector(__ret_436, __ret_436, 3, 2, 1, 0); \
52714  __ret_436; \
52715 })
52716 #endif
52717 
52718 #ifdef __LITTLE_ENDIAN__
52719 #define vmla_laneq_f32(__p0_437, __p1_437, __p2_437, __p3_437) __extension__ ({ \
52720  float32x2_t __s0_437 = __p0_437; \
52721  float32x2_t __s1_437 = __p1_437; \
52722  float32x4_t __s2_437 = __p2_437; \
52723  float32x2_t __ret_437; \
52724  __ret_437 = __s0_437 + __s1_437 * splat_laneq_f32(__s2_437, __p3_437); \
52725  __ret_437; \
52726 })
52727 #else
52728 #define vmla_laneq_f32(__p0_438, __p1_438, __p2_438, __p3_438) __extension__ ({ \
52729  float32x2_t __s0_438 = __p0_438; \
52730  float32x2_t __s1_438 = __p1_438; \
52731  float32x4_t __s2_438 = __p2_438; \
52732  float32x2_t __rev0_438; __rev0_438 = __builtin_shufflevector(__s0_438, __s0_438, 1, 0); \
52733  float32x2_t __rev1_438; __rev1_438 = __builtin_shufflevector(__s1_438, __s1_438, 1, 0); \
52734  float32x4_t __rev2_438; __rev2_438 = __builtin_shufflevector(__s2_438, __s2_438, 3, 2, 1, 0); \
52735  float32x2_t __ret_438; \
52736  __ret_438 = __rev0_438 + __rev1_438 * __noswap_splat_laneq_f32(__rev2_438, __p3_438); \
52737  __ret_438 = __builtin_shufflevector(__ret_438, __ret_438, 1, 0); \
52738  __ret_438; \
52739 })
52740 #endif
52741 
52742 #ifdef __LITTLE_ENDIAN__
52743 #define vmla_laneq_s32(__p0_439, __p1_439, __p2_439, __p3_439) __extension__ ({ \
52744  int32x2_t __s0_439 = __p0_439; \
52745  int32x2_t __s1_439 = __p1_439; \
52746  int32x4_t __s2_439 = __p2_439; \
52747  int32x2_t __ret_439; \
52748  __ret_439 = __s0_439 + __s1_439 * splat_laneq_s32(__s2_439, __p3_439); \
52749  __ret_439; \
52750 })
52751 #else
52752 #define vmla_laneq_s32(__p0_440, __p1_440, __p2_440, __p3_440) __extension__ ({ \
52753  int32x2_t __s0_440 = __p0_440; \
52754  int32x2_t __s1_440 = __p1_440; \
52755  int32x4_t __s2_440 = __p2_440; \
52756  int32x2_t __rev0_440; __rev0_440 = __builtin_shufflevector(__s0_440, __s0_440, 1, 0); \
52757  int32x2_t __rev1_440; __rev1_440 = __builtin_shufflevector(__s1_440, __s1_440, 1, 0); \
52758  int32x4_t __rev2_440; __rev2_440 = __builtin_shufflevector(__s2_440, __s2_440, 3, 2, 1, 0); \
52759  int32x2_t __ret_440; \
52760  __ret_440 = __rev0_440 + __rev1_440 * __noswap_splat_laneq_s32(__rev2_440, __p3_440); \
52761  __ret_440 = __builtin_shufflevector(__ret_440, __ret_440, 1, 0); \
52762  __ret_440; \
52763 })
52764 #endif
52765 
52766 #ifdef __LITTLE_ENDIAN__
52767 #define vmla_laneq_s16(__p0_441, __p1_441, __p2_441, __p3_441) __extension__ ({ \
52768  int16x4_t __s0_441 = __p0_441; \
52769  int16x4_t __s1_441 = __p1_441; \
52770  int16x8_t __s2_441 = __p2_441; \
52771  int16x4_t __ret_441; \
52772  __ret_441 = __s0_441 + __s1_441 * splat_laneq_s16(__s2_441, __p3_441); \
52773  __ret_441; \
52774 })
52775 #else
52776 #define vmla_laneq_s16(__p0_442, __p1_442, __p2_442, __p3_442) __extension__ ({ \
52777  int16x4_t __s0_442 = __p0_442; \
52778  int16x4_t __s1_442 = __p1_442; \
52779  int16x8_t __s2_442 = __p2_442; \
52780  int16x4_t __rev0_442; __rev0_442 = __builtin_shufflevector(__s0_442, __s0_442, 3, 2, 1, 0); \
52781  int16x4_t __rev1_442; __rev1_442 = __builtin_shufflevector(__s1_442, __s1_442, 3, 2, 1, 0); \
52782  int16x8_t __rev2_442; __rev2_442 = __builtin_shufflevector(__s2_442, __s2_442, 7, 6, 5, 4, 3, 2, 1, 0); \
52783  int16x4_t __ret_442; \
52784  __ret_442 = __rev0_442 + __rev1_442 * __noswap_splat_laneq_s16(__rev2_442, __p3_442); \
52785  __ret_442 = __builtin_shufflevector(__ret_442, __ret_442, 3, 2, 1, 0); \
52786  __ret_442; \
52787 })
52788 #endif
52789 
52790 #ifdef __LITTLE_ENDIAN__
52791 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52792  float64x2_t __ret;
52793  __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
52794  return __ret;
52795 }
52796 #else
52797 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52798  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52799  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52800  float64x2_t __ret;
52801  __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
52802  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52803  return __ret;
52804 }
52805 #endif
52806 
52807 #ifdef __LITTLE_ENDIAN__
52808 #define vmlal_high_lane_u32(__p0_443, __p1_443, __p2_443, __p3_443) __extension__ ({ \
52809  uint64x2_t __s0_443 = __p0_443; \
52810  uint32x4_t __s1_443 = __p1_443; \
52811  uint32x2_t __s2_443 = __p2_443; \
52812  uint64x2_t __ret_443; \
52813  __ret_443 = __s0_443 + vmull_u32(vget_high_u32(__s1_443), splat_lane_u32(__s2_443, __p3_443)); \
52814  __ret_443; \
52815 })
52816 #else
52817 #define vmlal_high_lane_u32(__p0_444, __p1_444, __p2_444, __p3_444) __extension__ ({ \
52818  uint64x2_t __s0_444 = __p0_444; \
52819  uint32x4_t __s1_444 = __p1_444; \
52820  uint32x2_t __s2_444 = __p2_444; \
52821  uint64x2_t __rev0_444; __rev0_444 = __builtin_shufflevector(__s0_444, __s0_444, 1, 0); \
52822  uint32x4_t __rev1_444; __rev1_444 = __builtin_shufflevector(__s1_444, __s1_444, 3, 2, 1, 0); \
52823  uint32x2_t __rev2_444; __rev2_444 = __builtin_shufflevector(__s2_444, __s2_444, 1, 0); \
52824  uint64x2_t __ret_444; \
52825  __ret_444 = __rev0_444 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_444), __noswap_splat_lane_u32(__rev2_444, __p3_444)); \
52826  __ret_444 = __builtin_shufflevector(__ret_444, __ret_444, 1, 0); \
52827  __ret_444; \
52828 })
52829 #endif
52830 
52831 #ifdef __LITTLE_ENDIAN__
52832 #define vmlal_high_lane_u16(__p0_445, __p1_445, __p2_445, __p3_445) __extension__ ({ \
52833  uint32x4_t __s0_445 = __p0_445; \
52834  uint16x8_t __s1_445 = __p1_445; \
52835  uint16x4_t __s2_445 = __p2_445; \
52836  uint32x4_t __ret_445; \
52837  __ret_445 = __s0_445 + vmull_u16(vget_high_u16(__s1_445), splat_lane_u16(__s2_445, __p3_445)); \
52838  __ret_445; \
52839 })
52840 #else
52841 #define vmlal_high_lane_u16(__p0_446, __p1_446, __p2_446, __p3_446) __extension__ ({ \
52842  uint32x4_t __s0_446 = __p0_446; \
52843  uint16x8_t __s1_446 = __p1_446; \
52844  uint16x4_t __s2_446 = __p2_446; \
52845  uint32x4_t __rev0_446; __rev0_446 = __builtin_shufflevector(__s0_446, __s0_446, 3, 2, 1, 0); \
52846  uint16x8_t __rev1_446; __rev1_446 = __builtin_shufflevector(__s1_446, __s1_446, 7, 6, 5, 4, 3, 2, 1, 0); \
52847  uint16x4_t __rev2_446; __rev2_446 = __builtin_shufflevector(__s2_446, __s2_446, 3, 2, 1, 0); \
52848  uint32x4_t __ret_446; \
52849  __ret_446 = __rev0_446 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_446), __noswap_splat_lane_u16(__rev2_446, __p3_446)); \
52850  __ret_446 = __builtin_shufflevector(__ret_446, __ret_446, 3, 2, 1, 0); \
52851  __ret_446; \
52852 })
52853 #endif
52854 
52855 #ifdef __LITTLE_ENDIAN__
52856 #define vmlal_high_lane_s32(__p0_447, __p1_447, __p2_447, __p3_447) __extension__ ({ \
52857  int64x2_t __s0_447 = __p0_447; \
52858  int32x4_t __s1_447 = __p1_447; \
52859  int32x2_t __s2_447 = __p2_447; \
52860  int64x2_t __ret_447; \
52861  __ret_447 = __s0_447 + vmull_s32(vget_high_s32(__s1_447), splat_lane_s32(__s2_447, __p3_447)); \
52862  __ret_447; \
52863 })
52864 #else
52865 #define vmlal_high_lane_s32(__p0_448, __p1_448, __p2_448, __p3_448) __extension__ ({ \
52866  int64x2_t __s0_448 = __p0_448; \
52867  int32x4_t __s1_448 = __p1_448; \
52868  int32x2_t __s2_448 = __p2_448; \
52869  int64x2_t __rev0_448; __rev0_448 = __builtin_shufflevector(__s0_448, __s0_448, 1, 0); \
52870  int32x4_t __rev1_448; __rev1_448 = __builtin_shufflevector(__s1_448, __s1_448, 3, 2, 1, 0); \
52871  int32x2_t __rev2_448; __rev2_448 = __builtin_shufflevector(__s2_448, __s2_448, 1, 0); \
52872  int64x2_t __ret_448; \
52873  __ret_448 = __rev0_448 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_448), __noswap_splat_lane_s32(__rev2_448, __p3_448)); \
52874  __ret_448 = __builtin_shufflevector(__ret_448, __ret_448, 1, 0); \
52875  __ret_448; \
52876 })
52877 #endif
52878 
52879 #ifdef __LITTLE_ENDIAN__
52880 #define vmlal_high_lane_s16(__p0_449, __p1_449, __p2_449, __p3_449) __extension__ ({ \
52881  int32x4_t __s0_449 = __p0_449; \
52882  int16x8_t __s1_449 = __p1_449; \
52883  int16x4_t __s2_449 = __p2_449; \
52884  int32x4_t __ret_449; \
52885  __ret_449 = __s0_449 + vmull_s16(vget_high_s16(__s1_449), splat_lane_s16(__s2_449, __p3_449)); \
52886  __ret_449; \
52887 })
52888 #else
52889 #define vmlal_high_lane_s16(__p0_450, __p1_450, __p2_450, __p3_450) __extension__ ({ \
52890  int32x4_t __s0_450 = __p0_450; \
52891  int16x8_t __s1_450 = __p1_450; \
52892  int16x4_t __s2_450 = __p2_450; \
52893  int32x4_t __rev0_450; __rev0_450 = __builtin_shufflevector(__s0_450, __s0_450, 3, 2, 1, 0); \
52894  int16x8_t __rev1_450; __rev1_450 = __builtin_shufflevector(__s1_450, __s1_450, 7, 6, 5, 4, 3, 2, 1, 0); \
52895  int16x4_t __rev2_450; __rev2_450 = __builtin_shufflevector(__s2_450, __s2_450, 3, 2, 1, 0); \
52896  int32x4_t __ret_450; \
52897  __ret_450 = __rev0_450 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_450), __noswap_splat_lane_s16(__rev2_450, __p3_450)); \
52898  __ret_450 = __builtin_shufflevector(__ret_450, __ret_450, 3, 2, 1, 0); \
52899  __ret_450; \
52900 })
52901 #endif
52902 
52903 #ifdef __LITTLE_ENDIAN__
52904 #define vmlal_high_laneq_u32(__p0_451, __p1_451, __p2_451, __p3_451) __extension__ ({ \
52905  uint64x2_t __s0_451 = __p0_451; \
52906  uint32x4_t __s1_451 = __p1_451; \
52907  uint32x4_t __s2_451 = __p2_451; \
52908  uint64x2_t __ret_451; \
52909  __ret_451 = __s0_451 + vmull_u32(vget_high_u32(__s1_451), splat_laneq_u32(__s2_451, __p3_451)); \
52910  __ret_451; \
52911 })
52912 #else
52913 #define vmlal_high_laneq_u32(__p0_452, __p1_452, __p2_452, __p3_452) __extension__ ({ \
52914  uint64x2_t __s0_452 = __p0_452; \
52915  uint32x4_t __s1_452 = __p1_452; \
52916  uint32x4_t __s2_452 = __p2_452; \
52917  uint64x2_t __rev0_452; __rev0_452 = __builtin_shufflevector(__s0_452, __s0_452, 1, 0); \
52918  uint32x4_t __rev1_452; __rev1_452 = __builtin_shufflevector(__s1_452, __s1_452, 3, 2, 1, 0); \
52919  uint32x4_t __rev2_452; __rev2_452 = __builtin_shufflevector(__s2_452, __s2_452, 3, 2, 1, 0); \
52920  uint64x2_t __ret_452; \
52921  __ret_452 = __rev0_452 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_452), __noswap_splat_laneq_u32(__rev2_452, __p3_452)); \
52922  __ret_452 = __builtin_shufflevector(__ret_452, __ret_452, 1, 0); \
52923  __ret_452; \
52924 })
52925 #endif
52926 
52927 #ifdef __LITTLE_ENDIAN__
52928 #define vmlal_high_laneq_u16(__p0_453, __p1_453, __p2_453, __p3_453) __extension__ ({ \
52929  uint32x4_t __s0_453 = __p0_453; \
52930  uint16x8_t __s1_453 = __p1_453; \
52931  uint16x8_t __s2_453 = __p2_453; \
52932  uint32x4_t __ret_453; \
52933  __ret_453 = __s0_453 + vmull_u16(vget_high_u16(__s1_453), splat_laneq_u16(__s2_453, __p3_453)); \
52934  __ret_453; \
52935 })
52936 #else
52937 #define vmlal_high_laneq_u16(__p0_454, __p1_454, __p2_454, __p3_454) __extension__ ({ \
52938  uint32x4_t __s0_454 = __p0_454; \
52939  uint16x8_t __s1_454 = __p1_454; \
52940  uint16x8_t __s2_454 = __p2_454; \
52941  uint32x4_t __rev0_454; __rev0_454 = __builtin_shufflevector(__s0_454, __s0_454, 3, 2, 1, 0); \
52942  uint16x8_t __rev1_454; __rev1_454 = __builtin_shufflevector(__s1_454, __s1_454, 7, 6, 5, 4, 3, 2, 1, 0); \
52943  uint16x8_t __rev2_454; __rev2_454 = __builtin_shufflevector(__s2_454, __s2_454, 7, 6, 5, 4, 3, 2, 1, 0); \
52944  uint32x4_t __ret_454; \
52945  __ret_454 = __rev0_454 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_454), __noswap_splat_laneq_u16(__rev2_454, __p3_454)); \
52946  __ret_454 = __builtin_shufflevector(__ret_454, __ret_454, 3, 2, 1, 0); \
52947  __ret_454; \
52948 })
52949 #endif
52950 
52951 #ifdef __LITTLE_ENDIAN__
52952 #define vmlal_high_laneq_s32(__p0_455, __p1_455, __p2_455, __p3_455) __extension__ ({ \
52953  int64x2_t __s0_455 = __p0_455; \
52954  int32x4_t __s1_455 = __p1_455; \
52955  int32x4_t __s2_455 = __p2_455; \
52956  int64x2_t __ret_455; \
52957  __ret_455 = __s0_455 + vmull_s32(vget_high_s32(__s1_455), splat_laneq_s32(__s2_455, __p3_455)); \
52958  __ret_455; \
52959 })
52960 #else
52961 #define vmlal_high_laneq_s32(__p0_456, __p1_456, __p2_456, __p3_456) __extension__ ({ \
52962  int64x2_t __s0_456 = __p0_456; \
52963  int32x4_t __s1_456 = __p1_456; \
52964  int32x4_t __s2_456 = __p2_456; \
52965  int64x2_t __rev0_456; __rev0_456 = __builtin_shufflevector(__s0_456, __s0_456, 1, 0); \
52966  int32x4_t __rev1_456; __rev1_456 = __builtin_shufflevector(__s1_456, __s1_456, 3, 2, 1, 0); \
52967  int32x4_t __rev2_456; __rev2_456 = __builtin_shufflevector(__s2_456, __s2_456, 3, 2, 1, 0); \
52968  int64x2_t __ret_456; \
52969  __ret_456 = __rev0_456 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_456), __noswap_splat_laneq_s32(__rev2_456, __p3_456)); \
52970  __ret_456 = __builtin_shufflevector(__ret_456, __ret_456, 1, 0); \
52971  __ret_456; \
52972 })
52973 #endif
52974 
52975 #ifdef __LITTLE_ENDIAN__
52976 #define vmlal_high_laneq_s16(__p0_457, __p1_457, __p2_457, __p3_457) __extension__ ({ \
52977  int32x4_t __s0_457 = __p0_457; \
52978  int16x8_t __s1_457 = __p1_457; \
52979  int16x8_t __s2_457 = __p2_457; \
52980  int32x4_t __ret_457; \
52981  __ret_457 = __s0_457 + vmull_s16(vget_high_s16(__s1_457), splat_laneq_s16(__s2_457, __p3_457)); \
52982  __ret_457; \
52983 })
52984 #else
52985 #define vmlal_high_laneq_s16(__p0_458, __p1_458, __p2_458, __p3_458) __extension__ ({ \
52986  int32x4_t __s0_458 = __p0_458; \
52987  int16x8_t __s1_458 = __p1_458; \
52988  int16x8_t __s2_458 = __p2_458; \
52989  int32x4_t __rev0_458; __rev0_458 = __builtin_shufflevector(__s0_458, __s0_458, 3, 2, 1, 0); \
52990  int16x8_t __rev1_458; __rev1_458 = __builtin_shufflevector(__s1_458, __s1_458, 7, 6, 5, 4, 3, 2, 1, 0); \
52991  int16x8_t __rev2_458; __rev2_458 = __builtin_shufflevector(__s2_458, __s2_458, 7, 6, 5, 4, 3, 2, 1, 0); \
52992  int32x4_t __ret_458; \
52993  __ret_458 = __rev0_458 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_458), __noswap_splat_laneq_s16(__rev2_458, __p3_458)); \
52994  __ret_458 = __builtin_shufflevector(__ret_458, __ret_458, 3, 2, 1, 0); \
52995  __ret_458; \
52996 })
52997 #endif
52998 
52999 #ifdef __LITTLE_ENDIAN__
53000 #define vmlal_laneq_u32(__p0_459, __p1_459, __p2_459, __p3_459) __extension__ ({ \
53001  uint64x2_t __s0_459 = __p0_459; \
53002  uint32x2_t __s1_459 = __p1_459; \
53003  uint32x4_t __s2_459 = __p2_459; \
53004  uint64x2_t __ret_459; \
53005  __ret_459 = __s0_459 + vmull_u32(__s1_459, splat_laneq_u32(__s2_459, __p3_459)); \
53006  __ret_459; \
53007 })
53008 #else
53009 #define vmlal_laneq_u32(__p0_460, __p1_460, __p2_460, __p3_460) __extension__ ({ \
53010  uint64x2_t __s0_460 = __p0_460; \
53011  uint32x2_t __s1_460 = __p1_460; \
53012  uint32x4_t __s2_460 = __p2_460; \
53013  uint64x2_t __rev0_460; __rev0_460 = __builtin_shufflevector(__s0_460, __s0_460, 1, 0); \
53014  uint32x2_t __rev1_460; __rev1_460 = __builtin_shufflevector(__s1_460, __s1_460, 1, 0); \
53015  uint32x4_t __rev2_460; __rev2_460 = __builtin_shufflevector(__s2_460, __s2_460, 3, 2, 1, 0); \
53016  uint64x2_t __ret_460; \
53017  __ret_460 = __rev0_460 + __noswap_vmull_u32(__rev1_460, __noswap_splat_laneq_u32(__rev2_460, __p3_460)); \
53018  __ret_460 = __builtin_shufflevector(__ret_460, __ret_460, 1, 0); \
53019  __ret_460; \
53020 })
53021 #endif
53022 
53023 #ifdef __LITTLE_ENDIAN__
53024 #define vmlal_laneq_u16(__p0_461, __p1_461, __p2_461, __p3_461) __extension__ ({ \
53025  uint32x4_t __s0_461 = __p0_461; \
53026  uint16x4_t __s1_461 = __p1_461; \
53027  uint16x8_t __s2_461 = __p2_461; \
53028  uint32x4_t __ret_461; \
53029  __ret_461 = __s0_461 + vmull_u16(__s1_461, splat_laneq_u16(__s2_461, __p3_461)); \
53030  __ret_461; \
53031 })
53032 #else
53033 #define vmlal_laneq_u16(__p0_462, __p1_462, __p2_462, __p3_462) __extension__ ({ \
53034  uint32x4_t __s0_462 = __p0_462; \
53035  uint16x4_t __s1_462 = __p1_462; \
53036  uint16x8_t __s2_462 = __p2_462; \
53037  uint32x4_t __rev0_462; __rev0_462 = __builtin_shufflevector(__s0_462, __s0_462, 3, 2, 1, 0); \
53038  uint16x4_t __rev1_462; __rev1_462 = __builtin_shufflevector(__s1_462, __s1_462, 3, 2, 1, 0); \
53039  uint16x8_t __rev2_462; __rev2_462 = __builtin_shufflevector(__s2_462, __s2_462, 7, 6, 5, 4, 3, 2, 1, 0); \
53040  uint32x4_t __ret_462; \
53041  __ret_462 = __rev0_462 + __noswap_vmull_u16(__rev1_462, __noswap_splat_laneq_u16(__rev2_462, __p3_462)); \
53042  __ret_462 = __builtin_shufflevector(__ret_462, __ret_462, 3, 2, 1, 0); \
53043  __ret_462; \
53044 })
53045 #endif
53046 
53047 #ifdef __LITTLE_ENDIAN__
53048 #define vmlal_laneq_s32(__p0_463, __p1_463, __p2_463, __p3_463) __extension__ ({ \
53049  int64x2_t __s0_463 = __p0_463; \
53050  int32x2_t __s1_463 = __p1_463; \
53051  int32x4_t __s2_463 = __p2_463; \
53052  int64x2_t __ret_463; \
53053  __ret_463 = __s0_463 + vmull_s32(__s1_463, splat_laneq_s32(__s2_463, __p3_463)); \
53054  __ret_463; \
53055 })
53056 #else
53057 #define vmlal_laneq_s32(__p0_464, __p1_464, __p2_464, __p3_464) __extension__ ({ \
53058  int64x2_t __s0_464 = __p0_464; \
53059  int32x2_t __s1_464 = __p1_464; \
53060  int32x4_t __s2_464 = __p2_464; \
53061  int64x2_t __rev0_464; __rev0_464 = __builtin_shufflevector(__s0_464, __s0_464, 1, 0); \
53062  int32x2_t __rev1_464; __rev1_464 = __builtin_shufflevector(__s1_464, __s1_464, 1, 0); \
53063  int32x4_t __rev2_464; __rev2_464 = __builtin_shufflevector(__s2_464, __s2_464, 3, 2, 1, 0); \
53064  int64x2_t __ret_464; \
53065  __ret_464 = __rev0_464 + __noswap_vmull_s32(__rev1_464, __noswap_splat_laneq_s32(__rev2_464, __p3_464)); \
53066  __ret_464 = __builtin_shufflevector(__ret_464, __ret_464, 1, 0); \
53067  __ret_464; \
53068 })
53069 #endif
53070 
53071 #ifdef __LITTLE_ENDIAN__
53072 #define vmlal_laneq_s16(__p0_465, __p1_465, __p2_465, __p3_465) __extension__ ({ \
53073  int32x4_t __s0_465 = __p0_465; \
53074  int16x4_t __s1_465 = __p1_465; \
53075  int16x8_t __s2_465 = __p2_465; \
53076  int32x4_t __ret_465; \
53077  __ret_465 = __s0_465 + vmull_s16(__s1_465, splat_laneq_s16(__s2_465, __p3_465)); \
53078  __ret_465; \
53079 })
53080 #else
53081 #define vmlal_laneq_s16(__p0_466, __p1_466, __p2_466, __p3_466) __extension__ ({ \
53082  int32x4_t __s0_466 = __p0_466; \
53083  int16x4_t __s1_466 = __p1_466; \
53084  int16x8_t __s2_466 = __p2_466; \
53085  int32x4_t __rev0_466; __rev0_466 = __builtin_shufflevector(__s0_466, __s0_466, 3, 2, 1, 0); \
53086  int16x4_t __rev1_466; __rev1_466 = __builtin_shufflevector(__s1_466, __s1_466, 3, 2, 1, 0); \
53087  int16x8_t __rev2_466; __rev2_466 = __builtin_shufflevector(__s2_466, __s2_466, 7, 6, 5, 4, 3, 2, 1, 0); \
53088  int32x4_t __ret_466; \
53089  __ret_466 = __rev0_466 + __noswap_vmull_s16(__rev1_466, __noswap_splat_laneq_s16(__rev2_466, __p3_466)); \
53090  __ret_466 = __builtin_shufflevector(__ret_466, __ret_466, 3, 2, 1, 0); \
53091  __ret_466; \
53092 })
53093 #endif
53094 
53095 #ifdef __LITTLE_ENDIAN__
53096 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
53097  float64x2_t __ret;
53098  __ret = __p0 - __p1 * __p2;
53099  return __ret;
53100 }
53101 #else
53102 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
53103  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53104  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53105  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
53106  float64x2_t __ret;
53107  __ret = __rev0 - __rev1 * __rev2;
53108  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53109  return __ret;
53110 }
53111 #endif
53112 
53113 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
53114  float64x1_t __ret;
53115  __ret = __p0 - __p1 * __p2;
53116  return __ret;
53117 }
53118 #ifdef __LITTLE_ENDIAN__
53119 #define vmlsq_laneq_u32(__p0_467, __p1_467, __p2_467, __p3_467) __extension__ ({ \
53120  uint32x4_t __s0_467 = __p0_467; \
53121  uint32x4_t __s1_467 = __p1_467; \
53122  uint32x4_t __s2_467 = __p2_467; \
53123  uint32x4_t __ret_467; \
53124  __ret_467 = __s0_467 - __s1_467 * splatq_laneq_u32(__s2_467, __p3_467); \
53125  __ret_467; \
53126 })
53127 #else
53128 #define vmlsq_laneq_u32(__p0_468, __p1_468, __p2_468, __p3_468) __extension__ ({ \
53129  uint32x4_t __s0_468 = __p0_468; \
53130  uint32x4_t __s1_468 = __p1_468; \
53131  uint32x4_t __s2_468 = __p2_468; \
53132  uint32x4_t __rev0_468; __rev0_468 = __builtin_shufflevector(__s0_468, __s0_468, 3, 2, 1, 0); \
53133  uint32x4_t __rev1_468; __rev1_468 = __builtin_shufflevector(__s1_468, __s1_468, 3, 2, 1, 0); \
53134  uint32x4_t __rev2_468; __rev2_468 = __builtin_shufflevector(__s2_468, __s2_468, 3, 2, 1, 0); \
53135  uint32x4_t __ret_468; \
53136  __ret_468 = __rev0_468 - __rev1_468 * __noswap_splatq_laneq_u32(__rev2_468, __p3_468); \
53137  __ret_468 = __builtin_shufflevector(__ret_468, __ret_468, 3, 2, 1, 0); \
53138  __ret_468; \
53139 })
53140 #endif
53141 
53142 #ifdef __LITTLE_ENDIAN__
53143 #define vmlsq_laneq_u16(__p0_469, __p1_469, __p2_469, __p3_469) __extension__ ({ \
53144  uint16x8_t __s0_469 = __p0_469; \
53145  uint16x8_t __s1_469 = __p1_469; \
53146  uint16x8_t __s2_469 = __p2_469; \
53147  uint16x8_t __ret_469; \
53148  __ret_469 = __s0_469 - __s1_469 * splatq_laneq_u16(__s2_469, __p3_469); \
53149  __ret_469; \
53150 })
53151 #else
53152 #define vmlsq_laneq_u16(__p0_470, __p1_470, __p2_470, __p3_470) __extension__ ({ \
53153  uint16x8_t __s0_470 = __p0_470; \
53154  uint16x8_t __s1_470 = __p1_470; \
53155  uint16x8_t __s2_470 = __p2_470; \
53156  uint16x8_t __rev0_470; __rev0_470 = __builtin_shufflevector(__s0_470, __s0_470, 7, 6, 5, 4, 3, 2, 1, 0); \
53157  uint16x8_t __rev1_470; __rev1_470 = __builtin_shufflevector(__s1_470, __s1_470, 7, 6, 5, 4, 3, 2, 1, 0); \
53158  uint16x8_t __rev2_470; __rev2_470 = __builtin_shufflevector(__s2_470, __s2_470, 7, 6, 5, 4, 3, 2, 1, 0); \
53159  uint16x8_t __ret_470; \
53160  __ret_470 = __rev0_470 - __rev1_470 * __noswap_splatq_laneq_u16(__rev2_470, __p3_470); \
53161  __ret_470 = __builtin_shufflevector(__ret_470, __ret_470, 7, 6, 5, 4, 3, 2, 1, 0); \
53162  __ret_470; \
53163 })
53164 #endif
53165 
53166 #ifdef __LITTLE_ENDIAN__
53167 #define vmlsq_laneq_f32(__p0_471, __p1_471, __p2_471, __p3_471) __extension__ ({ \
53168  float32x4_t __s0_471 = __p0_471; \
53169  float32x4_t __s1_471 = __p1_471; \
53170  float32x4_t __s2_471 = __p2_471; \
53171  float32x4_t __ret_471; \
53172  __ret_471 = __s0_471 - __s1_471 * splatq_laneq_f32(__s2_471, __p3_471); \
53173  __ret_471; \
53174 })
53175 #else
53176 #define vmlsq_laneq_f32(__p0_472, __p1_472, __p2_472, __p3_472) __extension__ ({ \
53177  float32x4_t __s0_472 = __p0_472; \
53178  float32x4_t __s1_472 = __p1_472; \
53179  float32x4_t __s2_472 = __p2_472; \
53180  float32x4_t __rev0_472; __rev0_472 = __builtin_shufflevector(__s0_472, __s0_472, 3, 2, 1, 0); \
53181  float32x4_t __rev1_472; __rev1_472 = __builtin_shufflevector(__s1_472, __s1_472, 3, 2, 1, 0); \
53182  float32x4_t __rev2_472; __rev2_472 = __builtin_shufflevector(__s2_472, __s2_472, 3, 2, 1, 0); \
53183  float32x4_t __ret_472; \
53184  __ret_472 = __rev0_472 - __rev1_472 * __noswap_splatq_laneq_f32(__rev2_472, __p3_472); \
53185  __ret_472 = __builtin_shufflevector(__ret_472, __ret_472, 3, 2, 1, 0); \
53186  __ret_472; \
53187 })
53188 #endif
53189 
53190 #ifdef __LITTLE_ENDIAN__
53191 #define vmlsq_laneq_s32(__p0_473, __p1_473, __p2_473, __p3_473) __extension__ ({ \
53192  int32x4_t __s0_473 = __p0_473; \
53193  int32x4_t __s1_473 = __p1_473; \
53194  int32x4_t __s2_473 = __p2_473; \
53195  int32x4_t __ret_473; \
53196  __ret_473 = __s0_473 - __s1_473 * splatq_laneq_s32(__s2_473, __p3_473); \
53197  __ret_473; \
53198 })
53199 #else
53200 #define vmlsq_laneq_s32(__p0_474, __p1_474, __p2_474, __p3_474) __extension__ ({ \
53201  int32x4_t __s0_474 = __p0_474; \
53202  int32x4_t __s1_474 = __p1_474; \
53203  int32x4_t __s2_474 = __p2_474; \
53204  int32x4_t __rev0_474; __rev0_474 = __builtin_shufflevector(__s0_474, __s0_474, 3, 2, 1, 0); \
53205  int32x4_t __rev1_474; __rev1_474 = __builtin_shufflevector(__s1_474, __s1_474, 3, 2, 1, 0); \
53206  int32x4_t __rev2_474; __rev2_474 = __builtin_shufflevector(__s2_474, __s2_474, 3, 2, 1, 0); \
53207  int32x4_t __ret_474; \
53208  __ret_474 = __rev0_474 - __rev1_474 * __noswap_splatq_laneq_s32(__rev2_474, __p3_474); \
53209  __ret_474 = __builtin_shufflevector(__ret_474, __ret_474, 3, 2, 1, 0); \
53210  __ret_474; \
53211 })
53212 #endif
53213 
53214 #ifdef __LITTLE_ENDIAN__
53215 #define vmlsq_laneq_s16(__p0_475, __p1_475, __p2_475, __p3_475) __extension__ ({ \
53216  int16x8_t __s0_475 = __p0_475; \
53217  int16x8_t __s1_475 = __p1_475; \
53218  int16x8_t __s2_475 = __p2_475; \
53219  int16x8_t __ret_475; \
53220  __ret_475 = __s0_475 - __s1_475 * splatq_laneq_s16(__s2_475, __p3_475); \
53221  __ret_475; \
53222 })
53223 #else
53224 #define vmlsq_laneq_s16(__p0_476, __p1_476, __p2_476, __p3_476) __extension__ ({ \
53225  int16x8_t __s0_476 = __p0_476; \
53226  int16x8_t __s1_476 = __p1_476; \
53227  int16x8_t __s2_476 = __p2_476; \
53228  int16x8_t __rev0_476; __rev0_476 = __builtin_shufflevector(__s0_476, __s0_476, 7, 6, 5, 4, 3, 2, 1, 0); \
53229  int16x8_t __rev1_476; __rev1_476 = __builtin_shufflevector(__s1_476, __s1_476, 7, 6, 5, 4, 3, 2, 1, 0); \
53230  int16x8_t __rev2_476; __rev2_476 = __builtin_shufflevector(__s2_476, __s2_476, 7, 6, 5, 4, 3, 2, 1, 0); \
53231  int16x8_t __ret_476; \
53232  __ret_476 = __rev0_476 - __rev1_476 * __noswap_splatq_laneq_s16(__rev2_476, __p3_476); \
53233  __ret_476 = __builtin_shufflevector(__ret_476, __ret_476, 7, 6, 5, 4, 3, 2, 1, 0); \
53234  __ret_476; \
53235 })
53236 #endif
53237 
53238 #ifdef __LITTLE_ENDIAN__
53239 #define vmls_laneq_u32(__p0_477, __p1_477, __p2_477, __p3_477) __extension__ ({ \
53240  uint32x2_t __s0_477 = __p0_477; \
53241  uint32x2_t __s1_477 = __p1_477; \
53242  uint32x4_t __s2_477 = __p2_477; \
53243  uint32x2_t __ret_477; \
53244  __ret_477 = __s0_477 - __s1_477 * splat_laneq_u32(__s2_477, __p3_477); \
53245  __ret_477; \
53246 })
53247 #else
53248 #define vmls_laneq_u32(__p0_478, __p1_478, __p2_478, __p3_478) __extension__ ({ \
53249  uint32x2_t __s0_478 = __p0_478; \
53250  uint32x2_t __s1_478 = __p1_478; \
53251  uint32x4_t __s2_478 = __p2_478; \
53252  uint32x2_t __rev0_478; __rev0_478 = __builtin_shufflevector(__s0_478, __s0_478, 1, 0); \
53253  uint32x2_t __rev1_478; __rev1_478 = __builtin_shufflevector(__s1_478, __s1_478, 1, 0); \
53254  uint32x4_t __rev2_478; __rev2_478 = __builtin_shufflevector(__s2_478, __s2_478, 3, 2, 1, 0); \
53255  uint32x2_t __ret_478; \
53256  __ret_478 = __rev0_478 - __rev1_478 * __noswap_splat_laneq_u32(__rev2_478, __p3_478); \
53257  __ret_478 = __builtin_shufflevector(__ret_478, __ret_478, 1, 0); \
53258  __ret_478; \
53259 })
53260 #endif
53261 
53262 #ifdef __LITTLE_ENDIAN__
53263 #define vmls_laneq_u16(__p0_479, __p1_479, __p2_479, __p3_479) __extension__ ({ \
53264  uint16x4_t __s0_479 = __p0_479; \
53265  uint16x4_t __s1_479 = __p1_479; \
53266  uint16x8_t __s2_479 = __p2_479; \
53267  uint16x4_t __ret_479; \
53268  __ret_479 = __s0_479 - __s1_479 * splat_laneq_u16(__s2_479, __p3_479); \
53269  __ret_479; \
53270 })
53271 #else
53272 #define vmls_laneq_u16(__p0_480, __p1_480, __p2_480, __p3_480) __extension__ ({ \
53273  uint16x4_t __s0_480 = __p0_480; \
53274  uint16x4_t __s1_480 = __p1_480; \
53275  uint16x8_t __s2_480 = __p2_480; \
53276  uint16x4_t __rev0_480; __rev0_480 = __builtin_shufflevector(__s0_480, __s0_480, 3, 2, 1, 0); \
53277  uint16x4_t __rev1_480; __rev1_480 = __builtin_shufflevector(__s1_480, __s1_480, 3, 2, 1, 0); \
53278  uint16x8_t __rev2_480; __rev2_480 = __builtin_shufflevector(__s2_480, __s2_480, 7, 6, 5, 4, 3, 2, 1, 0); \
53279  uint16x4_t __ret_480; \
53280  __ret_480 = __rev0_480 - __rev1_480 * __noswap_splat_laneq_u16(__rev2_480, __p3_480); \
53281  __ret_480 = __builtin_shufflevector(__ret_480, __ret_480, 3, 2, 1, 0); \
53282  __ret_480; \
53283 })
53284 #endif
53285 
53286 #ifdef __LITTLE_ENDIAN__
53287 #define vmls_laneq_f32(__p0_481, __p1_481, __p2_481, __p3_481) __extension__ ({ \
53288  float32x2_t __s0_481 = __p0_481; \
53289  float32x2_t __s1_481 = __p1_481; \
53290  float32x4_t __s2_481 = __p2_481; \
53291  float32x2_t __ret_481; \
53292  __ret_481 = __s0_481 - __s1_481 * splat_laneq_f32(__s2_481, __p3_481); \
53293  __ret_481; \
53294 })
53295 #else
53296 #define vmls_laneq_f32(__p0_482, __p1_482, __p2_482, __p3_482) __extension__ ({ \
53297  float32x2_t __s0_482 = __p0_482; \
53298  float32x2_t __s1_482 = __p1_482; \
53299  float32x4_t __s2_482 = __p2_482; \
53300  float32x2_t __rev0_482; __rev0_482 = __builtin_shufflevector(__s0_482, __s0_482, 1, 0); \
53301  float32x2_t __rev1_482; __rev1_482 = __builtin_shufflevector(__s1_482, __s1_482, 1, 0); \
53302  float32x4_t __rev2_482; __rev2_482 = __builtin_shufflevector(__s2_482, __s2_482, 3, 2, 1, 0); \
53303  float32x2_t __ret_482; \
53304  __ret_482 = __rev0_482 - __rev1_482 * __noswap_splat_laneq_f32(__rev2_482, __p3_482); \
53305  __ret_482 = __builtin_shufflevector(__ret_482, __ret_482, 1, 0); \
53306  __ret_482; \
53307 })
53308 #endif
53309 
53310 #ifdef __LITTLE_ENDIAN__
53311 #define vmls_laneq_s32(__p0_483, __p1_483, __p2_483, __p3_483) __extension__ ({ \
53312  int32x2_t __s0_483 = __p0_483; \
53313  int32x2_t __s1_483 = __p1_483; \
53314  int32x4_t __s2_483 = __p2_483; \
53315  int32x2_t __ret_483; \
53316  __ret_483 = __s0_483 - __s1_483 * splat_laneq_s32(__s2_483, __p3_483); \
53317  __ret_483; \
53318 })
53319 #else
53320 #define vmls_laneq_s32(__p0_484, __p1_484, __p2_484, __p3_484) __extension__ ({ \
53321  int32x2_t __s0_484 = __p0_484; \
53322  int32x2_t __s1_484 = __p1_484; \
53323  int32x4_t __s2_484 = __p2_484; \
53324  int32x2_t __rev0_484; __rev0_484 = __builtin_shufflevector(__s0_484, __s0_484, 1, 0); \
53325  int32x2_t __rev1_484; __rev1_484 = __builtin_shufflevector(__s1_484, __s1_484, 1, 0); \
53326  int32x4_t __rev2_484; __rev2_484 = __builtin_shufflevector(__s2_484, __s2_484, 3, 2, 1, 0); \
53327  int32x2_t __ret_484; \
53328  __ret_484 = __rev0_484 - __rev1_484 * __noswap_splat_laneq_s32(__rev2_484, __p3_484); \
53329  __ret_484 = __builtin_shufflevector(__ret_484, __ret_484, 1, 0); \
53330  __ret_484; \
53331 })
53332 #endif
53333 
53334 #ifdef __LITTLE_ENDIAN__
53335 #define vmls_laneq_s16(__p0_485, __p1_485, __p2_485, __p3_485) __extension__ ({ \
53336  int16x4_t __s0_485 = __p0_485; \
53337  int16x4_t __s1_485 = __p1_485; \
53338  int16x8_t __s2_485 = __p2_485; \
53339  int16x4_t __ret_485; \
53340  __ret_485 = __s0_485 - __s1_485 * splat_laneq_s16(__s2_485, __p3_485); \
53341  __ret_485; \
53342 })
53343 #else
53344 #define vmls_laneq_s16(__p0_486, __p1_486, __p2_486, __p3_486) __extension__ ({ \
53345  int16x4_t __s0_486 = __p0_486; \
53346  int16x4_t __s1_486 = __p1_486; \
53347  int16x8_t __s2_486 = __p2_486; \
53348  int16x4_t __rev0_486; __rev0_486 = __builtin_shufflevector(__s0_486, __s0_486, 3, 2, 1, 0); \
53349  int16x4_t __rev1_486; __rev1_486 = __builtin_shufflevector(__s1_486, __s1_486, 3, 2, 1, 0); \
53350  int16x8_t __rev2_486; __rev2_486 = __builtin_shufflevector(__s2_486, __s2_486, 7, 6, 5, 4, 3, 2, 1, 0); \
53351  int16x4_t __ret_486; \
53352  __ret_486 = __rev0_486 - __rev1_486 * __noswap_splat_laneq_s16(__rev2_486, __p3_486); \
53353  __ret_486 = __builtin_shufflevector(__ret_486, __ret_486, 3, 2, 1, 0); \
53354  __ret_486; \
53355 })
53356 #endif
53357 
53358 #ifdef __LITTLE_ENDIAN__
53359 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53360  float64x2_t __ret;
53361  __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
53362  return __ret;
53363 }
53364 #else
53365 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53366  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53367  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53368  float64x2_t __ret;
53369  __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
53370  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53371  return __ret;
53372 }
53373 #endif
53374 
53375 #ifdef __LITTLE_ENDIAN__
53376 #define vmlsl_high_lane_u32(__p0_487, __p1_487, __p2_487, __p3_487) __extension__ ({ \
53377  uint64x2_t __s0_487 = __p0_487; \
53378  uint32x4_t __s1_487 = __p1_487; \
53379  uint32x2_t __s2_487 = __p2_487; \
53380  uint64x2_t __ret_487; \
53381  __ret_487 = __s0_487 - vmull_u32(vget_high_u32(__s1_487), splat_lane_u32(__s2_487, __p3_487)); \
53382  __ret_487; \
53383 })
53384 #else
53385 #define vmlsl_high_lane_u32(__p0_488, __p1_488, __p2_488, __p3_488) __extension__ ({ \
53386  uint64x2_t __s0_488 = __p0_488; \
53387  uint32x4_t __s1_488 = __p1_488; \
53388  uint32x2_t __s2_488 = __p2_488; \
53389  uint64x2_t __rev0_488; __rev0_488 = __builtin_shufflevector(__s0_488, __s0_488, 1, 0); \
53390  uint32x4_t __rev1_488; __rev1_488 = __builtin_shufflevector(__s1_488, __s1_488, 3, 2, 1, 0); \
53391  uint32x2_t __rev2_488; __rev2_488 = __builtin_shufflevector(__s2_488, __s2_488, 1, 0); \
53392  uint64x2_t __ret_488; \
53393  __ret_488 = __rev0_488 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_488), __noswap_splat_lane_u32(__rev2_488, __p3_488)); \
53394  __ret_488 = __builtin_shufflevector(__ret_488, __ret_488, 1, 0); \
53395  __ret_488; \
53396 })
53397 #endif
53398 
53399 #ifdef __LITTLE_ENDIAN__
53400 #define vmlsl_high_lane_u16(__p0_489, __p1_489, __p2_489, __p3_489) __extension__ ({ \
53401  uint32x4_t __s0_489 = __p0_489; \
53402  uint16x8_t __s1_489 = __p1_489; \
53403  uint16x4_t __s2_489 = __p2_489; \
53404  uint32x4_t __ret_489; \
53405  __ret_489 = __s0_489 - vmull_u16(vget_high_u16(__s1_489), splat_lane_u16(__s2_489, __p3_489)); \
53406  __ret_489; \
53407 })
53408 #else
53409 #define vmlsl_high_lane_u16(__p0_490, __p1_490, __p2_490, __p3_490) __extension__ ({ \
53410  uint32x4_t __s0_490 = __p0_490; \
53411  uint16x8_t __s1_490 = __p1_490; \
53412  uint16x4_t __s2_490 = __p2_490; \
53413  uint32x4_t __rev0_490; __rev0_490 = __builtin_shufflevector(__s0_490, __s0_490, 3, 2, 1, 0); \
53414  uint16x8_t __rev1_490; __rev1_490 = __builtin_shufflevector(__s1_490, __s1_490, 7, 6, 5, 4, 3, 2, 1, 0); \
53415  uint16x4_t __rev2_490; __rev2_490 = __builtin_shufflevector(__s2_490, __s2_490, 3, 2, 1, 0); \
53416  uint32x4_t __ret_490; \
53417  __ret_490 = __rev0_490 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_490), __noswap_splat_lane_u16(__rev2_490, __p3_490)); \
53418  __ret_490 = __builtin_shufflevector(__ret_490, __ret_490, 3, 2, 1, 0); \
53419  __ret_490; \
53420 })
53421 #endif
53422 
53423 #ifdef __LITTLE_ENDIAN__
53424 #define vmlsl_high_lane_s32(__p0_491, __p1_491, __p2_491, __p3_491) __extension__ ({ \
53425  int64x2_t __s0_491 = __p0_491; \
53426  int32x4_t __s1_491 = __p1_491; \
53427  int32x2_t __s2_491 = __p2_491; \
53428  int64x2_t __ret_491; \
53429  __ret_491 = __s0_491 - vmull_s32(vget_high_s32(__s1_491), splat_lane_s32(__s2_491, __p3_491)); \
53430  __ret_491; \
53431 })
53432 #else
53433 #define vmlsl_high_lane_s32(__p0_492, __p1_492, __p2_492, __p3_492) __extension__ ({ \
53434  int64x2_t __s0_492 = __p0_492; \
53435  int32x4_t __s1_492 = __p1_492; \
53436  int32x2_t __s2_492 = __p2_492; \
53437  int64x2_t __rev0_492; __rev0_492 = __builtin_shufflevector(__s0_492, __s0_492, 1, 0); \
53438  int32x4_t __rev1_492; __rev1_492 = __builtin_shufflevector(__s1_492, __s1_492, 3, 2, 1, 0); \
53439  int32x2_t __rev2_492; __rev2_492 = __builtin_shufflevector(__s2_492, __s2_492, 1, 0); \
53440  int64x2_t __ret_492; \
53441  __ret_492 = __rev0_492 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_492), __noswap_splat_lane_s32(__rev2_492, __p3_492)); \
53442  __ret_492 = __builtin_shufflevector(__ret_492, __ret_492, 1, 0); \
53443  __ret_492; \
53444 })
53445 #endif
53446 
53447 #ifdef __LITTLE_ENDIAN__
53448 #define vmlsl_high_lane_s16(__p0_493, __p1_493, __p2_493, __p3_493) __extension__ ({ \
53449  int32x4_t __s0_493 = __p0_493; \
53450  int16x8_t __s1_493 = __p1_493; \
53451  int16x4_t __s2_493 = __p2_493; \
53452  int32x4_t __ret_493; \
53453  __ret_493 = __s0_493 - vmull_s16(vget_high_s16(__s1_493), splat_lane_s16(__s2_493, __p3_493)); \
53454  __ret_493; \
53455 })
53456 #else
53457 #define vmlsl_high_lane_s16(__p0_494, __p1_494, __p2_494, __p3_494) __extension__ ({ \
53458  int32x4_t __s0_494 = __p0_494; \
53459  int16x8_t __s1_494 = __p1_494; \
53460  int16x4_t __s2_494 = __p2_494; \
53461  int32x4_t __rev0_494; __rev0_494 = __builtin_shufflevector(__s0_494, __s0_494, 3, 2, 1, 0); \
53462  int16x8_t __rev1_494; __rev1_494 = __builtin_shufflevector(__s1_494, __s1_494, 7, 6, 5, 4, 3, 2, 1, 0); \
53463  int16x4_t __rev2_494; __rev2_494 = __builtin_shufflevector(__s2_494, __s2_494, 3, 2, 1, 0); \
53464  int32x4_t __ret_494; \
53465  __ret_494 = __rev0_494 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_494), __noswap_splat_lane_s16(__rev2_494, __p3_494)); \
53466  __ret_494 = __builtin_shufflevector(__ret_494, __ret_494, 3, 2, 1, 0); \
53467  __ret_494; \
53468 })
53469 #endif
53470 
53471 #ifdef __LITTLE_ENDIAN__
53472 #define vmlsl_high_laneq_u32(__p0_495, __p1_495, __p2_495, __p3_495) __extension__ ({ \
53473  uint64x2_t __s0_495 = __p0_495; \
53474  uint32x4_t __s1_495 = __p1_495; \
53475  uint32x4_t __s2_495 = __p2_495; \
53476  uint64x2_t __ret_495; \
53477  __ret_495 = __s0_495 - vmull_u32(vget_high_u32(__s1_495), splat_laneq_u32(__s2_495, __p3_495)); \
53478  __ret_495; \
53479 })
53480 #else
53481 #define vmlsl_high_laneq_u32(__p0_496, __p1_496, __p2_496, __p3_496) __extension__ ({ \
53482  uint64x2_t __s0_496 = __p0_496; \
53483  uint32x4_t __s1_496 = __p1_496; \
53484  uint32x4_t __s2_496 = __p2_496; \
53485  uint64x2_t __rev0_496; __rev0_496 = __builtin_shufflevector(__s0_496, __s0_496, 1, 0); \
53486  uint32x4_t __rev1_496; __rev1_496 = __builtin_shufflevector(__s1_496, __s1_496, 3, 2, 1, 0); \
53487  uint32x4_t __rev2_496; __rev2_496 = __builtin_shufflevector(__s2_496, __s2_496, 3, 2, 1, 0); \
53488  uint64x2_t __ret_496; \
53489  __ret_496 = __rev0_496 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1_496), __noswap_splat_laneq_u32(__rev2_496, __p3_496)); \
53490  __ret_496 = __builtin_shufflevector(__ret_496, __ret_496, 1, 0); \
53491  __ret_496; \
53492 })
53493 #endif
53494 
53495 #ifdef __LITTLE_ENDIAN__
53496 #define vmlsl_high_laneq_u16(__p0_497, __p1_497, __p2_497, __p3_497) __extension__ ({ \
53497  uint32x4_t __s0_497 = __p0_497; \
53498  uint16x8_t __s1_497 = __p1_497; \
53499  uint16x8_t __s2_497 = __p2_497; \
53500  uint32x4_t __ret_497; \
53501  __ret_497 = __s0_497 - vmull_u16(vget_high_u16(__s1_497), splat_laneq_u16(__s2_497, __p3_497)); \
53502  __ret_497; \
53503 })
53504 #else
53505 #define vmlsl_high_laneq_u16(__p0_498, __p1_498, __p2_498, __p3_498) __extension__ ({ \
53506  uint32x4_t __s0_498 = __p0_498; \
53507  uint16x8_t __s1_498 = __p1_498; \
53508  uint16x8_t __s2_498 = __p2_498; \
53509  uint32x4_t __rev0_498; __rev0_498 = __builtin_shufflevector(__s0_498, __s0_498, 3, 2, 1, 0); \
53510  uint16x8_t __rev1_498; __rev1_498 = __builtin_shufflevector(__s1_498, __s1_498, 7, 6, 5, 4, 3, 2, 1, 0); \
53511  uint16x8_t __rev2_498; __rev2_498 = __builtin_shufflevector(__s2_498, __s2_498, 7, 6, 5, 4, 3, 2, 1, 0); \
53512  uint32x4_t __ret_498; \
53513  __ret_498 = __rev0_498 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1_498), __noswap_splat_laneq_u16(__rev2_498, __p3_498)); \
53514  __ret_498 = __builtin_shufflevector(__ret_498, __ret_498, 3, 2, 1, 0); \
53515  __ret_498; \
53516 })
53517 #endif
53518 
53519 #ifdef __LITTLE_ENDIAN__
53520 #define vmlsl_high_laneq_s32(__p0_499, __p1_499, __p2_499, __p3_499) __extension__ ({ \
53521  int64x2_t __s0_499 = __p0_499; \
53522  int32x4_t __s1_499 = __p1_499; \
53523  int32x4_t __s2_499 = __p2_499; \
53524  int64x2_t __ret_499; \
53525  __ret_499 = __s0_499 - vmull_s32(vget_high_s32(__s1_499), splat_laneq_s32(__s2_499, __p3_499)); \
53526  __ret_499; \
53527 })
53528 #else
53529 #define vmlsl_high_laneq_s32(__p0_500, __p1_500, __p2_500, __p3_500) __extension__ ({ \
53530  int64x2_t __s0_500 = __p0_500; \
53531  int32x4_t __s1_500 = __p1_500; \
53532  int32x4_t __s2_500 = __p2_500; \
53533  int64x2_t __rev0_500; __rev0_500 = __builtin_shufflevector(__s0_500, __s0_500, 1, 0); \
53534  int32x4_t __rev1_500; __rev1_500 = __builtin_shufflevector(__s1_500, __s1_500, 3, 2, 1, 0); \
53535  int32x4_t __rev2_500; __rev2_500 = __builtin_shufflevector(__s2_500, __s2_500, 3, 2, 1, 0); \
53536  int64x2_t __ret_500; \
53537  __ret_500 = __rev0_500 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1_500), __noswap_splat_laneq_s32(__rev2_500, __p3_500)); \
53538  __ret_500 = __builtin_shufflevector(__ret_500, __ret_500, 1, 0); \
53539  __ret_500; \
53540 })
53541 #endif
53542 
53543 #ifdef __LITTLE_ENDIAN__
53544 #define vmlsl_high_laneq_s16(__p0_501, __p1_501, __p2_501, __p3_501) __extension__ ({ \
53545  int32x4_t __s0_501 = __p0_501; \
53546  int16x8_t __s1_501 = __p1_501; \
53547  int16x8_t __s2_501 = __p2_501; \
53548  int32x4_t __ret_501; \
53549  __ret_501 = __s0_501 - vmull_s16(vget_high_s16(__s1_501), splat_laneq_s16(__s2_501, __p3_501)); \
53550  __ret_501; \
53551 })
53552 #else
53553 #define vmlsl_high_laneq_s16(__p0_502, __p1_502, __p2_502, __p3_502) __extension__ ({ \
53554  int32x4_t __s0_502 = __p0_502; \
53555  int16x8_t __s1_502 = __p1_502; \
53556  int16x8_t __s2_502 = __p2_502; \
53557  int32x4_t __rev0_502; __rev0_502 = __builtin_shufflevector(__s0_502, __s0_502, 3, 2, 1, 0); \
53558  int16x8_t __rev1_502; __rev1_502 = __builtin_shufflevector(__s1_502, __s1_502, 7, 6, 5, 4, 3, 2, 1, 0); \
53559  int16x8_t __rev2_502; __rev2_502 = __builtin_shufflevector(__s2_502, __s2_502, 7, 6, 5, 4, 3, 2, 1, 0); \
53560  int32x4_t __ret_502; \
53561  __ret_502 = __rev0_502 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1_502), __noswap_splat_laneq_s16(__rev2_502, __p3_502)); \
53562  __ret_502 = __builtin_shufflevector(__ret_502, __ret_502, 3, 2, 1, 0); \
53563  __ret_502; \
53564 })
53565 #endif
53566 
53567 #ifdef __LITTLE_ENDIAN__
53568 #define vmlsl_laneq_u32(__p0_503, __p1_503, __p2_503, __p3_503) __extension__ ({ \
53569  uint64x2_t __s0_503 = __p0_503; \
53570  uint32x2_t __s1_503 = __p1_503; \
53571  uint32x4_t __s2_503 = __p2_503; \
53572  uint64x2_t __ret_503; \
53573  __ret_503 = __s0_503 - vmull_u32(__s1_503, splat_laneq_u32(__s2_503, __p3_503)); \
53574  __ret_503; \
53575 })
53576 #else
53577 #define vmlsl_laneq_u32(__p0_504, __p1_504, __p2_504, __p3_504) __extension__ ({ \
53578  uint64x2_t __s0_504 = __p0_504; \
53579  uint32x2_t __s1_504 = __p1_504; \
53580  uint32x4_t __s2_504 = __p2_504; \
53581  uint64x2_t __rev0_504; __rev0_504 = __builtin_shufflevector(__s0_504, __s0_504, 1, 0); \
53582  uint32x2_t __rev1_504; __rev1_504 = __builtin_shufflevector(__s1_504, __s1_504, 1, 0); \
53583  uint32x4_t __rev2_504; __rev2_504 = __builtin_shufflevector(__s2_504, __s2_504, 3, 2, 1, 0); \
53584  uint64x2_t __ret_504; \
53585  __ret_504 = __rev0_504 - __noswap_vmull_u32(__rev1_504, __noswap_splat_laneq_u32(__rev2_504, __p3_504)); \
53586  __ret_504 = __builtin_shufflevector(__ret_504, __ret_504, 1, 0); \
53587  __ret_504; \
53588 })
53589 #endif
53590 
53591 #ifdef __LITTLE_ENDIAN__
53592 #define vmlsl_laneq_u16(__p0_505, __p1_505, __p2_505, __p3_505) __extension__ ({ \
53593  uint32x4_t __s0_505 = __p0_505; \
53594  uint16x4_t __s1_505 = __p1_505; \
53595  uint16x8_t __s2_505 = __p2_505; \
53596  uint32x4_t __ret_505; \
53597  __ret_505 = __s0_505 - vmull_u16(__s1_505, splat_laneq_u16(__s2_505, __p3_505)); \
53598  __ret_505; \
53599 })
53600 #else
53601 #define vmlsl_laneq_u16(__p0_506, __p1_506, __p2_506, __p3_506) __extension__ ({ \
53602  uint32x4_t __s0_506 = __p0_506; \
53603  uint16x4_t __s1_506 = __p1_506; \
53604  uint16x8_t __s2_506 = __p2_506; \
53605  uint32x4_t __rev0_506; __rev0_506 = __builtin_shufflevector(__s0_506, __s0_506, 3, 2, 1, 0); \
53606  uint16x4_t __rev1_506; __rev1_506 = __builtin_shufflevector(__s1_506, __s1_506, 3, 2, 1, 0); \
53607  uint16x8_t __rev2_506; __rev2_506 = __builtin_shufflevector(__s2_506, __s2_506, 7, 6, 5, 4, 3, 2, 1, 0); \
53608  uint32x4_t __ret_506; \
53609  __ret_506 = __rev0_506 - __noswap_vmull_u16(__rev1_506, __noswap_splat_laneq_u16(__rev2_506, __p3_506)); \
53610  __ret_506 = __builtin_shufflevector(__ret_506, __ret_506, 3, 2, 1, 0); \
53611  __ret_506; \
53612 })
53613 #endif
53614 
53615 #ifdef __LITTLE_ENDIAN__
53616 #define vmlsl_laneq_s32(__p0_507, __p1_507, __p2_507, __p3_507) __extension__ ({ \
53617  int64x2_t __s0_507 = __p0_507; \
53618  int32x2_t __s1_507 = __p1_507; \
53619  int32x4_t __s2_507 = __p2_507; \
53620  int64x2_t __ret_507; \
53621  __ret_507 = __s0_507 - vmull_s32(__s1_507, splat_laneq_s32(__s2_507, __p3_507)); \
53622  __ret_507; \
53623 })
53624 #else
53625 #define vmlsl_laneq_s32(__p0_508, __p1_508, __p2_508, __p3_508) __extension__ ({ \
53626  int64x2_t __s0_508 = __p0_508; \
53627  int32x2_t __s1_508 = __p1_508; \
53628  int32x4_t __s2_508 = __p2_508; \
53629  int64x2_t __rev0_508; __rev0_508 = __builtin_shufflevector(__s0_508, __s0_508, 1, 0); \
53630  int32x2_t __rev1_508; __rev1_508 = __builtin_shufflevector(__s1_508, __s1_508, 1, 0); \
53631  int32x4_t __rev2_508; __rev2_508 = __builtin_shufflevector(__s2_508, __s2_508, 3, 2, 1, 0); \
53632  int64x2_t __ret_508; \
53633  __ret_508 = __rev0_508 - __noswap_vmull_s32(__rev1_508, __noswap_splat_laneq_s32(__rev2_508, __p3_508)); \
53634  __ret_508 = __builtin_shufflevector(__ret_508, __ret_508, 1, 0); \
53635  __ret_508; \
53636 })
53637 #endif
53638 
53639 #ifdef __LITTLE_ENDIAN__
53640 #define vmlsl_laneq_s16(__p0_509, __p1_509, __p2_509, __p3_509) __extension__ ({ \
53641  int32x4_t __s0_509 = __p0_509; \
53642  int16x4_t __s1_509 = __p1_509; \
53643  int16x8_t __s2_509 = __p2_509; \
53644  int32x4_t __ret_509; \
53645  __ret_509 = __s0_509 - vmull_s16(__s1_509, splat_laneq_s16(__s2_509, __p3_509)); \
53646  __ret_509; \
53647 })
53648 #else
53649 #define vmlsl_laneq_s16(__p0_510, __p1_510, __p2_510, __p3_510) __extension__ ({ \
53650  int32x4_t __s0_510 = __p0_510; \
53651  int16x4_t __s1_510 = __p1_510; \
53652  int16x8_t __s2_510 = __p2_510; \
53653  int32x4_t __rev0_510; __rev0_510 = __builtin_shufflevector(__s0_510, __s0_510, 3, 2, 1, 0); \
53654  int16x4_t __rev1_510; __rev1_510 = __builtin_shufflevector(__s1_510, __s1_510, 3, 2, 1, 0); \
53655  int16x8_t __rev2_510; __rev2_510 = __builtin_shufflevector(__s2_510, __s2_510, 7, 6, 5, 4, 3, 2, 1, 0); \
53656  int32x4_t __ret_510; \
53657  __ret_510 = __rev0_510 - __noswap_vmull_s16(__rev1_510, __noswap_splat_laneq_s16(__rev2_510, __p3_510)); \
53658  __ret_510 = __builtin_shufflevector(__ret_510, __ret_510, 3, 2, 1, 0); \
53659  __ret_510; \
53660 })
53661 #endif
53662 
53663 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53664  poly64x1_t __ret;
53665  __ret = (poly64x1_t) {__p0};
53666  return __ret;
53667 }
53668 #ifdef __LITTLE_ENDIAN__
53669 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53670  poly64x2_t __ret;
53671  __ret = (poly64x2_t) {__p0, __p0};
53672  return __ret;
53673 }
53674 #else
53675 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53676  poly64x2_t __ret;
53677  __ret = (poly64x2_t) {__p0, __p0};
53678  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53679  return __ret;
53680 }
53681 #endif
53682 
53683 #ifdef __LITTLE_ENDIAN__
53684 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
53685  float64x2_t __ret;
53686  __ret = (float64x2_t) {__p0, __p0};
53687  return __ret;
53688 }
53689 #else
53690 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
53691  float64x2_t __ret;
53692  __ret = (float64x2_t) {__p0, __p0};
53693  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53694  return __ret;
53695 }
53696 #endif
53697 
53698 __ai float64x1_t vmov_n_f64(float64_t __p0) {
53699  float64x1_t __ret;
53700  __ret = (float64x1_t) {__p0};
53701  return __ret;
53702 }
53703 #ifdef __LITTLE_ENDIAN__
53704 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_511) {
53705  uint16x8_t __ret_511;
53706  uint8x8_t __a1_511 = vget_high_u8(__p0_511);
53707  __ret_511 = (uint16x8_t)(vshll_n_u8(__a1_511, 0));
53708  return __ret_511;
53709 }
53710 #else
53711 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_512) {
53712  uint8x16_t __rev0_512; __rev0_512 = __builtin_shufflevector(__p0_512, __p0_512, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53713  uint16x8_t __ret_512;
53714  uint8x8_t __a1_512 = __noswap_vget_high_u8(__rev0_512);
53715  __ret_512 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_512, 0));
53716  __ret_512 = __builtin_shufflevector(__ret_512, __ret_512, 7, 6, 5, 4, 3, 2, 1, 0);
53717  return __ret_512;
53718 }
53719 __ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_513) {
53720  uint16x8_t __ret_513;
53721  uint8x8_t __a1_513 = __noswap_vget_high_u8(__p0_513);
53722  __ret_513 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_513, 0));
53723  return __ret_513;
53724 }
53725 #endif
53726 
53727 #ifdef __LITTLE_ENDIAN__
53728 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_514) {
53729  uint64x2_t __ret_514;
53730  uint32x2_t __a1_514 = vget_high_u32(__p0_514);
53731  __ret_514 = (uint64x2_t)(vshll_n_u32(__a1_514, 0));
53732  return __ret_514;
53733 }
53734 #else
53735 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_515) {
53736  uint32x4_t __rev0_515; __rev0_515 = __builtin_shufflevector(__p0_515, __p0_515, 3, 2, 1, 0);
53737  uint64x2_t __ret_515;
53738  uint32x2_t __a1_515 = __noswap_vget_high_u32(__rev0_515);
53739  __ret_515 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_515, 0));
53740  __ret_515 = __builtin_shufflevector(__ret_515, __ret_515, 1, 0);
53741  return __ret_515;
53742 }
53743 __ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_516) {
53744  uint64x2_t __ret_516;
53745  uint32x2_t __a1_516 = __noswap_vget_high_u32(__p0_516);
53746  __ret_516 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_516, 0));
53747  return __ret_516;
53748 }
53749 #endif
53750 
53751 #ifdef __LITTLE_ENDIAN__
53752 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_517) {
53753  uint32x4_t __ret_517;
53754  uint16x4_t __a1_517 = vget_high_u16(__p0_517);
53755  __ret_517 = (uint32x4_t)(vshll_n_u16(__a1_517, 0));
53756  return __ret_517;
53757 }
53758 #else
53759 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_518) {
53760  uint16x8_t __rev0_518; __rev0_518 = __builtin_shufflevector(__p0_518, __p0_518, 7, 6, 5, 4, 3, 2, 1, 0);
53761  uint32x4_t __ret_518;
53762  uint16x4_t __a1_518 = __noswap_vget_high_u16(__rev0_518);
53763  __ret_518 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_518, 0));
53764  __ret_518 = __builtin_shufflevector(__ret_518, __ret_518, 3, 2, 1, 0);
53765  return __ret_518;
53766 }
53767 __ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_519) {
53768  uint32x4_t __ret_519;
53769  uint16x4_t __a1_519 = __noswap_vget_high_u16(__p0_519);
53770  __ret_519 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_519, 0));
53771  return __ret_519;
53772 }
53773 #endif
53774 
53775 #ifdef __LITTLE_ENDIAN__
53776 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_520) {
53777  int16x8_t __ret_520;
53778  int8x8_t __a1_520 = vget_high_s8(__p0_520);
53779  __ret_520 = (int16x8_t)(vshll_n_s8(__a1_520, 0));
53780  return __ret_520;
53781 }
53782 #else
53783 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_521) {
53784  int8x16_t __rev0_521; __rev0_521 = __builtin_shufflevector(__p0_521, __p0_521, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53785  int16x8_t __ret_521;
53786  int8x8_t __a1_521 = __noswap_vget_high_s8(__rev0_521);
53787  __ret_521 = (int16x8_t)(__noswap_vshll_n_s8(__a1_521, 0));
53788  __ret_521 = __builtin_shufflevector(__ret_521, __ret_521, 7, 6, 5, 4, 3, 2, 1, 0);
53789  return __ret_521;
53790 }
53791 __ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_522) {
53792  int16x8_t __ret_522;
53793  int8x8_t __a1_522 = __noswap_vget_high_s8(__p0_522);
53794  __ret_522 = (int16x8_t)(__noswap_vshll_n_s8(__a1_522, 0));
53795  return __ret_522;
53796 }
53797 #endif
53798 
53799 #ifdef __LITTLE_ENDIAN__
53800 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_523) {
53801  int64x2_t __ret_523;
53802  int32x2_t __a1_523 = vget_high_s32(__p0_523);
53803  __ret_523 = (int64x2_t)(vshll_n_s32(__a1_523, 0));
53804  return __ret_523;
53805 }
53806 #else
53807 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_524) {
53808  int32x4_t __rev0_524; __rev0_524 = __builtin_shufflevector(__p0_524, __p0_524, 3, 2, 1, 0);
53809  int64x2_t __ret_524;
53810  int32x2_t __a1_524 = __noswap_vget_high_s32(__rev0_524);
53811  __ret_524 = (int64x2_t)(__noswap_vshll_n_s32(__a1_524, 0));
53812  __ret_524 = __builtin_shufflevector(__ret_524, __ret_524, 1, 0);
53813  return __ret_524;
53814 }
53815 __ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_525) {
53816  int64x2_t __ret_525;
53817  int32x2_t __a1_525 = __noswap_vget_high_s32(__p0_525);
53818  __ret_525 = (int64x2_t)(__noswap_vshll_n_s32(__a1_525, 0));
53819  return __ret_525;
53820 }
53821 #endif
53822 
53823 #ifdef __LITTLE_ENDIAN__
53824 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_526) {
53825  int32x4_t __ret_526;
53826  int16x4_t __a1_526 = vget_high_s16(__p0_526);
53827  __ret_526 = (int32x4_t)(vshll_n_s16(__a1_526, 0));
53828  return __ret_526;
53829 }
53830 #else
53831 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_527) {
53832  int16x8_t __rev0_527; __rev0_527 = __builtin_shufflevector(__p0_527, __p0_527, 7, 6, 5, 4, 3, 2, 1, 0);
53833  int32x4_t __ret_527;
53834  int16x4_t __a1_527 = __noswap_vget_high_s16(__rev0_527);
53835  __ret_527 = (int32x4_t)(__noswap_vshll_n_s16(__a1_527, 0));
53836  __ret_527 = __builtin_shufflevector(__ret_527, __ret_527, 3, 2, 1, 0);
53837  return __ret_527;
53838 }
53839 __ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_528) {
53840  int32x4_t __ret_528;
53841  int16x4_t __a1_528 = __noswap_vget_high_s16(__p0_528);
53842  __ret_528 = (int32x4_t)(__noswap_vshll_n_s16(__a1_528, 0));
53843  return __ret_528;
53844 }
53845 #endif
53846 
53847 #ifdef __LITTLE_ENDIAN__
53848 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53849  uint16x8_t __ret;
53850  __ret = vcombine_u16(__p0, vmovn_u32(__p1));
53851  return __ret;
53852 }
53853 #else
53854 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53855  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53856  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53857  uint16x8_t __ret;
53858  __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
53859  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53860  return __ret;
53861 }
53862 #endif
53863 
53864 #ifdef __LITTLE_ENDIAN__
53865 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53866  uint32x4_t __ret;
53867  __ret = vcombine_u32(__p0, vmovn_u64(__p1));
53868  return __ret;
53869 }
53870 #else
53871 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53872  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53873  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53874  uint32x4_t __ret;
53875  __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
53876  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53877  return __ret;
53878 }
53879 #endif
53880 
53881 #ifdef __LITTLE_ENDIAN__
53882 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53883  uint8x16_t __ret;
53884  __ret = vcombine_u8(__p0, vmovn_u16(__p1));
53885  return __ret;
53886 }
53887 #else
53888 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53889  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53890  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53891  uint8x16_t __ret;
53892  __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
53893  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53894  return __ret;
53895 }
53896 #endif
53897 
53898 #ifdef __LITTLE_ENDIAN__
53899 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53900  int16x8_t __ret;
53901  __ret = vcombine_s16(__p0, vmovn_s32(__p1));
53902  return __ret;
53903 }
53904 #else
53905 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53906  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53907  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53908  int16x8_t __ret;
53909  __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
53910  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53911  return __ret;
53912 }
53913 #endif
53914 
53915 #ifdef __LITTLE_ENDIAN__
53916 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53917  int32x4_t __ret;
53918  __ret = vcombine_s32(__p0, vmovn_s64(__p1));
53919  return __ret;
53920 }
53921 #else
53922 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53923  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53924  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53925  int32x4_t __ret;
53926  __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
53927  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53928  return __ret;
53929 }
53930 #endif
53931 
53932 #ifdef __LITTLE_ENDIAN__
53933 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53934  int8x16_t __ret;
53935  __ret = vcombine_s8(__p0, vmovn_s16(__p1));
53936  return __ret;
53937 }
53938 #else
53939 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53940  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53941  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53942  int8x16_t __ret;
53943  __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
53944  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53945  return __ret;
53946 }
53947 #endif
53948 
53949 #ifdef __LITTLE_ENDIAN__
53950 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53951  float64x2_t __ret;
53952  __ret = __p0 * __p1;
53953  return __ret;
53954 }
53955 #else
53956 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53957  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53958  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53959  float64x2_t __ret;
53960  __ret = __rev0 * __rev1;
53961  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53962  return __ret;
53963 }
53964 #endif
53965 
53966 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53967  float64x1_t __ret;
53968  __ret = __p0 * __p1;
53969  return __ret;
53970 }
53971 #define vmuld_lane_f64(__p0_529, __p1_529, __p2_529) __extension__ ({ \
53972  float64_t __s0_529 = __p0_529; \
53973  float64x1_t __s1_529 = __p1_529; \
53974  float64_t __ret_529; \
53975  __ret_529 = __s0_529 * vget_lane_f64(__s1_529, __p2_529); \
53976  __ret_529; \
53977 })
53978 #ifdef __LITTLE_ENDIAN__
53979 #define vmuls_lane_f32(__p0_530, __p1_530, __p2_530) __extension__ ({ \
53980  float32_t __s0_530 = __p0_530; \
53981  float32x2_t __s1_530 = __p1_530; \
53982  float32_t __ret_530; \
53983  __ret_530 = __s0_530 * vget_lane_f32(__s1_530, __p2_530); \
53984  __ret_530; \
53985 })
53986 #else
53987 #define vmuls_lane_f32(__p0_531, __p1_531, __p2_531) __extension__ ({ \
53988  float32_t __s0_531 = __p0_531; \
53989  float32x2_t __s1_531 = __p1_531; \
53990  float32x2_t __rev1_531; __rev1_531 = __builtin_shufflevector(__s1_531, __s1_531, 1, 0); \
53991  float32_t __ret_531; \
53992  __ret_531 = __s0_531 * __noswap_vget_lane_f32(__rev1_531, __p2_531); \
53993  __ret_531; \
53994 })
53995 #endif
53996 
53997 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53998  float64x1_t __s0 = __p0; \
53999  float64x1_t __s1 = __p1; \
54000  float64x1_t __ret; \
54001  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
54002  __ret; \
54003 })
54004 #ifdef __LITTLE_ENDIAN__
54005 #define vmulq_lane_f64(__p0_532, __p1_532, __p2_532) __extension__ ({ \
54006  float64x2_t __s0_532 = __p0_532; \
54007  float64x1_t __s1_532 = __p1_532; \
54008  float64x2_t __ret_532; \
54009  __ret_532 = __s0_532 * splatq_lane_f64(__s1_532, __p2_532); \
54010  __ret_532; \
54011 })
54012 #else
54013 #define vmulq_lane_f64(__p0_533, __p1_533, __p2_533) __extension__ ({ \
54014  float64x2_t __s0_533 = __p0_533; \
54015  float64x1_t __s1_533 = __p1_533; \
54016  float64x2_t __rev0_533; __rev0_533 = __builtin_shufflevector(__s0_533, __s0_533, 1, 0); \
54017  float64x2_t __ret_533; \
54018  __ret_533 = __rev0_533 * __noswap_splatq_lane_f64(__s1_533, __p2_533); \
54019  __ret_533 = __builtin_shufflevector(__ret_533, __ret_533, 1, 0); \
54020  __ret_533; \
54021 })
54022 #endif
54023 
54024 #ifdef __LITTLE_ENDIAN__
54025 #define vmuld_laneq_f64(__p0_534, __p1_534, __p2_534) __extension__ ({ \
54026  float64_t __s0_534 = __p0_534; \
54027  float64x2_t __s1_534 = __p1_534; \
54028  float64_t __ret_534; \
54029  __ret_534 = __s0_534 * vgetq_lane_f64(__s1_534, __p2_534); \
54030  __ret_534; \
54031 })
54032 #else
54033 #define vmuld_laneq_f64(__p0_535, __p1_535, __p2_535) __extension__ ({ \
54034  float64_t __s0_535 = __p0_535; \
54035  float64x2_t __s1_535 = __p1_535; \
54036  float64x2_t __rev1_535; __rev1_535 = __builtin_shufflevector(__s1_535, __s1_535, 1, 0); \
54037  float64_t __ret_535; \
54038  __ret_535 = __s0_535 * __noswap_vgetq_lane_f64(__rev1_535, __p2_535); \
54039  __ret_535; \
54040 })
54041 #endif
54042 
54043 #ifdef __LITTLE_ENDIAN__
54044 #define vmuls_laneq_f32(__p0_536, __p1_536, __p2_536) __extension__ ({ \
54045  float32_t __s0_536 = __p0_536; \
54046  float32x4_t __s1_536 = __p1_536; \
54047  float32_t __ret_536; \
54048  __ret_536 = __s0_536 * vgetq_lane_f32(__s1_536, __p2_536); \
54049  __ret_536; \
54050 })
54051 #else
54052 #define vmuls_laneq_f32(__p0_537, __p1_537, __p2_537) __extension__ ({ \
54053  float32_t __s0_537 = __p0_537; \
54054  float32x4_t __s1_537 = __p1_537; \
54055  float32x4_t __rev1_537; __rev1_537 = __builtin_shufflevector(__s1_537, __s1_537, 3, 2, 1, 0); \
54056  float32_t __ret_537; \
54057  __ret_537 = __s0_537 * __noswap_vgetq_lane_f32(__rev1_537, __p2_537); \
54058  __ret_537; \
54059 })
54060 #endif
54061 
54062 #ifdef __LITTLE_ENDIAN__
54063 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54064  float64x1_t __s0 = __p0; \
54065  float64x2_t __s1 = __p1; \
54066  float64x1_t __ret; \
54067  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
54068  __ret; \
54069 })
54070 #else
54071 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54072  float64x1_t __s0 = __p0; \
54073  float64x2_t __s1 = __p1; \
54074  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54075  float64x1_t __ret; \
54076  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
54077  __ret; \
54078 })
54079 #endif
54080 
54081 #ifdef __LITTLE_ENDIAN__
54082 #define vmulq_laneq_u32(__p0_538, __p1_538, __p2_538) __extension__ ({ \
54083  uint32x4_t __s0_538 = __p0_538; \
54084  uint32x4_t __s1_538 = __p1_538; \
54085  uint32x4_t __ret_538; \
54086  __ret_538 = __s0_538 * splatq_laneq_u32(__s1_538, __p2_538); \
54087  __ret_538; \
54088 })
54089 #else
54090 #define vmulq_laneq_u32(__p0_539, __p1_539, __p2_539) __extension__ ({ \
54091  uint32x4_t __s0_539 = __p0_539; \
54092  uint32x4_t __s1_539 = __p1_539; \
54093  uint32x4_t __rev0_539; __rev0_539 = __builtin_shufflevector(__s0_539, __s0_539, 3, 2, 1, 0); \
54094  uint32x4_t __rev1_539; __rev1_539 = __builtin_shufflevector(__s1_539, __s1_539, 3, 2, 1, 0); \
54095  uint32x4_t __ret_539; \
54096  __ret_539 = __rev0_539 * __noswap_splatq_laneq_u32(__rev1_539, __p2_539); \
54097  __ret_539 = __builtin_shufflevector(__ret_539, __ret_539, 3, 2, 1, 0); \
54098  __ret_539; \
54099 })
54100 #endif
54101 
54102 #ifdef __LITTLE_ENDIAN__
54103 #define vmulq_laneq_u16(__p0_540, __p1_540, __p2_540) __extension__ ({ \
54104  uint16x8_t __s0_540 = __p0_540; \
54105  uint16x8_t __s1_540 = __p1_540; \
54106  uint16x8_t __ret_540; \
54107  __ret_540 = __s0_540 * splatq_laneq_u16(__s1_540, __p2_540); \
54108  __ret_540; \
54109 })
54110 #else
54111 #define vmulq_laneq_u16(__p0_541, __p1_541, __p2_541) __extension__ ({ \
54112  uint16x8_t __s0_541 = __p0_541; \
54113  uint16x8_t __s1_541 = __p1_541; \
54114  uint16x8_t __rev0_541; __rev0_541 = __builtin_shufflevector(__s0_541, __s0_541, 7, 6, 5, 4, 3, 2, 1, 0); \
54115  uint16x8_t __rev1_541; __rev1_541 = __builtin_shufflevector(__s1_541, __s1_541, 7, 6, 5, 4, 3, 2, 1, 0); \
54116  uint16x8_t __ret_541; \
54117  __ret_541 = __rev0_541 * __noswap_splatq_laneq_u16(__rev1_541, __p2_541); \
54118  __ret_541 = __builtin_shufflevector(__ret_541, __ret_541, 7, 6, 5, 4, 3, 2, 1, 0); \
54119  __ret_541; \
54120 })
54121 #endif
54122 
54123 #ifdef __LITTLE_ENDIAN__
54124 #define vmulq_laneq_f64(__p0_542, __p1_542, __p2_542) __extension__ ({ \
54125  float64x2_t __s0_542 = __p0_542; \
54126  float64x2_t __s1_542 = __p1_542; \
54127  float64x2_t __ret_542; \
54128  __ret_542 = __s0_542 * splatq_laneq_f64(__s1_542, __p2_542); \
54129  __ret_542; \
54130 })
54131 #else
54132 #define vmulq_laneq_f64(__p0_543, __p1_543, __p2_543) __extension__ ({ \
54133  float64x2_t __s0_543 = __p0_543; \
54134  float64x2_t __s1_543 = __p1_543; \
54135  float64x2_t __rev0_543; __rev0_543 = __builtin_shufflevector(__s0_543, __s0_543, 1, 0); \
54136  float64x2_t __rev1_543; __rev1_543 = __builtin_shufflevector(__s1_543, __s1_543, 1, 0); \
54137  float64x2_t __ret_543; \
54138  __ret_543 = __rev0_543 * __noswap_splatq_laneq_f64(__rev1_543, __p2_543); \
54139  __ret_543 = __builtin_shufflevector(__ret_543, __ret_543, 1, 0); \
54140  __ret_543; \
54141 })
54142 #endif
54143 
54144 #ifdef __LITTLE_ENDIAN__
54145 #define vmulq_laneq_f32(__p0_544, __p1_544, __p2_544) __extension__ ({ \
54146  float32x4_t __s0_544 = __p0_544; \
54147  float32x4_t __s1_544 = __p1_544; \
54148  float32x4_t __ret_544; \
54149  __ret_544 = __s0_544 * splatq_laneq_f32(__s1_544, __p2_544); \
54150  __ret_544; \
54151 })
54152 #else
54153 #define vmulq_laneq_f32(__p0_545, __p1_545, __p2_545) __extension__ ({ \
54154  float32x4_t __s0_545 = __p0_545; \
54155  float32x4_t __s1_545 = __p1_545; \
54156  float32x4_t __rev0_545; __rev0_545 = __builtin_shufflevector(__s0_545, __s0_545, 3, 2, 1, 0); \
54157  float32x4_t __rev1_545; __rev1_545 = __builtin_shufflevector(__s1_545, __s1_545, 3, 2, 1, 0); \
54158  float32x4_t __ret_545; \
54159  __ret_545 = __rev0_545 * __noswap_splatq_laneq_f32(__rev1_545, __p2_545); \
54160  __ret_545 = __builtin_shufflevector(__ret_545, __ret_545, 3, 2, 1, 0); \
54161  __ret_545; \
54162 })
54163 #endif
54164 
54165 #ifdef __LITTLE_ENDIAN__
54166 #define vmulq_laneq_s32(__p0_546, __p1_546, __p2_546) __extension__ ({ \
54167  int32x4_t __s0_546 = __p0_546; \
54168  int32x4_t __s1_546 = __p1_546; \
54169  int32x4_t __ret_546; \
54170  __ret_546 = __s0_546 * splatq_laneq_s32(__s1_546, __p2_546); \
54171  __ret_546; \
54172 })
54173 #else
54174 #define vmulq_laneq_s32(__p0_547, __p1_547, __p2_547) __extension__ ({ \
54175  int32x4_t __s0_547 = __p0_547; \
54176  int32x4_t __s1_547 = __p1_547; \
54177  int32x4_t __rev0_547; __rev0_547 = __builtin_shufflevector(__s0_547, __s0_547, 3, 2, 1, 0); \
54178  int32x4_t __rev1_547; __rev1_547 = __builtin_shufflevector(__s1_547, __s1_547, 3, 2, 1, 0); \
54179  int32x4_t __ret_547; \
54180  __ret_547 = __rev0_547 * __noswap_splatq_laneq_s32(__rev1_547, __p2_547); \
54181  __ret_547 = __builtin_shufflevector(__ret_547, __ret_547, 3, 2, 1, 0); \
54182  __ret_547; \
54183 })
54184 #endif
54185 
54186 #ifdef __LITTLE_ENDIAN__
54187 #define vmulq_laneq_s16(__p0_548, __p1_548, __p2_548) __extension__ ({ \
54188  int16x8_t __s0_548 = __p0_548; \
54189  int16x8_t __s1_548 = __p1_548; \
54190  int16x8_t __ret_548; \
54191  __ret_548 = __s0_548 * splatq_laneq_s16(__s1_548, __p2_548); \
54192  __ret_548; \
54193 })
54194 #else
54195 #define vmulq_laneq_s16(__p0_549, __p1_549, __p2_549) __extension__ ({ \
54196  int16x8_t __s0_549 = __p0_549; \
54197  int16x8_t __s1_549 = __p1_549; \
54198  int16x8_t __rev0_549; __rev0_549 = __builtin_shufflevector(__s0_549, __s0_549, 7, 6, 5, 4, 3, 2, 1, 0); \
54199  int16x8_t __rev1_549; __rev1_549 = __builtin_shufflevector(__s1_549, __s1_549, 7, 6, 5, 4, 3, 2, 1, 0); \
54200  int16x8_t __ret_549; \
54201  __ret_549 = __rev0_549 * __noswap_splatq_laneq_s16(__rev1_549, __p2_549); \
54202  __ret_549 = __builtin_shufflevector(__ret_549, __ret_549, 7, 6, 5, 4, 3, 2, 1, 0); \
54203  __ret_549; \
54204 })
54205 #endif
54206 
54207 #ifdef __LITTLE_ENDIAN__
54208 #define vmul_laneq_u32(__p0_550, __p1_550, __p2_550) __extension__ ({ \
54209  uint32x2_t __s0_550 = __p0_550; \
54210  uint32x4_t __s1_550 = __p1_550; \
54211  uint32x2_t __ret_550; \
54212  __ret_550 = __s0_550 * splat_laneq_u32(__s1_550, __p2_550); \
54213  __ret_550; \
54214 })
54215 #else
54216 #define vmul_laneq_u32(__p0_551, __p1_551, __p2_551) __extension__ ({ \
54217  uint32x2_t __s0_551 = __p0_551; \
54218  uint32x4_t __s1_551 = __p1_551; \
54219  uint32x2_t __rev0_551; __rev0_551 = __builtin_shufflevector(__s0_551, __s0_551, 1, 0); \
54220  uint32x4_t __rev1_551; __rev1_551 = __builtin_shufflevector(__s1_551, __s1_551, 3, 2, 1, 0); \
54221  uint32x2_t __ret_551; \
54222  __ret_551 = __rev0_551 * __noswap_splat_laneq_u32(__rev1_551, __p2_551); \
54223  __ret_551 = __builtin_shufflevector(__ret_551, __ret_551, 1, 0); \
54224  __ret_551; \
54225 })
54226 #endif
54227 
54228 #ifdef __LITTLE_ENDIAN__
54229 #define vmul_laneq_u16(__p0_552, __p1_552, __p2_552) __extension__ ({ \
54230  uint16x4_t __s0_552 = __p0_552; \
54231  uint16x8_t __s1_552 = __p1_552; \
54232  uint16x4_t __ret_552; \
54233  __ret_552 = __s0_552 * splat_laneq_u16(__s1_552, __p2_552); \
54234  __ret_552; \
54235 })
54236 #else
54237 #define vmul_laneq_u16(__p0_553, __p1_553, __p2_553) __extension__ ({ \
54238  uint16x4_t __s0_553 = __p0_553; \
54239  uint16x8_t __s1_553 = __p1_553; \
54240  uint16x4_t __rev0_553; __rev0_553 = __builtin_shufflevector(__s0_553, __s0_553, 3, 2, 1, 0); \
54241  uint16x8_t __rev1_553; __rev1_553 = __builtin_shufflevector(__s1_553, __s1_553, 7, 6, 5, 4, 3, 2, 1, 0); \
54242  uint16x4_t __ret_553; \
54243  __ret_553 = __rev0_553 * __noswap_splat_laneq_u16(__rev1_553, __p2_553); \
54244  __ret_553 = __builtin_shufflevector(__ret_553, __ret_553, 3, 2, 1, 0); \
54245  __ret_553; \
54246 })
54247 #endif
54248 
54249 #ifdef __LITTLE_ENDIAN__
54250 #define vmul_laneq_f32(__p0_554, __p1_554, __p2_554) __extension__ ({ \
54251  float32x2_t __s0_554 = __p0_554; \
54252  float32x4_t __s1_554 = __p1_554; \
54253  float32x2_t __ret_554; \
54254  __ret_554 = __s0_554 * splat_laneq_f32(__s1_554, __p2_554); \
54255  __ret_554; \
54256 })
54257 #else
54258 #define vmul_laneq_f32(__p0_555, __p1_555, __p2_555) __extension__ ({ \
54259  float32x2_t __s0_555 = __p0_555; \
54260  float32x4_t __s1_555 = __p1_555; \
54261  float32x2_t __rev0_555; __rev0_555 = __builtin_shufflevector(__s0_555, __s0_555, 1, 0); \
54262  float32x4_t __rev1_555; __rev1_555 = __builtin_shufflevector(__s1_555, __s1_555, 3, 2, 1, 0); \
54263  float32x2_t __ret_555; \
54264  __ret_555 = __rev0_555 * __noswap_splat_laneq_f32(__rev1_555, __p2_555); \
54265  __ret_555 = __builtin_shufflevector(__ret_555, __ret_555, 1, 0); \
54266  __ret_555; \
54267 })
54268 #endif
54269 
54270 #ifdef __LITTLE_ENDIAN__
54271 #define vmul_laneq_s32(__p0_556, __p1_556, __p2_556) __extension__ ({ \
54272  int32x2_t __s0_556 = __p0_556; \
54273  int32x4_t __s1_556 = __p1_556; \
54274  int32x2_t __ret_556; \
54275  __ret_556 = __s0_556 * splat_laneq_s32(__s1_556, __p2_556); \
54276  __ret_556; \
54277 })
54278 #else
54279 #define vmul_laneq_s32(__p0_557, __p1_557, __p2_557) __extension__ ({ \
54280  int32x2_t __s0_557 = __p0_557; \
54281  int32x4_t __s1_557 = __p1_557; \
54282  int32x2_t __rev0_557; __rev0_557 = __builtin_shufflevector(__s0_557, __s0_557, 1, 0); \
54283  int32x4_t __rev1_557; __rev1_557 = __builtin_shufflevector(__s1_557, __s1_557, 3, 2, 1, 0); \
54284  int32x2_t __ret_557; \
54285  __ret_557 = __rev0_557 * __noswap_splat_laneq_s32(__rev1_557, __p2_557); \
54286  __ret_557 = __builtin_shufflevector(__ret_557, __ret_557, 1, 0); \
54287  __ret_557; \
54288 })
54289 #endif
54290 
54291 #ifdef __LITTLE_ENDIAN__
54292 #define vmul_laneq_s16(__p0_558, __p1_558, __p2_558) __extension__ ({ \
54293  int16x4_t __s0_558 = __p0_558; \
54294  int16x8_t __s1_558 = __p1_558; \
54295  int16x4_t __ret_558; \
54296  __ret_558 = __s0_558 * splat_laneq_s16(__s1_558, __p2_558); \
54297  __ret_558; \
54298 })
54299 #else
54300 #define vmul_laneq_s16(__p0_559, __p1_559, __p2_559) __extension__ ({ \
54301  int16x4_t __s0_559 = __p0_559; \
54302  int16x8_t __s1_559 = __p1_559; \
54303  int16x4_t __rev0_559; __rev0_559 = __builtin_shufflevector(__s0_559, __s0_559, 3, 2, 1, 0); \
54304  int16x8_t __rev1_559; __rev1_559 = __builtin_shufflevector(__s1_559, __s1_559, 7, 6, 5, 4, 3, 2, 1, 0); \
54305  int16x4_t __ret_559; \
54306  __ret_559 = __rev0_559 * __noswap_splat_laneq_s16(__rev1_559, __p2_559); \
54307  __ret_559 = __builtin_shufflevector(__ret_559, __ret_559, 3, 2, 1, 0); \
54308  __ret_559; \
54309 })
54310 #endif
54311 
54312 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54313  float64x1_t __ret;
54314  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((float64x1_t)__p0, __p1);
54315  return __ret;
54316 }
54317 #ifdef __LITTLE_ENDIAN__
54318 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54319  float64x2_t __ret;
54320  __ret = __p0 * (float64x2_t) {__p1, __p1};
54321  return __ret;
54322 }
54323 #else
54324 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54325  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54326  float64x2_t __ret;
54327  __ret = __rev0 * (float64x2_t) {__p1, __p1};
54328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54329  return __ret;
54330 }
54331 #endif
54332 
54333 __ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54334  poly128_t __ret;
54335  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54336  return __ret;
54337 }
54338 #ifdef __LITTLE_ENDIAN__
54339 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54340  poly16x8_t __ret;
54341  __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
54342  return __ret;
54343 }
54344 #else
54345 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54346  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54347  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54348  poly16x8_t __ret;
54349  __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
54350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54351  return __ret;
54352 }
54353 #endif
54354 
54355 #ifdef __LITTLE_ENDIAN__
54356 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54357  uint16x8_t __ret;
54358  __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
54359  return __ret;
54360 }
54361 #else
54362 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54363  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54364  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54365  uint16x8_t __ret;
54366  __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
54367  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54368  return __ret;
54369 }
54370 #endif
54371 
54372 #ifdef __LITTLE_ENDIAN__
54373 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54374  uint64x2_t __ret;
54375  __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
54376  return __ret;
54377 }
54378 #else
54379 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54380  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54381  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54382  uint64x2_t __ret;
54383  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
54384  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54385  return __ret;
54386 }
54387 #endif
54388 
54389 #ifdef __LITTLE_ENDIAN__
54390 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54391  uint32x4_t __ret;
54392  __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
54393  return __ret;
54394 }
54395 #else
54396 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54397  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54398  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54399  uint32x4_t __ret;
54400  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
54401  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54402  return __ret;
54403 }
54404 #endif
54405 
54406 #ifdef __LITTLE_ENDIAN__
54407 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54408  int16x8_t __ret;
54409  __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
54410  return __ret;
54411 }
54412 #else
54413 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54414  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54415  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54416  int16x8_t __ret;
54417  __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
54418  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54419  return __ret;
54420 }
54421 #endif
54422 
54423 #ifdef __LITTLE_ENDIAN__
54424 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54425  int64x2_t __ret;
54426  __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
54427  return __ret;
54428 }
54429 #else
54430 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54431  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54432  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54433  int64x2_t __ret;
54434  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
54435  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54436  return __ret;
54437 }
54438 #endif
54439 
54440 #ifdef __LITTLE_ENDIAN__
54441 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54442  int32x4_t __ret;
54443  __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
54444  return __ret;
54445 }
54446 #else
54447 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54448  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54449  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54450  int32x4_t __ret;
54451  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
54452  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54453  return __ret;
54454 }
54455 #endif
54456 
54457 #ifdef __LITTLE_ENDIAN__
54458 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54459  poly128_t __ret;
54460  __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
54461  return __ret;
54462 }
54463 #else
54464 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54465  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54466  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54467  poly128_t __ret;
54468  __ret = vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
54469  return __ret;
54470 }
54471 #endif
54472 
54473 #ifdef __LITTLE_ENDIAN__
54474 #define vmull_high_lane_u32(__p0_560, __p1_560, __p2_560) __extension__ ({ \
54475  uint32x4_t __s0_560 = __p0_560; \
54476  uint32x2_t __s1_560 = __p1_560; \
54477  uint64x2_t __ret_560; \
54478  __ret_560 = vmull_u32(vget_high_u32(__s0_560), splat_lane_u32(__s1_560, __p2_560)); \
54479  __ret_560; \
54480 })
54481 #else
54482 #define vmull_high_lane_u32(__p0_561, __p1_561, __p2_561) __extension__ ({ \
54483  uint32x4_t __s0_561 = __p0_561; \
54484  uint32x2_t __s1_561 = __p1_561; \
54485  uint32x4_t __rev0_561; __rev0_561 = __builtin_shufflevector(__s0_561, __s0_561, 3, 2, 1, 0); \
54486  uint32x2_t __rev1_561; __rev1_561 = __builtin_shufflevector(__s1_561, __s1_561, 1, 0); \
54487  uint64x2_t __ret_561; \
54488  __ret_561 = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0_561), __noswap_splat_lane_u32(__rev1_561, __p2_561)); \
54489  __ret_561 = __builtin_shufflevector(__ret_561, __ret_561, 1, 0); \
54490  __ret_561; \
54491 })
54492 #endif
54493 
54494 #ifdef __LITTLE_ENDIAN__
54495 #define vmull_high_lane_u16(__p0_562, __p1_562, __p2_562) __extension__ ({ \
54496  uint16x8_t __s0_562 = __p0_562; \
54497  uint16x4_t __s1_562 = __p1_562; \
54498  uint32x4_t __ret_562; \
54499  __ret_562 = vmull_u16(vget_high_u16(__s0_562), splat_lane_u16(__s1_562, __p2_562)); \
54500  __ret_562; \
54501 })
54502 #else
54503 #define vmull_high_lane_u16(__p0_563, __p1_563, __p2_563) __extension__ ({ \
54504  uint16x8_t __s0_563 = __p0_563; \
54505  uint16x4_t __s1_563 = __p1_563; \
54506  uint16x8_t __rev0_563; __rev0_563 = __builtin_shufflevector(__s0_563, __s0_563, 7, 6, 5, 4, 3, 2, 1, 0); \
54507  uint16x4_t __rev1_563; __rev1_563 = __builtin_shufflevector(__s1_563, __s1_563, 3, 2, 1, 0); \
54508  uint32x4_t __ret_563; \
54509  __ret_563 = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0_563), __noswap_splat_lane_u16(__rev1_563, __p2_563)); \
54510  __ret_563 = __builtin_shufflevector(__ret_563, __ret_563, 3, 2, 1, 0); \
54511  __ret_563; \
54512 })
54513 #endif
54514 
54515 #ifdef __LITTLE_ENDIAN__
54516 #define vmull_high_lane_s32(__p0_564, __p1_564, __p2_564) __extension__ ({ \
54517  int32x4_t __s0_564 = __p0_564; \
54518  int32x2_t __s1_564 = __p1_564; \
54519  int64x2_t __ret_564; \
54520  __ret_564 = vmull_s32(vget_high_s32(__s0_564), splat_lane_s32(__s1_564, __p2_564)); \
54521  __ret_564; \
54522 })
54523 #else
54524 #define vmull_high_lane_s32(__p0_565, __p1_565, __p2_565) __extension__ ({ \
54525  int32x4_t __s0_565 = __p0_565; \
54526  int32x2_t __s1_565 = __p1_565; \
54527  int32x4_t __rev0_565; __rev0_565 = __builtin_shufflevector(__s0_565, __s0_565, 3, 2, 1, 0); \
54528  int32x2_t __rev1_565; __rev1_565 = __builtin_shufflevector(__s1_565, __s1_565, 1, 0); \
54529  int64x2_t __ret_565; \
54530  __ret_565 = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0_565), __noswap_splat_lane_s32(__rev1_565, __p2_565)); \
54531  __ret_565 = __builtin_shufflevector(__ret_565, __ret_565, 1, 0); \
54532  __ret_565; \
54533 })
54534 #endif
54535 
54536 #ifdef __LITTLE_ENDIAN__
54537 #define vmull_high_lane_s16(__p0_566, __p1_566, __p2_566) __extension__ ({ \
54538  int16x8_t __s0_566 = __p0_566; \
54539  int16x4_t __s1_566 = __p1_566; \
54540  int32x4_t __ret_566; \
54541  __ret_566 = vmull_s16(vget_high_s16(__s0_566), splat_lane_s16(__s1_566, __p2_566)); \
54542  __ret_566; \
54543 })
54544 #else
54545 #define vmull_high_lane_s16(__p0_567, __p1_567, __p2_567) __extension__ ({ \
54546  int16x8_t __s0_567 = __p0_567; \
54547  int16x4_t __s1_567 = __p1_567; \
54548  int16x8_t __rev0_567; __rev0_567 = __builtin_shufflevector(__s0_567, __s0_567, 7, 6, 5, 4, 3, 2, 1, 0); \
54549  int16x4_t __rev1_567; __rev1_567 = __builtin_shufflevector(__s1_567, __s1_567, 3, 2, 1, 0); \
54550  int32x4_t __ret_567; \
54551  __ret_567 = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0_567), __noswap_splat_lane_s16(__rev1_567, __p2_567)); \
54552  __ret_567 = __builtin_shufflevector(__ret_567, __ret_567, 3, 2, 1, 0); \
54553  __ret_567; \
54554 })
54555 #endif
54556 
54557 #ifdef __LITTLE_ENDIAN__
54558 #define vmull_high_laneq_u32(__p0_568, __p1_568, __p2_568) __extension__ ({ \
54559  uint32x4_t __s0_568 = __p0_568; \
54560  uint32x4_t __s1_568 = __p1_568; \
54561  uint64x2_t __ret_568; \
54562  __ret_568 = vmull_u32(vget_high_u32(__s0_568), splat_laneq_u32(__s1_568, __p2_568)); \
54563  __ret_568; \
54564 })
54565 #else
54566 #define vmull_high_laneq_u32(__p0_569, __p1_569, __p2_569) __extension__ ({ \
54567  uint32x4_t __s0_569 = __p0_569; \
54568  uint32x4_t __s1_569 = __p1_569; \
54569  uint32x4_t __rev0_569; __rev0_569 = __builtin_shufflevector(__s0_569, __s0_569, 3, 2, 1, 0); \
54570  uint32x4_t __rev1_569; __rev1_569 = __builtin_shufflevector(__s1_569, __s1_569, 3, 2, 1, 0); \
54571  uint64x2_t __ret_569; \
54572  __ret_569 = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0_569), __noswap_splat_laneq_u32(__rev1_569, __p2_569)); \
54573  __ret_569 = __builtin_shufflevector(__ret_569, __ret_569, 1, 0); \
54574  __ret_569; \
54575 })
54576 #endif
54577 
54578 #ifdef __LITTLE_ENDIAN__
54579 #define vmull_high_laneq_u16(__p0_570, __p1_570, __p2_570) __extension__ ({ \
54580  uint16x8_t __s0_570 = __p0_570; \
54581  uint16x8_t __s1_570 = __p1_570; \
54582  uint32x4_t __ret_570; \
54583  __ret_570 = vmull_u16(vget_high_u16(__s0_570), splat_laneq_u16(__s1_570, __p2_570)); \
54584  __ret_570; \
54585 })
54586 #else
54587 #define vmull_high_laneq_u16(__p0_571, __p1_571, __p2_571) __extension__ ({ \
54588  uint16x8_t __s0_571 = __p0_571; \
54589  uint16x8_t __s1_571 = __p1_571; \
54590  uint16x8_t __rev0_571; __rev0_571 = __builtin_shufflevector(__s0_571, __s0_571, 7, 6, 5, 4, 3, 2, 1, 0); \
54591  uint16x8_t __rev1_571; __rev1_571 = __builtin_shufflevector(__s1_571, __s1_571, 7, 6, 5, 4, 3, 2, 1, 0); \
54592  uint32x4_t __ret_571; \
54593  __ret_571 = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0_571), __noswap_splat_laneq_u16(__rev1_571, __p2_571)); \
54594  __ret_571 = __builtin_shufflevector(__ret_571, __ret_571, 3, 2, 1, 0); \
54595  __ret_571; \
54596 })
54597 #endif
54598 
54599 #ifdef __LITTLE_ENDIAN__
54600 #define vmull_high_laneq_s32(__p0_572, __p1_572, __p2_572) __extension__ ({ \
54601  int32x4_t __s0_572 = __p0_572; \
54602  int32x4_t __s1_572 = __p1_572; \
54603  int64x2_t __ret_572; \
54604  __ret_572 = vmull_s32(vget_high_s32(__s0_572), splat_laneq_s32(__s1_572, __p2_572)); \
54605  __ret_572; \
54606 })
54607 #else
54608 #define vmull_high_laneq_s32(__p0_573, __p1_573, __p2_573) __extension__ ({ \
54609  int32x4_t __s0_573 = __p0_573; \
54610  int32x4_t __s1_573 = __p1_573; \
54611  int32x4_t __rev0_573; __rev0_573 = __builtin_shufflevector(__s0_573, __s0_573, 3, 2, 1, 0); \
54612  int32x4_t __rev1_573; __rev1_573 = __builtin_shufflevector(__s1_573, __s1_573, 3, 2, 1, 0); \
54613  int64x2_t __ret_573; \
54614  __ret_573 = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0_573), __noswap_splat_laneq_s32(__rev1_573, __p2_573)); \
54615  __ret_573 = __builtin_shufflevector(__ret_573, __ret_573, 1, 0); \
54616  __ret_573; \
54617 })
54618 #endif
54619 
54620 #ifdef __LITTLE_ENDIAN__
54621 #define vmull_high_laneq_s16(__p0_574, __p1_574, __p2_574) __extension__ ({ \
54622  int16x8_t __s0_574 = __p0_574; \
54623  int16x8_t __s1_574 = __p1_574; \
54624  int32x4_t __ret_574; \
54625  __ret_574 = vmull_s16(vget_high_s16(__s0_574), splat_laneq_s16(__s1_574, __p2_574)); \
54626  __ret_574; \
54627 })
54628 #else
54629 #define vmull_high_laneq_s16(__p0_575, __p1_575, __p2_575) __extension__ ({ \
54630  int16x8_t __s0_575 = __p0_575; \
54631  int16x8_t __s1_575 = __p1_575; \
54632  int16x8_t __rev0_575; __rev0_575 = __builtin_shufflevector(__s0_575, __s0_575, 7, 6, 5, 4, 3, 2, 1, 0); \
54633  int16x8_t __rev1_575; __rev1_575 = __builtin_shufflevector(__s1_575, __s1_575, 7, 6, 5, 4, 3, 2, 1, 0); \
54634  int32x4_t __ret_575; \
54635  __ret_575 = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0_575), __noswap_splat_laneq_s16(__rev1_575, __p2_575)); \
54636  __ret_575 = __builtin_shufflevector(__ret_575, __ret_575, 3, 2, 1, 0); \
54637  __ret_575; \
54638 })
54639 #endif
54640 
54641 #ifdef __LITTLE_ENDIAN__
54642 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54643  uint64x2_t __ret;
54644  __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
54645  return __ret;
54646 }
54647 #else
54648 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54649  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54650  uint64x2_t __ret;
54651  __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
54652  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54653  return __ret;
54654 }
54655 #endif
54656 
54657 #ifdef __LITTLE_ENDIAN__
54658 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54659  uint32x4_t __ret;
54660  __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
54661  return __ret;
54662 }
54663 #else
54664 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54665  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54666  uint32x4_t __ret;
54667  __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
54668  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54669  return __ret;
54670 }
54671 #endif
54672 
54673 #ifdef __LITTLE_ENDIAN__
54674 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54675  int64x2_t __ret;
54676  __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
54677  return __ret;
54678 }
54679 #else
54680 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54681  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54682  int64x2_t __ret;
54683  __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
54684  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54685  return __ret;
54686 }
54687 #endif
54688 
54689 #ifdef __LITTLE_ENDIAN__
54690 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54691  int32x4_t __ret;
54692  __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
54693  return __ret;
54694 }
54695 #else
54696 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54697  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54698  int32x4_t __ret;
54699  __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
54700  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54701  return __ret;
54702 }
54703 #endif
54704 
54705 #ifdef __LITTLE_ENDIAN__
54706 #define vmull_laneq_u32(__p0_576, __p1_576, __p2_576) __extension__ ({ \
54707  uint32x2_t __s0_576 = __p0_576; \
54708  uint32x4_t __s1_576 = __p1_576; \
54709  uint64x2_t __ret_576; \
54710  __ret_576 = vmull_u32(__s0_576, splat_laneq_u32(__s1_576, __p2_576)); \
54711  __ret_576; \
54712 })
54713 #else
54714 #define vmull_laneq_u32(__p0_577, __p1_577, __p2_577) __extension__ ({ \
54715  uint32x2_t __s0_577 = __p0_577; \
54716  uint32x4_t __s1_577 = __p1_577; \
54717  uint32x2_t __rev0_577; __rev0_577 = __builtin_shufflevector(__s0_577, __s0_577, 1, 0); \
54718  uint32x4_t __rev1_577; __rev1_577 = __builtin_shufflevector(__s1_577, __s1_577, 3, 2, 1, 0); \
54719  uint64x2_t __ret_577; \
54720  __ret_577 = __noswap_vmull_u32(__rev0_577, __noswap_splat_laneq_u32(__rev1_577, __p2_577)); \
54721  __ret_577 = __builtin_shufflevector(__ret_577, __ret_577, 1, 0); \
54722  __ret_577; \
54723 })
54724 #endif
54725 
54726 #ifdef __LITTLE_ENDIAN__
54727 #define vmull_laneq_u16(__p0_578, __p1_578, __p2_578) __extension__ ({ \
54728  uint16x4_t __s0_578 = __p0_578; \
54729  uint16x8_t __s1_578 = __p1_578; \
54730  uint32x4_t __ret_578; \
54731  __ret_578 = vmull_u16(__s0_578, splat_laneq_u16(__s1_578, __p2_578)); \
54732  __ret_578; \
54733 })
54734 #else
54735 #define vmull_laneq_u16(__p0_579, __p1_579, __p2_579) __extension__ ({ \
54736  uint16x4_t __s0_579 = __p0_579; \
54737  uint16x8_t __s1_579 = __p1_579; \
54738  uint16x4_t __rev0_579; __rev0_579 = __builtin_shufflevector(__s0_579, __s0_579, 3, 2, 1, 0); \
54739  uint16x8_t __rev1_579; __rev1_579 = __builtin_shufflevector(__s1_579, __s1_579, 7, 6, 5, 4, 3, 2, 1, 0); \
54740  uint32x4_t __ret_579; \
54741  __ret_579 = __noswap_vmull_u16(__rev0_579, __noswap_splat_laneq_u16(__rev1_579, __p2_579)); \
54742  __ret_579 = __builtin_shufflevector(__ret_579, __ret_579, 3, 2, 1, 0); \
54743  __ret_579; \
54744 })
54745 #endif
54746 
54747 #ifdef __LITTLE_ENDIAN__
54748 #define vmull_laneq_s32(__p0_580, __p1_580, __p2_580) __extension__ ({ \
54749  int32x2_t __s0_580 = __p0_580; \
54750  int32x4_t __s1_580 = __p1_580; \
54751  int64x2_t __ret_580; \
54752  __ret_580 = vmull_s32(__s0_580, splat_laneq_s32(__s1_580, __p2_580)); \
54753  __ret_580; \
54754 })
54755 #else
54756 #define vmull_laneq_s32(__p0_581, __p1_581, __p2_581) __extension__ ({ \
54757  int32x2_t __s0_581 = __p0_581; \
54758  int32x4_t __s1_581 = __p1_581; \
54759  int32x2_t __rev0_581; __rev0_581 = __builtin_shufflevector(__s0_581, __s0_581, 1, 0); \
54760  int32x4_t __rev1_581; __rev1_581 = __builtin_shufflevector(__s1_581, __s1_581, 3, 2, 1, 0); \
54761  int64x2_t __ret_581; \
54762  __ret_581 = __noswap_vmull_s32(__rev0_581, __noswap_splat_laneq_s32(__rev1_581, __p2_581)); \
54763  __ret_581 = __builtin_shufflevector(__ret_581, __ret_581, 1, 0); \
54764  __ret_581; \
54765 })
54766 #endif
54767 
54768 #ifdef __LITTLE_ENDIAN__
54769 #define vmull_laneq_s16(__p0_582, __p1_582, __p2_582) __extension__ ({ \
54770  int16x4_t __s0_582 = __p0_582; \
54771  int16x8_t __s1_582 = __p1_582; \
54772  int32x4_t __ret_582; \
54773  __ret_582 = vmull_s16(__s0_582, splat_laneq_s16(__s1_582, __p2_582)); \
54774  __ret_582; \
54775 })
54776 #else
54777 #define vmull_laneq_s16(__p0_583, __p1_583, __p2_583) __extension__ ({ \
54778  int16x4_t __s0_583 = __p0_583; \
54779  int16x8_t __s1_583 = __p1_583; \
54780  int16x4_t __rev0_583; __rev0_583 = __builtin_shufflevector(__s0_583, __s0_583, 3, 2, 1, 0); \
54781  int16x8_t __rev1_583; __rev1_583 = __builtin_shufflevector(__s1_583, __s1_583, 7, 6, 5, 4, 3, 2, 1, 0); \
54782  int32x4_t __ret_583; \
54783  __ret_583 = __noswap_vmull_s16(__rev0_583, __noswap_splat_laneq_s16(__rev1_583, __p2_583)); \
54784  __ret_583 = __builtin_shufflevector(__ret_583, __ret_583, 3, 2, 1, 0); \
54785  __ret_583; \
54786 })
54787 #endif
54788 
54789 #ifdef __LITTLE_ENDIAN__
54790 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54791  float64x2_t __ret;
54792  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54793  return __ret;
54794 }
54795 #else
54796 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54797  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54798  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54799  float64x2_t __ret;
54800  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54802  return __ret;
54803 }
54804 __ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54805  float64x2_t __ret;
54806  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54807  return __ret;
54808 }
54809 #endif
54810 
54811 #ifdef __LITTLE_ENDIAN__
54812 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54813  float32x4_t __ret;
54814  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54815  return __ret;
54816 }
54817 #else
54818 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54819  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54820  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54821  float32x4_t __ret;
54822  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54824  return __ret;
54825 }
54826 __ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54827  float32x4_t __ret;
54828  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54829  return __ret;
54830 }
54831 #endif
54832 
54833 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54834  float64x1_t __ret;
54835  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54836  return __ret;
54837 }
54838 #ifdef __LITTLE_ENDIAN__
54839 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54840  float32x2_t __ret;
54841  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54842  return __ret;
54843 }
54844 #else
54845 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54846  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54847  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54848  float32x2_t __ret;
54849  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54851  return __ret;
54852 }
54853 __ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54854  float32x2_t __ret;
54855  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54856  return __ret;
54857 }
54858 #endif
54859 
54860 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54861  float64_t __ret;
54862  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54863  return __ret;
54864 }
54865 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54866  float32_t __ret;
54867  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54868  return __ret;
54869 }
54870 #define vmulxd_lane_f64(__p0_584, __p1_584, __p2_584) __extension__ ({ \
54871  float64_t __s0_584 = __p0_584; \
54872  float64x1_t __s1_584 = __p1_584; \
54873  float64_t __ret_584; \
54874  __ret_584 = vmulxd_f64(__s0_584, vget_lane_f64(__s1_584, __p2_584)); \
54875  __ret_584; \
54876 })
54877 #ifdef __LITTLE_ENDIAN__
54878 #define vmulxs_lane_f32(__p0_585, __p1_585, __p2_585) __extension__ ({ \
54879  float32_t __s0_585 = __p0_585; \
54880  float32x2_t __s1_585 = __p1_585; \
54881  float32_t __ret_585; \
54882  __ret_585 = vmulxs_f32(__s0_585, vget_lane_f32(__s1_585, __p2_585)); \
54883  __ret_585; \
54884 })
54885 #else
54886 #define vmulxs_lane_f32(__p0_586, __p1_586, __p2_586) __extension__ ({ \
54887  float32_t __s0_586 = __p0_586; \
54888  float32x2_t __s1_586 = __p1_586; \
54889  float32x2_t __rev1_586; __rev1_586 = __builtin_shufflevector(__s1_586, __s1_586, 1, 0); \
54890  float32_t __ret_586; \
54891  __ret_586 = vmulxs_f32(__s0_586, __noswap_vget_lane_f32(__rev1_586, __p2_586)); \
54892  __ret_586; \
54893 })
54894 #endif
54895 
54896 #ifdef __LITTLE_ENDIAN__
54897 #define vmulxq_lane_f64(__p0_587, __p1_587, __p2_587) __extension__ ({ \
54898  float64x2_t __s0_587 = __p0_587; \
54899  float64x1_t __s1_587 = __p1_587; \
54900  float64x2_t __ret_587; \
54901  __ret_587 = vmulxq_f64(__s0_587, splatq_lane_f64(__s1_587, __p2_587)); \
54902  __ret_587; \
54903 })
54904 #else
54905 #define vmulxq_lane_f64(__p0_588, __p1_588, __p2_588) __extension__ ({ \
54906  float64x2_t __s0_588 = __p0_588; \
54907  float64x1_t __s1_588 = __p1_588; \
54908  float64x2_t __rev0_588; __rev0_588 = __builtin_shufflevector(__s0_588, __s0_588, 1, 0); \
54909  float64x2_t __ret_588; \
54910  __ret_588 = __noswap_vmulxq_f64(__rev0_588, __noswap_splatq_lane_f64(__s1_588, __p2_588)); \
54911  __ret_588 = __builtin_shufflevector(__ret_588, __ret_588, 1, 0); \
54912  __ret_588; \
54913 })
54914 #endif
54915 
54916 #ifdef __LITTLE_ENDIAN__
54917 #define vmulxq_lane_f32(__p0_589, __p1_589, __p2_589) __extension__ ({ \
54918  float32x4_t __s0_589 = __p0_589; \
54919  float32x2_t __s1_589 = __p1_589; \
54920  float32x4_t __ret_589; \
54921  __ret_589 = vmulxq_f32(__s0_589, splatq_lane_f32(__s1_589, __p2_589)); \
54922  __ret_589; \
54923 })
54924 #else
54925 #define vmulxq_lane_f32(__p0_590, __p1_590, __p2_590) __extension__ ({ \
54926  float32x4_t __s0_590 = __p0_590; \
54927  float32x2_t __s1_590 = __p1_590; \
54928  float32x4_t __rev0_590; __rev0_590 = __builtin_shufflevector(__s0_590, __s0_590, 3, 2, 1, 0); \
54929  float32x2_t __rev1_590; __rev1_590 = __builtin_shufflevector(__s1_590, __s1_590, 1, 0); \
54930  float32x4_t __ret_590; \
54931  __ret_590 = __noswap_vmulxq_f32(__rev0_590, __noswap_splatq_lane_f32(__rev1_590, __p2_590)); \
54932  __ret_590 = __builtin_shufflevector(__ret_590, __ret_590, 3, 2, 1, 0); \
54933  __ret_590; \
54934 })
54935 #endif
54936 
54937 #ifdef __LITTLE_ENDIAN__
54938 #define vmulx_lane_f32(__p0_591, __p1_591, __p2_591) __extension__ ({ \
54939  float32x2_t __s0_591 = __p0_591; \
54940  float32x2_t __s1_591 = __p1_591; \
54941  float32x2_t __ret_591; \
54942  __ret_591 = vmulx_f32(__s0_591, splat_lane_f32(__s1_591, __p2_591)); \
54943  __ret_591; \
54944 })
54945 #else
54946 #define vmulx_lane_f32(__p0_592, __p1_592, __p2_592) __extension__ ({ \
54947  float32x2_t __s0_592 = __p0_592; \
54948  float32x2_t __s1_592 = __p1_592; \
54949  float32x2_t __rev0_592; __rev0_592 = __builtin_shufflevector(__s0_592, __s0_592, 1, 0); \
54950  float32x2_t __rev1_592; __rev1_592 = __builtin_shufflevector(__s1_592, __s1_592, 1, 0); \
54951  float32x2_t __ret_592; \
54952  __ret_592 = __noswap_vmulx_f32(__rev0_592, __noswap_splat_lane_f32(__rev1_592, __p2_592)); \
54953  __ret_592 = __builtin_shufflevector(__ret_592, __ret_592, 1, 0); \
54954  __ret_592; \
54955 })
54956 #endif
54957 
54958 #ifdef __LITTLE_ENDIAN__
54959 #define vmulxd_laneq_f64(__p0_593, __p1_593, __p2_593) __extension__ ({ \
54960  float64_t __s0_593 = __p0_593; \
54961  float64x2_t __s1_593 = __p1_593; \
54962  float64_t __ret_593; \
54963  __ret_593 = vmulxd_f64(__s0_593, vgetq_lane_f64(__s1_593, __p2_593)); \
54964  __ret_593; \
54965 })
54966 #else
54967 #define vmulxd_laneq_f64(__p0_594, __p1_594, __p2_594) __extension__ ({ \
54968  float64_t __s0_594 = __p0_594; \
54969  float64x2_t __s1_594 = __p1_594; \
54970  float64x2_t __rev1_594; __rev1_594 = __builtin_shufflevector(__s1_594, __s1_594, 1, 0); \
54971  float64_t __ret_594; \
54972  __ret_594 = vmulxd_f64(__s0_594, __noswap_vgetq_lane_f64(__rev1_594, __p2_594)); \
54973  __ret_594; \
54974 })
54975 #endif
54976 
54977 #ifdef __LITTLE_ENDIAN__
54978 #define vmulxs_laneq_f32(__p0_595, __p1_595, __p2_595) __extension__ ({ \
54979  float32_t __s0_595 = __p0_595; \
54980  float32x4_t __s1_595 = __p1_595; \
54981  float32_t __ret_595; \
54982  __ret_595 = vmulxs_f32(__s0_595, vgetq_lane_f32(__s1_595, __p2_595)); \
54983  __ret_595; \
54984 })
54985 #else
54986 #define vmulxs_laneq_f32(__p0_596, __p1_596, __p2_596) __extension__ ({ \
54987  float32_t __s0_596 = __p0_596; \
54988  float32x4_t __s1_596 = __p1_596; \
54989  float32x4_t __rev1_596; __rev1_596 = __builtin_shufflevector(__s1_596, __s1_596, 3, 2, 1, 0); \
54990  float32_t __ret_596; \
54991  __ret_596 = vmulxs_f32(__s0_596, __noswap_vgetq_lane_f32(__rev1_596, __p2_596)); \
54992  __ret_596; \
54993 })
54994 #endif
54995 
54996 #ifdef __LITTLE_ENDIAN__
54997 #define vmulxq_laneq_f64(__p0_597, __p1_597, __p2_597) __extension__ ({ \
54998  float64x2_t __s0_597 = __p0_597; \
54999  float64x2_t __s1_597 = __p1_597; \
55000  float64x2_t __ret_597; \
55001  __ret_597 = vmulxq_f64(__s0_597, splatq_laneq_f64(__s1_597, __p2_597)); \
55002  __ret_597; \
55003 })
55004 #else
55005 #define vmulxq_laneq_f64(__p0_598, __p1_598, __p2_598) __extension__ ({ \
55006  float64x2_t __s0_598 = __p0_598; \
55007  float64x2_t __s1_598 = __p1_598; \
55008  float64x2_t __rev0_598; __rev0_598 = __builtin_shufflevector(__s0_598, __s0_598, 1, 0); \
55009  float64x2_t __rev1_598; __rev1_598 = __builtin_shufflevector(__s1_598, __s1_598, 1, 0); \
55010  float64x2_t __ret_598; \
55011  __ret_598 = __noswap_vmulxq_f64(__rev0_598, __noswap_splatq_laneq_f64(__rev1_598, __p2_598)); \
55012  __ret_598 = __builtin_shufflevector(__ret_598, __ret_598, 1, 0); \
55013  __ret_598; \
55014 })
55015 #endif
55016 
55017 #ifdef __LITTLE_ENDIAN__
55018 #define vmulxq_laneq_f32(__p0_599, __p1_599, __p2_599) __extension__ ({ \
55019  float32x4_t __s0_599 = __p0_599; \
55020  float32x4_t __s1_599 = __p1_599; \
55021  float32x4_t __ret_599; \
55022  __ret_599 = vmulxq_f32(__s0_599, splatq_laneq_f32(__s1_599, __p2_599)); \
55023  __ret_599; \
55024 })
55025 #else
55026 #define vmulxq_laneq_f32(__p0_600, __p1_600, __p2_600) __extension__ ({ \
55027  float32x4_t __s0_600 = __p0_600; \
55028  float32x4_t __s1_600 = __p1_600; \
55029  float32x4_t __rev0_600; __rev0_600 = __builtin_shufflevector(__s0_600, __s0_600, 3, 2, 1, 0); \
55030  float32x4_t __rev1_600; __rev1_600 = __builtin_shufflevector(__s1_600, __s1_600, 3, 2, 1, 0); \
55031  float32x4_t __ret_600; \
55032  __ret_600 = __noswap_vmulxq_f32(__rev0_600, __noswap_splatq_laneq_f32(__rev1_600, __p2_600)); \
55033  __ret_600 = __builtin_shufflevector(__ret_600, __ret_600, 3, 2, 1, 0); \
55034  __ret_600; \
55035 })
55036 #endif
55037 
55038 #ifdef __LITTLE_ENDIAN__
55039 #define vmulx_laneq_f32(__p0_601, __p1_601, __p2_601) __extension__ ({ \
55040  float32x2_t __s0_601 = __p0_601; \
55041  float32x4_t __s1_601 = __p1_601; \
55042  float32x2_t __ret_601; \
55043  __ret_601 = vmulx_f32(__s0_601, splat_laneq_f32(__s1_601, __p2_601)); \
55044  __ret_601; \
55045 })
55046 #else
55047 #define vmulx_laneq_f32(__p0_602, __p1_602, __p2_602) __extension__ ({ \
55048  float32x2_t __s0_602 = __p0_602; \
55049  float32x4_t __s1_602 = __p1_602; \
55050  float32x2_t __rev0_602; __rev0_602 = __builtin_shufflevector(__s0_602, __s0_602, 1, 0); \
55051  float32x4_t __rev1_602; __rev1_602 = __builtin_shufflevector(__s1_602, __s1_602, 3, 2, 1, 0); \
55052  float32x2_t __ret_602; \
55053  __ret_602 = __noswap_vmulx_f32(__rev0_602, __noswap_splat_laneq_f32(__rev1_602, __p2_602)); \
55054  __ret_602 = __builtin_shufflevector(__ret_602, __ret_602, 1, 0); \
55055  __ret_602; \
55056 })
55057 #endif
55058 
55059 #ifdef __LITTLE_ENDIAN__
55060 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
55061  float64x2_t __ret;
55062  __ret = -__p0;
55063  return __ret;
55064 }
55065 #else
55066 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
55067  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55068  float64x2_t __ret;
55069  __ret = -__rev0;
55070  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55071  return __ret;
55072 }
55073 #endif
55074 
55075 #ifdef __LITTLE_ENDIAN__
55076 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
55077  int64x2_t __ret;
55078  __ret = -__p0;
55079  return __ret;
55080 }
55081 #else
55082 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
55083  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55084  int64x2_t __ret;
55085  __ret = -__rev0;
55086  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55087  return __ret;
55088 }
55089 #endif
55090 
55091 __ai float64x1_t vneg_f64(float64x1_t __p0) {
55092  float64x1_t __ret;
55093  __ret = -__p0;
55094  return __ret;
55095 }
55096 __ai int64x1_t vneg_s64(int64x1_t __p0) {
55097  int64x1_t __ret;
55098  __ret = -__p0;
55099  return __ret;
55100 }
55101 __ai int64_t vnegd_s64(int64_t __p0) {
55102  int64_t __ret;
55103  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55104  return __ret;
55105 }
55106 #ifdef __LITTLE_ENDIAN__
55107 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55108  uint8x16_t __ret;
55109  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55110  return __ret;
55111 }
55112 #else
55113 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55114  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55115  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55116  uint8x16_t __ret;
55117  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55118  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55119  return __ret;
55120 }
55121 #endif
55122 
55123 #ifdef __LITTLE_ENDIAN__
55124 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55125  uint32x4_t __ret;
55126  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55127  return __ret;
55128 }
55129 #else
55130 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55131  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55132  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55133  uint32x4_t __ret;
55134  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55135  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55136  return __ret;
55137 }
55138 #endif
55139 
55140 #ifdef __LITTLE_ENDIAN__
55141 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55142  uint64x2_t __ret;
55143  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
55144  return __ret;
55145 }
55146 #else
55147 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55148  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55149  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55150  uint64x2_t __ret;
55151  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
55152  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55153  return __ret;
55154 }
55155 #endif
55156 
55157 #ifdef __LITTLE_ENDIAN__
55158 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55159  uint16x8_t __ret;
55160  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55161  return __ret;
55162 }
55163 #else
55164 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55165  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55166  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55167  uint16x8_t __ret;
55168  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55169  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55170  return __ret;
55171 }
55172 #endif
55173 
55174 #ifdef __LITTLE_ENDIAN__
55175 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55176  int8x16_t __ret;
55177  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55178  return __ret;
55179 }
55180 #else
55181 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55182  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55183  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55184  int8x16_t __ret;
55185  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55186  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55187  return __ret;
55188 }
55189 #endif
55190 
55191 #ifdef __LITTLE_ENDIAN__
55192 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55193  float64x2_t __ret;
55194  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55195  return __ret;
55196 }
55197 #else
55198 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55199  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55200  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55201  float64x2_t __ret;
55202  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55203  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55204  return __ret;
55205 }
55206 #endif
55207 
55208 #ifdef __LITTLE_ENDIAN__
55209 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55210  float32x4_t __ret;
55211  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55212  return __ret;
55213 }
55214 #else
55215 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55216  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55217  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55218  float32x4_t __ret;
55219  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55220  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55221  return __ret;
55222 }
55223 #endif
55224 
55225 #ifdef __LITTLE_ENDIAN__
55226 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55227  int32x4_t __ret;
55228  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55229  return __ret;
55230 }
55231 #else
55232 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55233  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55234  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55235  int32x4_t __ret;
55236  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55237  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55238  return __ret;
55239 }
55240 #endif
55241 
55242 #ifdef __LITTLE_ENDIAN__
55243 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55244  int64x2_t __ret;
55245  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
55246  return __ret;
55247 }
55248 #else
55249 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55250  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55251  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55252  int64x2_t __ret;
55253  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
55254  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55255  return __ret;
55256 }
55257 #endif
55258 
55259 #ifdef __LITTLE_ENDIAN__
55260 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55261  int16x8_t __ret;
55262  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55263  return __ret;
55264 }
55265 #else
55266 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55267  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55268  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55269  int16x8_t __ret;
55270  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55271  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55272  return __ret;
55273 }
55274 #endif
55275 
55276 #ifdef __LITTLE_ENDIAN__
55277 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55278  uint64_t __ret;
55279  __ret = (uint64_t) __builtin_neon_vpaddd_u64(__p0);
55280  return __ret;
55281 }
55282 #else
55283 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55284  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55285  uint64_t __ret;
55286  __ret = (uint64_t) __builtin_neon_vpaddd_u64(__rev0);
55287  return __ret;
55288 }
55289 #endif
55290 
55291 #ifdef __LITTLE_ENDIAN__
55292 __ai float64_t vpaddd_f64(float64x2_t __p0) {
55293  float64_t __ret;
55294  __ret = (float64_t) __builtin_neon_vpaddd_f64(__p0);
55295  return __ret;
55296 }
55297 #else
55298 __ai float64_t vpaddd_f64(float64x2_t __p0) {
55299  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55300  float64_t __ret;
55301  __ret = (float64_t) __builtin_neon_vpaddd_f64(__rev0);
55302  return __ret;
55303 }
55304 #endif
55305 
55306 #ifdef __LITTLE_ENDIAN__
55307 __ai int64_t vpaddd_s64(int64x2_t __p0) {
55308  int64_t __ret;
55309  __ret = (int64_t) __builtin_neon_vpaddd_s64(__p0);
55310  return __ret;
55311 }
55312 #else
55313 __ai int64_t vpaddd_s64(int64x2_t __p0) {
55314  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55315  int64_t __ret;
55316  __ret = (int64_t) __builtin_neon_vpaddd_s64(__rev0);
55317  return __ret;
55318 }
55319 #endif
55320 
55321 #ifdef __LITTLE_ENDIAN__
55322 __ai float32_t vpadds_f32(float32x2_t __p0) {
55323  float32_t __ret;
55324  __ret = (float32_t) __builtin_neon_vpadds_f32(__p0);
55325  return __ret;
55326 }
55327 #else
55328 __ai float32_t vpadds_f32(float32x2_t __p0) {
55329  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55330  float32_t __ret;
55331  __ret = (float32_t) __builtin_neon_vpadds_f32(__rev0);
55332  return __ret;
55333 }
55334 #endif
55335 
55336 #ifdef __LITTLE_ENDIAN__
55337 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55338  uint8x16_t __ret;
55339  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55340  return __ret;
55341 }
55342 #else
55343 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55344  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55345  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55346  uint8x16_t __ret;
55347  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55348  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55349  return __ret;
55350 }
55351 #endif
55352 
55353 #ifdef __LITTLE_ENDIAN__
55354 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55355  uint32x4_t __ret;
55356  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55357  return __ret;
55358 }
55359 #else
55360 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55361  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55362  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55363  uint32x4_t __ret;
55364  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55366  return __ret;
55367 }
55368 #endif
55369 
55370 #ifdef __LITTLE_ENDIAN__
55371 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55372  uint16x8_t __ret;
55373  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55374  return __ret;
55375 }
55376 #else
55377 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55378  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55379  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55380  uint16x8_t __ret;
55381  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55382  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55383  return __ret;
55384 }
55385 #endif
55386 
55387 #ifdef __LITTLE_ENDIAN__
55388 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55389  int8x16_t __ret;
55390  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55391  return __ret;
55392 }
55393 #else
55394 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55395  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55396  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55397  int8x16_t __ret;
55398  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55399  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55400  return __ret;
55401 }
55402 #endif
55403 
55404 #ifdef __LITTLE_ENDIAN__
55405 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55406  float64x2_t __ret;
55407  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55408  return __ret;
55409 }
55410 #else
55411 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55412  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55413  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55414  float64x2_t __ret;
55415  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55416  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55417  return __ret;
55418 }
55419 #endif
55420 
55421 #ifdef __LITTLE_ENDIAN__
55422 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55423  float32x4_t __ret;
55424  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55425  return __ret;
55426 }
55427 #else
55428 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55429  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55430  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55431  float32x4_t __ret;
55432  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55434  return __ret;
55435 }
55436 #endif
55437 
55438 #ifdef __LITTLE_ENDIAN__
55439 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55440  int32x4_t __ret;
55441  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55442  return __ret;
55443 }
55444 #else
55445 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55446  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55447  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55448  int32x4_t __ret;
55449  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55450  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55451  return __ret;
55452 }
55453 #endif
55454 
55455 #ifdef __LITTLE_ENDIAN__
55456 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55457  int16x8_t __ret;
55458  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55459  return __ret;
55460 }
55461 #else
55462 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55463  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55464  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55465  int16x8_t __ret;
55466  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55467  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55468  return __ret;
55469 }
55470 #endif
55471 
55472 #ifdef __LITTLE_ENDIAN__
55473 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55474  float64_t __ret;
55475  __ret = (float64_t) __builtin_neon_vpmaxqd_f64(__p0);
55476  return __ret;
55477 }
55478 #else
55479 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55480  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55481  float64_t __ret;
55482  __ret = (float64_t) __builtin_neon_vpmaxqd_f64(__rev0);
55483  return __ret;
55484 }
55485 #endif
55486 
55487 #ifdef __LITTLE_ENDIAN__
55488 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
55489  float32_t __ret;
55490  __ret = (float32_t) __builtin_neon_vpmaxs_f32(__p0);
55491  return __ret;
55492 }
55493 #else
55494 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
55495  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55496  float32_t __ret;
55497  __ret = (float32_t) __builtin_neon_vpmaxs_f32(__rev0);
55498  return __ret;
55499 }
55500 #endif
55501 
55502 #ifdef __LITTLE_ENDIAN__
55503 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55504  float64x2_t __ret;
55505  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55506  return __ret;
55507 }
55508 #else
55509 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55510  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55511  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55512  float64x2_t __ret;
55513  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55515  return __ret;
55516 }
55517 #endif
55518 
55519 #ifdef __LITTLE_ENDIAN__
55520 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55521  float32x4_t __ret;
55522  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55523  return __ret;
55524 }
55525 #else
55526 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55527  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55528  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55529  float32x4_t __ret;
55530  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55531  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55532  return __ret;
55533 }
55534 #endif
55535 
55536 #ifdef __LITTLE_ENDIAN__
55537 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55538  float32x2_t __ret;
55539  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55540  return __ret;
55541 }
55542 #else
55543 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55544  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55545  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55546  float32x2_t __ret;
55547  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55548  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55549  return __ret;
55550 }
55551 #endif
55552 
55553 #ifdef __LITTLE_ENDIAN__
55554 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55555  float64_t __ret;
55556  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64(__p0);
55557  return __ret;
55558 }
55559 #else
55560 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55561  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55562  float64_t __ret;
55563  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64(__rev0);
55564  return __ret;
55565 }
55566 #endif
55567 
55568 #ifdef __LITTLE_ENDIAN__
55569 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55570  float32_t __ret;
55571  __ret = (float32_t) __builtin_neon_vpmaxnms_f32(__p0);
55572  return __ret;
55573 }
55574 #else
55575 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55576  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55577  float32_t __ret;
55578  __ret = (float32_t) __builtin_neon_vpmaxnms_f32(__rev0);
55579  return __ret;
55580 }
55581 #endif
55582 
55583 #ifdef __LITTLE_ENDIAN__
55584 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55585  uint8x16_t __ret;
55586  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55587  return __ret;
55588 }
55589 #else
55590 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55591  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55592  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55593  uint8x16_t __ret;
55594  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55595  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55596  return __ret;
55597 }
55598 #endif
55599 
55600 #ifdef __LITTLE_ENDIAN__
55601 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55602  uint32x4_t __ret;
55603  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55604  return __ret;
55605 }
55606 #else
55607 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55608  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55609  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55610  uint32x4_t __ret;
55611  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55612  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55613  return __ret;
55614 }
55615 #endif
55616 
55617 #ifdef __LITTLE_ENDIAN__
55618 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55619  uint16x8_t __ret;
55620  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55621  return __ret;
55622 }
55623 #else
55624 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55625  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55626  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55627  uint16x8_t __ret;
55628  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55629  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55630  return __ret;
55631 }
55632 #endif
55633 
55634 #ifdef __LITTLE_ENDIAN__
55635 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55636  int8x16_t __ret;
55637  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55638  return __ret;
55639 }
55640 #else
55641 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55642  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55643  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55644  int8x16_t __ret;
55645  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55646  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55647  return __ret;
55648 }
55649 #endif
55650 
55651 #ifdef __LITTLE_ENDIAN__
55652 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55653  float64x2_t __ret;
55654  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55655  return __ret;
55656 }
55657 #else
55658 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55659  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55660  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55661  float64x2_t __ret;
55662  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55663  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55664  return __ret;
55665 }
55666 #endif
55667 
55668 #ifdef __LITTLE_ENDIAN__
55669 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55670  float32x4_t __ret;
55671  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55672  return __ret;
55673 }
55674 #else
55675 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55676  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55677  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55678  float32x4_t __ret;
55679  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55681  return __ret;
55682 }
55683 #endif
55684 
55685 #ifdef __LITTLE_ENDIAN__
55686 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55687  int32x4_t __ret;
55688  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55689  return __ret;
55690 }
55691 #else
55692 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55693  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55694  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55695  int32x4_t __ret;
55696  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55698  return __ret;
55699 }
55700 #endif
55701 
55702 #ifdef __LITTLE_ENDIAN__
55703 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55704  int16x8_t __ret;
55705  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55706  return __ret;
55707 }
55708 #else
55709 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55710  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55711  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55712  int16x8_t __ret;
55713  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55714  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55715  return __ret;
55716 }
55717 #endif
55718 
55719 #ifdef __LITTLE_ENDIAN__
55720 __ai float64_t vpminqd_f64(float64x2_t __p0) {
55721  float64_t __ret;
55722  __ret = (float64_t) __builtin_neon_vpminqd_f64(__p0);
55723  return __ret;
55724 }
55725 #else
55726 __ai float64_t vpminqd_f64(float64x2_t __p0) {
55727  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55728  float64_t __ret;
55729  __ret = (float64_t) __builtin_neon_vpminqd_f64(__rev0);
55730  return __ret;
55731 }
55732 #endif
55733 
55734 #ifdef __LITTLE_ENDIAN__
55735 __ai float32_t vpmins_f32(float32x2_t __p0) {
55736  float32_t __ret;
55737  __ret = (float32_t) __builtin_neon_vpmins_f32(__p0);
55738  return __ret;
55739 }
55740 #else
55741 __ai float32_t vpmins_f32(float32x2_t __p0) {
55742  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55743  float32_t __ret;
55744  __ret = (float32_t) __builtin_neon_vpmins_f32(__rev0);
55745  return __ret;
55746 }
55747 #endif
55748 
55749 #ifdef __LITTLE_ENDIAN__
55750 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55751  float64x2_t __ret;
55752  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55753  return __ret;
55754 }
55755 #else
55756 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55757  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55758  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55759  float64x2_t __ret;
55760  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55761  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55762  return __ret;
55763 }
55764 #endif
55765 
55766 #ifdef __LITTLE_ENDIAN__
55767 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55768  float32x4_t __ret;
55769  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55770  return __ret;
55771 }
55772 #else
55773 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55774  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55775  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55776  float32x4_t __ret;
55777  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55778  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55779  return __ret;
55780 }
55781 #endif
55782 
55783 #ifdef __LITTLE_ENDIAN__
55784 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55785  float32x2_t __ret;
55786  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55787  return __ret;
55788 }
55789 #else
55790 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55791  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55792  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55793  float32x2_t __ret;
55794  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55795  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55796  return __ret;
55797 }
55798 #endif
55799 
55800 #ifdef __LITTLE_ENDIAN__
55801 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55802  float64_t __ret;
55803  __ret = (float64_t) __builtin_neon_vpminnmqd_f64(__p0);
55804  return __ret;
55805 }
55806 #else
55807 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55808  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55809  float64_t __ret;
55810  __ret = (float64_t) __builtin_neon_vpminnmqd_f64(__rev0);
55811  return __ret;
55812 }
55813 #endif
55814 
55815 #ifdef __LITTLE_ENDIAN__
55816 __ai float32_t vpminnms_f32(float32x2_t __p0) {
55817  float32_t __ret;
55818  __ret = (float32_t) __builtin_neon_vpminnms_f32(__p0);
55819  return __ret;
55820 }
55821 #else
55822 __ai float32_t vpminnms_f32(float32x2_t __p0) {
55823  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55824  float32_t __ret;
55825  __ret = (float32_t) __builtin_neon_vpminnms_f32(__rev0);
55826  return __ret;
55827 }
55828 #endif
55829 
55830 #ifdef __LITTLE_ENDIAN__
55831 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55832  int64x2_t __ret;
55833  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
55834  return __ret;
55835 }
55836 #else
55837 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55838  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55839  int64x2_t __ret;
55840  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
55841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55842  return __ret;
55843 }
55844 #endif
55845 
55846 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
55847  int64x1_t __ret;
55848  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55849  return __ret;
55850 }
55851 __ai int8_t vqabsb_s8(int8_t __p0) {
55852  int8_t __ret;
55853  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55854  return __ret;
55855 }
55856 __ai int32_t vqabss_s32(int32_t __p0) {
55857  int32_t __ret;
55858  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55859  return __ret;
55860 }
55861 __ai int64_t vqabsd_s64(int64_t __p0) {
55862  int64_t __ret;
55863  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55864  return __ret;
55865 }
55866 __ai int16_t vqabsh_s16(int16_t __p0) {
55867  int16_t __ret;
55868  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55869  return __ret;
55870 }
55871 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55872  uint8_t __ret;
55873  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55874  return __ret;
55875 }
55876 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55877  uint32_t __ret;
55878  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55879  return __ret;
55880 }
55881 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55882  uint64_t __ret;
55883  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55884  return __ret;
55885 }
55886 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55887  uint16_t __ret;
55888  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55889  return __ret;
55890 }
55891 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55892  int8_t __ret;
55893  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55894  return __ret;
55895 }
55896 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55897  int32_t __ret;
55898  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55899  return __ret;
55900 }
55901 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55902  int64_t __ret;
55903  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55904  return __ret;
55905 }
55906 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55907  int16_t __ret;
55908  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55909  return __ret;
55910 }
55911 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55912  int64_t __ret;
55913  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
55914  return __ret;
55915 }
55916 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
55917  int32_t __ret;
55918  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
55919  return __ret;
55920 }
55921 #ifdef __LITTLE_ENDIAN__
55922 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55923  int64x2_t __ret;
55924  __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
55925  return __ret;
55926 }
55927 #else
55928 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55929  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55930  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55931  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
55932  int64x2_t __ret;
55933  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
55934  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55935  return __ret;
55936 }
55937 #endif
55938 
55939 #ifdef __LITTLE_ENDIAN__
55940 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55941  int32x4_t __ret;
55942  __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
55943  return __ret;
55944 }
55945 #else
55946 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55947  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55948  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55949  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55950  int32x4_t __ret;
55951  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
55952  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55953  return __ret;
55954 }
55955 #endif
55956 
55957 #ifdef __LITTLE_ENDIAN__
55958 #define vqdmlal_high_lane_s32(__p0_603, __p1_603, __p2_603, __p3_603) __extension__ ({ \
55959  int64x2_t __s0_603 = __p0_603; \
55960  int32x4_t __s1_603 = __p1_603; \
55961  int32x2_t __s2_603 = __p2_603; \
55962  int64x2_t __ret_603; \
55963  __ret_603 = vqdmlal_s32(__s0_603, vget_high_s32(__s1_603), splat_lane_s32(__s2_603, __p3_603)); \
55964  __ret_603; \
55965 })
55966 #else
55967 #define vqdmlal_high_lane_s32(__p0_604, __p1_604, __p2_604, __p3_604) __extension__ ({ \
55968  int64x2_t __s0_604 = __p0_604; \
55969  int32x4_t __s1_604 = __p1_604; \
55970  int32x2_t __s2_604 = __p2_604; \
55971  int64x2_t __rev0_604; __rev0_604 = __builtin_shufflevector(__s0_604, __s0_604, 1, 0); \
55972  int32x4_t __rev1_604; __rev1_604 = __builtin_shufflevector(__s1_604, __s1_604, 3, 2, 1, 0); \
55973  int32x2_t __rev2_604; __rev2_604 = __builtin_shufflevector(__s2_604, __s2_604, 1, 0); \
55974  int64x2_t __ret_604; \
55975  __ret_604 = __noswap_vqdmlal_s32(__rev0_604, __noswap_vget_high_s32(__rev1_604), __noswap_splat_lane_s32(__rev2_604, __p3_604)); \
55976  __ret_604 = __builtin_shufflevector(__ret_604, __ret_604, 1, 0); \
55977  __ret_604; \
55978 })
55979 #endif
55980 
55981 #ifdef __LITTLE_ENDIAN__
55982 #define vqdmlal_high_lane_s16(__p0_605, __p1_605, __p2_605, __p3_605) __extension__ ({ \
55983  int32x4_t __s0_605 = __p0_605; \
55984  int16x8_t __s1_605 = __p1_605; \
55985  int16x4_t __s2_605 = __p2_605; \
55986  int32x4_t __ret_605; \
55987  __ret_605 = vqdmlal_s16(__s0_605, vget_high_s16(__s1_605), splat_lane_s16(__s2_605, __p3_605)); \
55988  __ret_605; \
55989 })
55990 #else
55991 #define vqdmlal_high_lane_s16(__p0_606, __p1_606, __p2_606, __p3_606) __extension__ ({ \
55992  int32x4_t __s0_606 = __p0_606; \
55993  int16x8_t __s1_606 = __p1_606; \
55994  int16x4_t __s2_606 = __p2_606; \
55995  int32x4_t __rev0_606; __rev0_606 = __builtin_shufflevector(__s0_606, __s0_606, 3, 2, 1, 0); \
55996  int16x8_t __rev1_606; __rev1_606 = __builtin_shufflevector(__s1_606, __s1_606, 7, 6, 5, 4, 3, 2, 1, 0); \
55997  int16x4_t __rev2_606; __rev2_606 = __builtin_shufflevector(__s2_606, __s2_606, 3, 2, 1, 0); \
55998  int32x4_t __ret_606; \
55999  __ret_606 = __noswap_vqdmlal_s16(__rev0_606, __noswap_vget_high_s16(__rev1_606), __noswap_splat_lane_s16(__rev2_606, __p3_606)); \
56000  __ret_606 = __builtin_shufflevector(__ret_606, __ret_606, 3, 2, 1, 0); \
56001  __ret_606; \
56002 })
56003 #endif
56004 
56005 #ifdef __LITTLE_ENDIAN__
56006 #define vqdmlal_high_laneq_s32(__p0_607, __p1_607, __p2_607, __p3_607) __extension__ ({ \
56007  int64x2_t __s0_607 = __p0_607; \
56008  int32x4_t __s1_607 = __p1_607; \
56009  int32x4_t __s2_607 = __p2_607; \
56010  int64x2_t __ret_607; \
56011  __ret_607 = vqdmlal_s32(__s0_607, vget_high_s32(__s1_607), splat_laneq_s32(__s2_607, __p3_607)); \
56012  __ret_607; \
56013 })
56014 #else
56015 #define vqdmlal_high_laneq_s32(__p0_608, __p1_608, __p2_608, __p3_608) __extension__ ({ \
56016  int64x2_t __s0_608 = __p0_608; \
56017  int32x4_t __s1_608 = __p1_608; \
56018  int32x4_t __s2_608 = __p2_608; \
56019  int64x2_t __rev0_608; __rev0_608 = __builtin_shufflevector(__s0_608, __s0_608, 1, 0); \
56020  int32x4_t __rev1_608; __rev1_608 = __builtin_shufflevector(__s1_608, __s1_608, 3, 2, 1, 0); \
56021  int32x4_t __rev2_608; __rev2_608 = __builtin_shufflevector(__s2_608, __s2_608, 3, 2, 1, 0); \
56022  int64x2_t __ret_608; \
56023  __ret_608 = __noswap_vqdmlal_s32(__rev0_608, __noswap_vget_high_s32(__rev1_608), __noswap_splat_laneq_s32(__rev2_608, __p3_608)); \
56024  __ret_608 = __builtin_shufflevector(__ret_608, __ret_608, 1, 0); \
56025  __ret_608; \
56026 })
56027 #endif
56028 
56029 #ifdef __LITTLE_ENDIAN__
56030 #define vqdmlal_high_laneq_s16(__p0_609, __p1_609, __p2_609, __p3_609) __extension__ ({ \
56031  int32x4_t __s0_609 = __p0_609; \
56032  int16x8_t __s1_609 = __p1_609; \
56033  int16x8_t __s2_609 = __p2_609; \
56034  int32x4_t __ret_609; \
56035  __ret_609 = vqdmlal_s16(__s0_609, vget_high_s16(__s1_609), splat_laneq_s16(__s2_609, __p3_609)); \
56036  __ret_609; \
56037 })
56038 #else
56039 #define vqdmlal_high_laneq_s16(__p0_610, __p1_610, __p2_610, __p3_610) __extension__ ({ \
56040  int32x4_t __s0_610 = __p0_610; \
56041  int16x8_t __s1_610 = __p1_610; \
56042  int16x8_t __s2_610 = __p2_610; \
56043  int32x4_t __rev0_610; __rev0_610 = __builtin_shufflevector(__s0_610, __s0_610, 3, 2, 1, 0); \
56044  int16x8_t __rev1_610; __rev1_610 = __builtin_shufflevector(__s1_610, __s1_610, 7, 6, 5, 4, 3, 2, 1, 0); \
56045  int16x8_t __rev2_610; __rev2_610 = __builtin_shufflevector(__s2_610, __s2_610, 7, 6, 5, 4, 3, 2, 1, 0); \
56046  int32x4_t __ret_610; \
56047  __ret_610 = __noswap_vqdmlal_s16(__rev0_610, __noswap_vget_high_s16(__rev1_610), __noswap_splat_laneq_s16(__rev2_610, __p3_610)); \
56048  __ret_610 = __builtin_shufflevector(__ret_610, __ret_610, 3, 2, 1, 0); \
56049  __ret_610; \
56050 })
56051 #endif
56052 
56053 #ifdef __LITTLE_ENDIAN__
56054 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56055  int64x2_t __ret;
56056  __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
56057  return __ret;
56058 }
56059 #else
56060 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56061  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56062  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56063  int64x2_t __ret;
56064  __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56065  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56066  return __ret;
56067 }
56068 #endif
56069 
56070 #ifdef __LITTLE_ENDIAN__
56071 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56072  int32x4_t __ret;
56073  __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
56074  return __ret;
56075 }
56076 #else
56077 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56078  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56079  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56080  int32x4_t __ret;
56081  __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56082  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56083  return __ret;
56084 }
56085 #endif
56086 
56087 #ifdef __LITTLE_ENDIAN__
56088 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56089  int64_t __s0 = __p0; \
56090  int32_t __s1 = __p1; \
56091  int32x2_t __s2 = __p2; \
56092  int64_t __ret; \
56093  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, __s2, __p3); \
56094  __ret; \
56095 })
56096 #else
56097 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56098  int64_t __s0 = __p0; \
56099  int32_t __s1 = __p1; \
56100  int32x2_t __s2 = __p2; \
56101  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56102  int64_t __ret; \
56103  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, __rev2, __p3); \
56104  __ret; \
56105 })
56106 #endif
56107 
56108 #ifdef __LITTLE_ENDIAN__
56109 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56110  int32_t __s0 = __p0; \
56111  int16_t __s1 = __p1; \
56112  int16x4_t __s2 = __p2; \
56113  int32_t __ret; \
56114  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, __s2, __p3); \
56115  __ret; \
56116 })
56117 #else
56118 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56119  int32_t __s0 = __p0; \
56120  int16_t __s1 = __p1; \
56121  int16x4_t __s2 = __p2; \
56122  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56123  int32_t __ret; \
56124  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, __rev2, __p3); \
56125  __ret; \
56126 })
56127 #endif
56128 
56129 #ifdef __LITTLE_ENDIAN__
56130 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56131  int64_t __s0 = __p0; \
56132  int32_t __s1 = __p1; \
56133  int32x4_t __s2 = __p2; \
56134  int64_t __ret; \
56135  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, __s2, __p3); \
56136  __ret; \
56137 })
56138 #else
56139 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56140  int64_t __s0 = __p0; \
56141  int32_t __s1 = __p1; \
56142  int32x4_t __s2 = __p2; \
56143  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56144  int64_t __ret; \
56145  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, __rev2, __p3); \
56146  __ret; \
56147 })
56148 #endif
56149 
56150 #ifdef __LITTLE_ENDIAN__
56151 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56152  int32_t __s0 = __p0; \
56153  int16_t __s1 = __p1; \
56154  int16x8_t __s2 = __p2; \
56155  int32_t __ret; \
56156  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, __s2, __p3); \
56157  __ret; \
56158 })
56159 #else
56160 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56161  int32_t __s0 = __p0; \
56162  int16_t __s1 = __p1; \
56163  int16x8_t __s2 = __p2; \
56164  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56165  int32_t __ret; \
56166  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, __rev2, __p3); \
56167  __ret; \
56168 })
56169 #endif
56170 
56171 #ifdef __LITTLE_ENDIAN__
56172 #define vqdmlal_laneq_s32(__p0_611, __p1_611, __p2_611, __p3_611) __extension__ ({ \
56173  int64x2_t __s0_611 = __p0_611; \
56174  int32x2_t __s1_611 = __p1_611; \
56175  int32x4_t __s2_611 = __p2_611; \
56176  int64x2_t __ret_611; \
56177  __ret_611 = vqdmlal_s32(__s0_611, __s1_611, splat_laneq_s32(__s2_611, __p3_611)); \
56178  __ret_611; \
56179 })
56180 #else
56181 #define vqdmlal_laneq_s32(__p0_612, __p1_612, __p2_612, __p3_612) __extension__ ({ \
56182  int64x2_t __s0_612 = __p0_612; \
56183  int32x2_t __s1_612 = __p1_612; \
56184  int32x4_t __s2_612 = __p2_612; \
56185  int64x2_t __rev0_612; __rev0_612 = __builtin_shufflevector(__s0_612, __s0_612, 1, 0); \
56186  int32x2_t __rev1_612; __rev1_612 = __builtin_shufflevector(__s1_612, __s1_612, 1, 0); \
56187  int32x4_t __rev2_612; __rev2_612 = __builtin_shufflevector(__s2_612, __s2_612, 3, 2, 1, 0); \
56188  int64x2_t __ret_612; \
56189  __ret_612 = __noswap_vqdmlal_s32(__rev0_612, __rev1_612, __noswap_splat_laneq_s32(__rev2_612, __p3_612)); \
56190  __ret_612 = __builtin_shufflevector(__ret_612, __ret_612, 1, 0); \
56191  __ret_612; \
56192 })
56193 #endif
56194 
56195 #ifdef __LITTLE_ENDIAN__
56196 #define vqdmlal_laneq_s16(__p0_613, __p1_613, __p2_613, __p3_613) __extension__ ({ \
56197  int32x4_t __s0_613 = __p0_613; \
56198  int16x4_t __s1_613 = __p1_613; \
56199  int16x8_t __s2_613 = __p2_613; \
56200  int32x4_t __ret_613; \
56201  __ret_613 = vqdmlal_s16(__s0_613, __s1_613, splat_laneq_s16(__s2_613, __p3_613)); \
56202  __ret_613; \
56203 })
56204 #else
56205 #define vqdmlal_laneq_s16(__p0_614, __p1_614, __p2_614, __p3_614) __extension__ ({ \
56206  int32x4_t __s0_614 = __p0_614; \
56207  int16x4_t __s1_614 = __p1_614; \
56208  int16x8_t __s2_614 = __p2_614; \
56209  int32x4_t __rev0_614; __rev0_614 = __builtin_shufflevector(__s0_614, __s0_614, 3, 2, 1, 0); \
56210  int16x4_t __rev1_614; __rev1_614 = __builtin_shufflevector(__s1_614, __s1_614, 3, 2, 1, 0); \
56211  int16x8_t __rev2_614; __rev2_614 = __builtin_shufflevector(__s2_614, __s2_614, 7, 6, 5, 4, 3, 2, 1, 0); \
56212  int32x4_t __ret_614; \
56213  __ret_614 = __noswap_vqdmlal_s16(__rev0_614, __rev1_614, __noswap_splat_laneq_s16(__rev2_614, __p3_614)); \
56214  __ret_614 = __builtin_shufflevector(__ret_614, __ret_614, 3, 2, 1, 0); \
56215  __ret_614; \
56216 })
56217 #endif
56218 
56219 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56220  int64_t __ret;
56221  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56222  return __ret;
56223 }
56224 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56225  int32_t __ret;
56226  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56227  return __ret;
56228 }
56229 #ifdef __LITTLE_ENDIAN__
56230 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56231  int64x2_t __ret;
56232  __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56233  return __ret;
56234 }
56235 #else
56236 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56237  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56238  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56239  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56240  int64x2_t __ret;
56241  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56243  return __ret;
56244 }
56245 #endif
56246 
56247 #ifdef __LITTLE_ENDIAN__
56248 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56249  int32x4_t __ret;
56250  __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56251  return __ret;
56252 }
56253 #else
56254 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56255  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56256  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56257  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56258  int32x4_t __ret;
56259  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56260  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56261  return __ret;
56262 }
56263 #endif
56264 
56265 #ifdef __LITTLE_ENDIAN__
56266 #define vqdmlsl_high_lane_s32(__p0_615, __p1_615, __p2_615, __p3_615) __extension__ ({ \
56267  int64x2_t __s0_615 = __p0_615; \
56268  int32x4_t __s1_615 = __p1_615; \
56269  int32x2_t __s2_615 = __p2_615; \
56270  int64x2_t __ret_615; \
56271  __ret_615 = vqdmlsl_s32(__s0_615, vget_high_s32(__s1_615), splat_lane_s32(__s2_615, __p3_615)); \
56272  __ret_615; \
56273 })
56274 #else
56275 #define vqdmlsl_high_lane_s32(__p0_616, __p1_616, __p2_616, __p3_616) __extension__ ({ \
56276  int64x2_t __s0_616 = __p0_616; \
56277  int32x4_t __s1_616 = __p1_616; \
56278  int32x2_t __s2_616 = __p2_616; \
56279  int64x2_t __rev0_616; __rev0_616 = __builtin_shufflevector(__s0_616, __s0_616, 1, 0); \
56280  int32x4_t __rev1_616; __rev1_616 = __builtin_shufflevector(__s1_616, __s1_616, 3, 2, 1, 0); \
56281  int32x2_t __rev2_616; __rev2_616 = __builtin_shufflevector(__s2_616, __s2_616, 1, 0); \
56282  int64x2_t __ret_616; \
56283  __ret_616 = __noswap_vqdmlsl_s32(__rev0_616, __noswap_vget_high_s32(__rev1_616), __noswap_splat_lane_s32(__rev2_616, __p3_616)); \
56284  __ret_616 = __builtin_shufflevector(__ret_616, __ret_616, 1, 0); \
56285  __ret_616; \
56286 })
56287 #endif
56288 
56289 #ifdef __LITTLE_ENDIAN__
56290 #define vqdmlsl_high_lane_s16(__p0_617, __p1_617, __p2_617, __p3_617) __extension__ ({ \
56291  int32x4_t __s0_617 = __p0_617; \
56292  int16x8_t __s1_617 = __p1_617; \
56293  int16x4_t __s2_617 = __p2_617; \
56294  int32x4_t __ret_617; \
56295  __ret_617 = vqdmlsl_s16(__s0_617, vget_high_s16(__s1_617), splat_lane_s16(__s2_617, __p3_617)); \
56296  __ret_617; \
56297 })
56298 #else
56299 #define vqdmlsl_high_lane_s16(__p0_618, __p1_618, __p2_618, __p3_618) __extension__ ({ \
56300  int32x4_t __s0_618 = __p0_618; \
56301  int16x8_t __s1_618 = __p1_618; \
56302  int16x4_t __s2_618 = __p2_618; \
56303  int32x4_t __rev0_618; __rev0_618 = __builtin_shufflevector(__s0_618, __s0_618, 3, 2, 1, 0); \
56304  int16x8_t __rev1_618; __rev1_618 = __builtin_shufflevector(__s1_618, __s1_618, 7, 6, 5, 4, 3, 2, 1, 0); \
56305  int16x4_t __rev2_618; __rev2_618 = __builtin_shufflevector(__s2_618, __s2_618, 3, 2, 1, 0); \
56306  int32x4_t __ret_618; \
56307  __ret_618 = __noswap_vqdmlsl_s16(__rev0_618, __noswap_vget_high_s16(__rev1_618), __noswap_splat_lane_s16(__rev2_618, __p3_618)); \
56308  __ret_618 = __builtin_shufflevector(__ret_618, __ret_618, 3, 2, 1, 0); \
56309  __ret_618; \
56310 })
56311 #endif
56312 
56313 #ifdef __LITTLE_ENDIAN__
56314 #define vqdmlsl_high_laneq_s32(__p0_619, __p1_619, __p2_619, __p3_619) __extension__ ({ \
56315  int64x2_t __s0_619 = __p0_619; \
56316  int32x4_t __s1_619 = __p1_619; \
56317  int32x4_t __s2_619 = __p2_619; \
56318  int64x2_t __ret_619; \
56319  __ret_619 = vqdmlsl_s32(__s0_619, vget_high_s32(__s1_619), splat_laneq_s32(__s2_619, __p3_619)); \
56320  __ret_619; \
56321 })
56322 #else
56323 #define vqdmlsl_high_laneq_s32(__p0_620, __p1_620, __p2_620, __p3_620) __extension__ ({ \
56324  int64x2_t __s0_620 = __p0_620; \
56325  int32x4_t __s1_620 = __p1_620; \
56326  int32x4_t __s2_620 = __p2_620; \
56327  int64x2_t __rev0_620; __rev0_620 = __builtin_shufflevector(__s0_620, __s0_620, 1, 0); \
56328  int32x4_t __rev1_620; __rev1_620 = __builtin_shufflevector(__s1_620, __s1_620, 3, 2, 1, 0); \
56329  int32x4_t __rev2_620; __rev2_620 = __builtin_shufflevector(__s2_620, __s2_620, 3, 2, 1, 0); \
56330  int64x2_t __ret_620; \
56331  __ret_620 = __noswap_vqdmlsl_s32(__rev0_620, __noswap_vget_high_s32(__rev1_620), __noswap_splat_laneq_s32(__rev2_620, __p3_620)); \
56332  __ret_620 = __builtin_shufflevector(__ret_620, __ret_620, 1, 0); \
56333  __ret_620; \
56334 })
56335 #endif
56336 
56337 #ifdef __LITTLE_ENDIAN__
56338 #define vqdmlsl_high_laneq_s16(__p0_621, __p1_621, __p2_621, __p3_621) __extension__ ({ \
56339  int32x4_t __s0_621 = __p0_621; \
56340  int16x8_t __s1_621 = __p1_621; \
56341  int16x8_t __s2_621 = __p2_621; \
56342  int32x4_t __ret_621; \
56343  __ret_621 = vqdmlsl_s16(__s0_621, vget_high_s16(__s1_621), splat_laneq_s16(__s2_621, __p3_621)); \
56344  __ret_621; \
56345 })
56346 #else
56347 #define vqdmlsl_high_laneq_s16(__p0_622, __p1_622, __p2_622, __p3_622) __extension__ ({ \
56348  int32x4_t __s0_622 = __p0_622; \
56349  int16x8_t __s1_622 = __p1_622; \
56350  int16x8_t __s2_622 = __p2_622; \
56351  int32x4_t __rev0_622; __rev0_622 = __builtin_shufflevector(__s0_622, __s0_622, 3, 2, 1, 0); \
56352  int16x8_t __rev1_622; __rev1_622 = __builtin_shufflevector(__s1_622, __s1_622, 7, 6, 5, 4, 3, 2, 1, 0); \
56353  int16x8_t __rev2_622; __rev2_622 = __builtin_shufflevector(__s2_622, __s2_622, 7, 6, 5, 4, 3, 2, 1, 0); \
56354  int32x4_t __ret_622; \
56355  __ret_622 = __noswap_vqdmlsl_s16(__rev0_622, __noswap_vget_high_s16(__rev1_622), __noswap_splat_laneq_s16(__rev2_622, __p3_622)); \
56356  __ret_622 = __builtin_shufflevector(__ret_622, __ret_622, 3, 2, 1, 0); \
56357  __ret_622; \
56358 })
56359 #endif
56360 
56361 #ifdef __LITTLE_ENDIAN__
56362 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56363  int64x2_t __ret;
56364  __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
56365  return __ret;
56366 }
56367 #else
56368 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56369  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56370  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56371  int64x2_t __ret;
56372  __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56373  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56374  return __ret;
56375 }
56376 #endif
56377 
56378 #ifdef __LITTLE_ENDIAN__
56379 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56380  int32x4_t __ret;
56381  __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
56382  return __ret;
56383 }
56384 #else
56385 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56386  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56387  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56388  int32x4_t __ret;
56389  __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56391  return __ret;
56392 }
56393 #endif
56394 
56395 #ifdef __LITTLE_ENDIAN__
56396 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56397  int64_t __s0 = __p0; \
56398  int32_t __s1 = __p1; \
56399  int32x2_t __s2 = __p2; \
56400  int64_t __ret; \
56401  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, __s2, __p3); \
56402  __ret; \
56403 })
56404 #else
56405 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56406  int64_t __s0 = __p0; \
56407  int32_t __s1 = __p1; \
56408  int32x2_t __s2 = __p2; \
56409  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56410  int64_t __ret; \
56411  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, __rev2, __p3); \
56412  __ret; \
56413 })
56414 #endif
56415 
56416 #ifdef __LITTLE_ENDIAN__
56417 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56418  int32_t __s0 = __p0; \
56419  int16_t __s1 = __p1; \
56420  int16x4_t __s2 = __p2; \
56421  int32_t __ret; \
56422  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, __s2, __p3); \
56423  __ret; \
56424 })
56425 #else
56426 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56427  int32_t __s0 = __p0; \
56428  int16_t __s1 = __p1; \
56429  int16x4_t __s2 = __p2; \
56430  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56431  int32_t __ret; \
56432  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, __rev2, __p3); \
56433  __ret; \
56434 })
56435 #endif
56436 
56437 #ifdef __LITTLE_ENDIAN__
56438 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56439  int64_t __s0 = __p0; \
56440  int32_t __s1 = __p1; \
56441  int32x4_t __s2 = __p2; \
56442  int64_t __ret; \
56443  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, __s2, __p3); \
56444  __ret; \
56445 })
56446 #else
56447 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56448  int64_t __s0 = __p0; \
56449  int32_t __s1 = __p1; \
56450  int32x4_t __s2 = __p2; \
56451  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56452  int64_t __ret; \
56453  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, __rev2, __p3); \
56454  __ret; \
56455 })
56456 #endif
56457 
56458 #ifdef __LITTLE_ENDIAN__
56459 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56460  int32_t __s0 = __p0; \
56461  int16_t __s1 = __p1; \
56462  int16x8_t __s2 = __p2; \
56463  int32_t __ret; \
56464  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, __s2, __p3); \
56465  __ret; \
56466 })
56467 #else
56468 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56469  int32_t __s0 = __p0; \
56470  int16_t __s1 = __p1; \
56471  int16x8_t __s2 = __p2; \
56472  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56473  int32_t __ret; \
56474  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, __rev2, __p3); \
56475  __ret; \
56476 })
56477 #endif
56478 
56479 #ifdef __LITTLE_ENDIAN__
56480 #define vqdmlsl_laneq_s32(__p0_623, __p1_623, __p2_623, __p3_623) __extension__ ({ \
56481  int64x2_t __s0_623 = __p0_623; \
56482  int32x2_t __s1_623 = __p1_623; \
56483  int32x4_t __s2_623 = __p2_623; \
56484  int64x2_t __ret_623; \
56485  __ret_623 = vqdmlsl_s32(__s0_623, __s1_623, splat_laneq_s32(__s2_623, __p3_623)); \
56486  __ret_623; \
56487 })
56488 #else
56489 #define vqdmlsl_laneq_s32(__p0_624, __p1_624, __p2_624, __p3_624) __extension__ ({ \
56490  int64x2_t __s0_624 = __p0_624; \
56491  int32x2_t __s1_624 = __p1_624; \
56492  int32x4_t __s2_624 = __p2_624; \
56493  int64x2_t __rev0_624; __rev0_624 = __builtin_shufflevector(__s0_624, __s0_624, 1, 0); \
56494  int32x2_t __rev1_624; __rev1_624 = __builtin_shufflevector(__s1_624, __s1_624, 1, 0); \
56495  int32x4_t __rev2_624; __rev2_624 = __builtin_shufflevector(__s2_624, __s2_624, 3, 2, 1, 0); \
56496  int64x2_t __ret_624; \
56497  __ret_624 = __noswap_vqdmlsl_s32(__rev0_624, __rev1_624, __noswap_splat_laneq_s32(__rev2_624, __p3_624)); \
56498  __ret_624 = __builtin_shufflevector(__ret_624, __ret_624, 1, 0); \
56499  __ret_624; \
56500 })
56501 #endif
56502 
56503 #ifdef __LITTLE_ENDIAN__
56504 #define vqdmlsl_laneq_s16(__p0_625, __p1_625, __p2_625, __p3_625) __extension__ ({ \
56505  int32x4_t __s0_625 = __p0_625; \
56506  int16x4_t __s1_625 = __p1_625; \
56507  int16x8_t __s2_625 = __p2_625; \
56508  int32x4_t __ret_625; \
56509  __ret_625 = vqdmlsl_s16(__s0_625, __s1_625, splat_laneq_s16(__s2_625, __p3_625)); \
56510  __ret_625; \
56511 })
56512 #else
56513 #define vqdmlsl_laneq_s16(__p0_626, __p1_626, __p2_626, __p3_626) __extension__ ({ \
56514  int32x4_t __s0_626 = __p0_626; \
56515  int16x4_t __s1_626 = __p1_626; \
56516  int16x8_t __s2_626 = __p2_626; \
56517  int32x4_t __rev0_626; __rev0_626 = __builtin_shufflevector(__s0_626, __s0_626, 3, 2, 1, 0); \
56518  int16x4_t __rev1_626; __rev1_626 = __builtin_shufflevector(__s1_626, __s1_626, 3, 2, 1, 0); \
56519  int16x8_t __rev2_626; __rev2_626 = __builtin_shufflevector(__s2_626, __s2_626, 7, 6, 5, 4, 3, 2, 1, 0); \
56520  int32x4_t __ret_626; \
56521  __ret_626 = __noswap_vqdmlsl_s16(__rev0_626, __rev1_626, __noswap_splat_laneq_s16(__rev2_626, __p3_626)); \
56522  __ret_626 = __builtin_shufflevector(__ret_626, __ret_626, 3, 2, 1, 0); \
56523  __ret_626; \
56524 })
56525 #endif
56526 
56527 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56528  int32_t __ret;
56529  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56530  return __ret;
56531 }
56532 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56533  int16_t __ret;
56534  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56535  return __ret;
56536 }
56537 #ifdef __LITTLE_ENDIAN__
56538 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56539  int32x4_t __s0 = __p0; \
56540  int32x2_t __s1 = __p1; \
56541  int32x4_t __ret; \
56542  __ret = (int32x4_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 2); \
56543  __ret; \
56544 })
56545 #else
56546 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56547  int32x4_t __s0 = __p0; \
56548  int32x2_t __s1 = __p1; \
56549  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56550  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56551  int32x4_t __ret; \
56552  __ret = (int32x4_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
56553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56554  __ret; \
56555 })
56556 #endif
56557 
56558 #ifdef __LITTLE_ENDIAN__
56559 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56560  int16x8_t __s0 = __p0; \
56561  int16x4_t __s1 = __p1; \
56562  int16x8_t __ret; \
56563  __ret = (int16x8_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 1); \
56564  __ret; \
56565 })
56566 #else
56567 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56568  int16x8_t __s0 = __p0; \
56569  int16x4_t __s1 = __p1; \
56570  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56571  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56572  int16x8_t __ret; \
56573  __ret = (int16x8_t) __builtin_neon_vqdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
56574  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56575  __ret; \
56576 })
56577 #endif
56578 
56579 #ifdef __LITTLE_ENDIAN__
56580 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56581  int32x2_t __s0 = __p0; \
56582  int32x2_t __s1 = __p1; \
56583  int32x2_t __ret; \
56584  __ret = (int32x2_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
56585  __ret; \
56586 })
56587 #else
56588 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56589  int32x2_t __s0 = __p0; \
56590  int32x2_t __s1 = __p1; \
56591  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56592  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56593  int32x2_t __ret; \
56594  __ret = (int32x2_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
56595  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56596  __ret; \
56597 })
56598 #endif
56599 
56600 #ifdef __LITTLE_ENDIAN__
56601 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56602  int16x4_t __s0 = __p0; \
56603  int16x4_t __s1 = __p1; \
56604  int16x4_t __ret; \
56605  __ret = (int16x4_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
56606  __ret; \
56607 })
56608 #else
56609 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56610  int16x4_t __s0 = __p0; \
56611  int16x4_t __s1 = __p1; \
56612  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56613  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56614  int16x4_t __ret; \
56615  __ret = (int16x4_t) __builtin_neon_vqdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
56616  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56617  __ret; \
56618 })
56619 #endif
56620 
56621 #ifdef __LITTLE_ENDIAN__
56622 #define vqdmulhs_lane_s32(__p0_627, __p1_627, __p2_627) __extension__ ({ \
56623  int32_t __s0_627 = __p0_627; \
56624  int32x2_t __s1_627 = __p1_627; \
56625  int32_t __ret_627; \
56626  __ret_627 = vqdmulhs_s32(__s0_627, vget_lane_s32(__s1_627, __p2_627)); \
56627  __ret_627; \
56628 })
56629 #else
56630 #define vqdmulhs_lane_s32(__p0_628, __p1_628, __p2_628) __extension__ ({ \
56631  int32_t __s0_628 = __p0_628; \
56632  int32x2_t __s1_628 = __p1_628; \
56633  int32x2_t __rev1_628; __rev1_628 = __builtin_shufflevector(__s1_628, __s1_628, 1, 0); \
56634  int32_t __ret_628; \
56635  __ret_628 = vqdmulhs_s32(__s0_628, __noswap_vget_lane_s32(__rev1_628, __p2_628)); \
56636  __ret_628; \
56637 })
56638 #endif
56639 
56640 #ifdef __LITTLE_ENDIAN__
56641 #define vqdmulhh_lane_s16(__p0_629, __p1_629, __p2_629) __extension__ ({ \
56642  int16_t __s0_629 = __p0_629; \
56643  int16x4_t __s1_629 = __p1_629; \
56644  int16_t __ret_629; \
56645  __ret_629 = vqdmulhh_s16(__s0_629, vget_lane_s16(__s1_629, __p2_629)); \
56646  __ret_629; \
56647 })
56648 #else
56649 #define vqdmulhh_lane_s16(__p0_630, __p1_630, __p2_630) __extension__ ({ \
56650  int16_t __s0_630 = __p0_630; \
56651  int16x4_t __s1_630 = __p1_630; \
56652  int16x4_t __rev1_630; __rev1_630 = __builtin_shufflevector(__s1_630, __s1_630, 3, 2, 1, 0); \
56653  int16_t __ret_630; \
56654  __ret_630 = vqdmulhh_s16(__s0_630, __noswap_vget_lane_s16(__rev1_630, __p2_630)); \
56655  __ret_630; \
56656 })
56657 #endif
56658 
56659 #ifdef __LITTLE_ENDIAN__
56660 #define vqdmulhs_laneq_s32(__p0_631, __p1_631, __p2_631) __extension__ ({ \
56661  int32_t __s0_631 = __p0_631; \
56662  int32x4_t __s1_631 = __p1_631; \
56663  int32_t __ret_631; \
56664  __ret_631 = vqdmulhs_s32(__s0_631, vgetq_lane_s32(__s1_631, __p2_631)); \
56665  __ret_631; \
56666 })
56667 #else
56668 #define vqdmulhs_laneq_s32(__p0_632, __p1_632, __p2_632) __extension__ ({ \
56669  int32_t __s0_632 = __p0_632; \
56670  int32x4_t __s1_632 = __p1_632; \
56671  int32x4_t __rev1_632; __rev1_632 = __builtin_shufflevector(__s1_632, __s1_632, 3, 2, 1, 0); \
56672  int32_t __ret_632; \
56673  __ret_632 = vqdmulhs_s32(__s0_632, __noswap_vgetq_lane_s32(__rev1_632, __p2_632)); \
56674  __ret_632; \
56675 })
56676 #endif
56677 
56678 #ifdef __LITTLE_ENDIAN__
56679 #define vqdmulhh_laneq_s16(__p0_633, __p1_633, __p2_633) __extension__ ({ \
56680  int16_t __s0_633 = __p0_633; \
56681  int16x8_t __s1_633 = __p1_633; \
56682  int16_t __ret_633; \
56683  __ret_633 = vqdmulhh_s16(__s0_633, vgetq_lane_s16(__s1_633, __p2_633)); \
56684  __ret_633; \
56685 })
56686 #else
56687 #define vqdmulhh_laneq_s16(__p0_634, __p1_634, __p2_634) __extension__ ({ \
56688  int16_t __s0_634 = __p0_634; \
56689  int16x8_t __s1_634 = __p1_634; \
56690  int16x8_t __rev1_634; __rev1_634 = __builtin_shufflevector(__s1_634, __s1_634, 7, 6, 5, 4, 3, 2, 1, 0); \
56691  int16_t __ret_634; \
56692  __ret_634 = vqdmulhh_s16(__s0_634, __noswap_vgetq_lane_s16(__rev1_634, __p2_634)); \
56693  __ret_634; \
56694 })
56695 #endif
56696 
56697 #ifdef __LITTLE_ENDIAN__
56698 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56699  int32x4_t __s0 = __p0; \
56700  int32x4_t __s1 = __p1; \
56701  int32x4_t __ret; \
56702  __ret = (int32x4_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
56703  __ret; \
56704 })
56705 #else
56706 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56707  int32x4_t __s0 = __p0; \
56708  int32x4_t __s1 = __p1; \
56709  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56710  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56711  int32x4_t __ret; \
56712  __ret = (int32x4_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
56713  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56714  __ret; \
56715 })
56716 #endif
56717 
56718 #ifdef __LITTLE_ENDIAN__
56719 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56720  int16x8_t __s0 = __p0; \
56721  int16x8_t __s1 = __p1; \
56722  int16x8_t __ret; \
56723  __ret = (int16x8_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
56724  __ret; \
56725 })
56726 #else
56727 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56728  int16x8_t __s0 = __p0; \
56729  int16x8_t __s1 = __p1; \
56730  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56731  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56732  int16x8_t __ret; \
56733  __ret = (int16x8_t) __builtin_neon_vqdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
56734  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56735  __ret; \
56736 })
56737 #endif
56738 
56739 #ifdef __LITTLE_ENDIAN__
56740 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56741  int32x2_t __s0 = __p0; \
56742  int32x4_t __s1 = __p1; \
56743  int32x2_t __ret; \
56744  __ret = (int32x2_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 2); \
56745  __ret; \
56746 })
56747 #else
56748 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56749  int32x2_t __s0 = __p0; \
56750  int32x4_t __s1 = __p1; \
56751  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56752  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56753  int32x2_t __ret; \
56754  __ret = (int32x2_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 2); \
56755  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56756  __ret; \
56757 })
56758 #endif
56759 
56760 #ifdef __LITTLE_ENDIAN__
56761 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56762  int16x4_t __s0 = __p0; \
56763  int16x8_t __s1 = __p1; \
56764  int16x4_t __ret; \
56765  __ret = (int16x4_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 1); \
56766  __ret; \
56767 })
56768 #else
56769 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56770  int16x4_t __s0 = __p0; \
56771  int16x8_t __s1 = __p1; \
56772  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56773  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56774  int16x4_t __ret; \
56775  __ret = (int16x4_t) __builtin_neon_vqdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 1); \
56776  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56777  __ret; \
56778 })
56779 #endif
56780 
56781 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56782  int64_t __ret;
56783  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56784  return __ret;
56785 }
56786 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56787  int32_t __ret;
56788  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56789  return __ret;
56790 }
56791 #ifdef __LITTLE_ENDIAN__
56792 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56793  int64x2_t __ret;
56794  __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
56795  return __ret;
56796 }
56797 #else
56798 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56799  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56800  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56801  int64x2_t __ret;
56802  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
56803  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56804  return __ret;
56805 }
56806 #endif
56807 
56808 #ifdef __LITTLE_ENDIAN__
56809 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56810  int32x4_t __ret;
56811  __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
56812  return __ret;
56813 }
56814 #else
56815 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56816  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56817  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56818  int32x4_t __ret;
56819  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
56820  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56821  return __ret;
56822 }
56823 #endif
56824 
56825 #ifdef __LITTLE_ENDIAN__
56826 #define vqdmull_high_lane_s32(__p0_635, __p1_635, __p2_635) __extension__ ({ \
56827  int32x4_t __s0_635 = __p0_635; \
56828  int32x2_t __s1_635 = __p1_635; \
56829  int64x2_t __ret_635; \
56830  __ret_635 = vqdmull_s32(vget_high_s32(__s0_635), splat_lane_s32(__s1_635, __p2_635)); \
56831  __ret_635; \
56832 })
56833 #else
56834 #define vqdmull_high_lane_s32(__p0_636, __p1_636, __p2_636) __extension__ ({ \
56835  int32x4_t __s0_636 = __p0_636; \
56836  int32x2_t __s1_636 = __p1_636; \
56837  int32x4_t __rev0_636; __rev0_636 = __builtin_shufflevector(__s0_636, __s0_636, 3, 2, 1, 0); \
56838  int32x2_t __rev1_636; __rev1_636 = __builtin_shufflevector(__s1_636, __s1_636, 1, 0); \
56839  int64x2_t __ret_636; \
56840  __ret_636 = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0_636), __noswap_splat_lane_s32(__rev1_636, __p2_636)); \
56841  __ret_636 = __builtin_shufflevector(__ret_636, __ret_636, 1, 0); \
56842  __ret_636; \
56843 })
56844 #endif
56845 
56846 #ifdef __LITTLE_ENDIAN__
56847 #define vqdmull_high_lane_s16(__p0_637, __p1_637, __p2_637) __extension__ ({ \
56848  int16x8_t __s0_637 = __p0_637; \
56849  int16x4_t __s1_637 = __p1_637; \
56850  int32x4_t __ret_637; \
56851  __ret_637 = vqdmull_s16(vget_high_s16(__s0_637), splat_lane_s16(__s1_637, __p2_637)); \
56852  __ret_637; \
56853 })
56854 #else
56855 #define vqdmull_high_lane_s16(__p0_638, __p1_638, __p2_638) __extension__ ({ \
56856  int16x8_t __s0_638 = __p0_638; \
56857  int16x4_t __s1_638 = __p1_638; \
56858  int16x8_t __rev0_638; __rev0_638 = __builtin_shufflevector(__s0_638, __s0_638, 7, 6, 5, 4, 3, 2, 1, 0); \
56859  int16x4_t __rev1_638; __rev1_638 = __builtin_shufflevector(__s1_638, __s1_638, 3, 2, 1, 0); \
56860  int32x4_t __ret_638; \
56861  __ret_638 = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0_638), __noswap_splat_lane_s16(__rev1_638, __p2_638)); \
56862  __ret_638 = __builtin_shufflevector(__ret_638, __ret_638, 3, 2, 1, 0); \
56863  __ret_638; \
56864 })
56865 #endif
56866 
56867 #ifdef __LITTLE_ENDIAN__
56868 #define vqdmull_high_laneq_s32(__p0_639, __p1_639, __p2_639) __extension__ ({ \
56869  int32x4_t __s0_639 = __p0_639; \
56870  int32x4_t __s1_639 = __p1_639; \
56871  int64x2_t __ret_639; \
56872  __ret_639 = vqdmull_s32(vget_high_s32(__s0_639), splat_laneq_s32(__s1_639, __p2_639)); \
56873  __ret_639; \
56874 })
56875 #else
56876 #define vqdmull_high_laneq_s32(__p0_640, __p1_640, __p2_640) __extension__ ({ \
56877  int32x4_t __s0_640 = __p0_640; \
56878  int32x4_t __s1_640 = __p1_640; \
56879  int32x4_t __rev0_640; __rev0_640 = __builtin_shufflevector(__s0_640, __s0_640, 3, 2, 1, 0); \
56880  int32x4_t __rev1_640; __rev1_640 = __builtin_shufflevector(__s1_640, __s1_640, 3, 2, 1, 0); \
56881  int64x2_t __ret_640; \
56882  __ret_640 = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0_640), __noswap_splat_laneq_s32(__rev1_640, __p2_640)); \
56883  __ret_640 = __builtin_shufflevector(__ret_640, __ret_640, 1, 0); \
56884  __ret_640; \
56885 })
56886 #endif
56887 
56888 #ifdef __LITTLE_ENDIAN__
56889 #define vqdmull_high_laneq_s16(__p0_641, __p1_641, __p2_641) __extension__ ({ \
56890  int16x8_t __s0_641 = __p0_641; \
56891  int16x8_t __s1_641 = __p1_641; \
56892  int32x4_t __ret_641; \
56893  __ret_641 = vqdmull_s16(vget_high_s16(__s0_641), splat_laneq_s16(__s1_641, __p2_641)); \
56894  __ret_641; \
56895 })
56896 #else
56897 #define vqdmull_high_laneq_s16(__p0_642, __p1_642, __p2_642) __extension__ ({ \
56898  int16x8_t __s0_642 = __p0_642; \
56899  int16x8_t __s1_642 = __p1_642; \
56900  int16x8_t __rev0_642; __rev0_642 = __builtin_shufflevector(__s0_642, __s0_642, 7, 6, 5, 4, 3, 2, 1, 0); \
56901  int16x8_t __rev1_642; __rev1_642 = __builtin_shufflevector(__s1_642, __s1_642, 7, 6, 5, 4, 3, 2, 1, 0); \
56902  int32x4_t __ret_642; \
56903  __ret_642 = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0_642), __noswap_splat_laneq_s16(__rev1_642, __p2_642)); \
56904  __ret_642 = __builtin_shufflevector(__ret_642, __ret_642, 3, 2, 1, 0); \
56905  __ret_642; \
56906 })
56907 #endif
56908 
56909 #ifdef __LITTLE_ENDIAN__
56910 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
56911  int64x2_t __ret;
56912  __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
56913  return __ret;
56914 }
56915 #else
56916 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
56917  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56918  int64x2_t __ret;
56919  __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
56920  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56921  return __ret;
56922 }
56923 #endif
56924 
56925 #ifdef __LITTLE_ENDIAN__
56926 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
56927  int32x4_t __ret;
56928  __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
56929  return __ret;
56930 }
56931 #else
56932 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
56933  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56934  int32x4_t __ret;
56935  __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
56936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56937  return __ret;
56938 }
56939 #endif
56940 
56941 #ifdef __LITTLE_ENDIAN__
56942 #define vqdmulls_lane_s32(__p0_643, __p1_643, __p2_643) __extension__ ({ \
56943  int32_t __s0_643 = __p0_643; \
56944  int32x2_t __s1_643 = __p1_643; \
56945  int64_t __ret_643; \
56946  __ret_643 = vqdmulls_s32(__s0_643, vget_lane_s32(__s1_643, __p2_643)); \
56947  __ret_643; \
56948 })
56949 #else
56950 #define vqdmulls_lane_s32(__p0_644, __p1_644, __p2_644) __extension__ ({ \
56951  int32_t __s0_644 = __p0_644; \
56952  int32x2_t __s1_644 = __p1_644; \
56953  int32x2_t __rev1_644; __rev1_644 = __builtin_shufflevector(__s1_644, __s1_644, 1, 0); \
56954  int64_t __ret_644; \
56955  __ret_644 = vqdmulls_s32(__s0_644, __noswap_vget_lane_s32(__rev1_644, __p2_644)); \
56956  __ret_644; \
56957 })
56958 #endif
56959 
56960 #ifdef __LITTLE_ENDIAN__
56961 #define vqdmullh_lane_s16(__p0_645, __p1_645, __p2_645) __extension__ ({ \
56962  int16_t __s0_645 = __p0_645; \
56963  int16x4_t __s1_645 = __p1_645; \
56964  int32_t __ret_645; \
56965  __ret_645 = vqdmullh_s16(__s0_645, vget_lane_s16(__s1_645, __p2_645)); \
56966  __ret_645; \
56967 })
56968 #else
56969 #define vqdmullh_lane_s16(__p0_646, __p1_646, __p2_646) __extension__ ({ \
56970  int16_t __s0_646 = __p0_646; \
56971  int16x4_t __s1_646 = __p1_646; \
56972  int16x4_t __rev1_646; __rev1_646 = __builtin_shufflevector(__s1_646, __s1_646, 3, 2, 1, 0); \
56973  int32_t __ret_646; \
56974  __ret_646 = vqdmullh_s16(__s0_646, __noswap_vget_lane_s16(__rev1_646, __p2_646)); \
56975  __ret_646; \
56976 })
56977 #endif
56978 
56979 #ifdef __LITTLE_ENDIAN__
56980 #define vqdmulls_laneq_s32(__p0_647, __p1_647, __p2_647) __extension__ ({ \
56981  int32_t __s0_647 = __p0_647; \
56982  int32x4_t __s1_647 = __p1_647; \
56983  int64_t __ret_647; \
56984  __ret_647 = vqdmulls_s32(__s0_647, vgetq_lane_s32(__s1_647, __p2_647)); \
56985  __ret_647; \
56986 })
56987 #else
56988 #define vqdmulls_laneq_s32(__p0_648, __p1_648, __p2_648) __extension__ ({ \
56989  int32_t __s0_648 = __p0_648; \
56990  int32x4_t __s1_648 = __p1_648; \
56991  int32x4_t __rev1_648; __rev1_648 = __builtin_shufflevector(__s1_648, __s1_648, 3, 2, 1, 0); \
56992  int64_t __ret_648; \
56993  __ret_648 = vqdmulls_s32(__s0_648, __noswap_vgetq_lane_s32(__rev1_648, __p2_648)); \
56994  __ret_648; \
56995 })
56996 #endif
56997 
56998 #ifdef __LITTLE_ENDIAN__
56999 #define vqdmullh_laneq_s16(__p0_649, __p1_649, __p2_649) __extension__ ({ \
57000  int16_t __s0_649 = __p0_649; \
57001  int16x8_t __s1_649 = __p1_649; \
57002  int32_t __ret_649; \
57003  __ret_649 = vqdmullh_s16(__s0_649, vgetq_lane_s16(__s1_649, __p2_649)); \
57004  __ret_649; \
57005 })
57006 #else
57007 #define vqdmullh_laneq_s16(__p0_650, __p1_650, __p2_650) __extension__ ({ \
57008  int16_t __s0_650 = __p0_650; \
57009  int16x8_t __s1_650 = __p1_650; \
57010  int16x8_t __rev1_650; __rev1_650 = __builtin_shufflevector(__s1_650, __s1_650, 7, 6, 5, 4, 3, 2, 1, 0); \
57011  int32_t __ret_650; \
57012  __ret_650 = vqdmullh_s16(__s0_650, __noswap_vgetq_lane_s16(__rev1_650, __p2_650)); \
57013  __ret_650; \
57014 })
57015 #endif
57016 
57017 #ifdef __LITTLE_ENDIAN__
57018 #define vqdmull_laneq_s32(__p0_651, __p1_651, __p2_651) __extension__ ({ \
57019  int32x2_t __s0_651 = __p0_651; \
57020  int32x4_t __s1_651 = __p1_651; \
57021  int64x2_t __ret_651; \
57022  __ret_651 = vqdmull_s32(__s0_651, splat_laneq_s32(__s1_651, __p2_651)); \
57023  __ret_651; \
57024 })
57025 #else
57026 #define vqdmull_laneq_s32(__p0_652, __p1_652, __p2_652) __extension__ ({ \
57027  int32x2_t __s0_652 = __p0_652; \
57028  int32x4_t __s1_652 = __p1_652; \
57029  int32x2_t __rev0_652; __rev0_652 = __builtin_shufflevector(__s0_652, __s0_652, 1, 0); \
57030  int32x4_t __rev1_652; __rev1_652 = __builtin_shufflevector(__s1_652, __s1_652, 3, 2, 1, 0); \
57031  int64x2_t __ret_652; \
57032  __ret_652 = __noswap_vqdmull_s32(__rev0_652, __noswap_splat_laneq_s32(__rev1_652, __p2_652)); \
57033  __ret_652 = __builtin_shufflevector(__ret_652, __ret_652, 1, 0); \
57034  __ret_652; \
57035 })
57036 #endif
57037 
57038 #ifdef __LITTLE_ENDIAN__
57039 #define vqdmull_laneq_s16(__p0_653, __p1_653, __p2_653) __extension__ ({ \
57040  int16x4_t __s0_653 = __p0_653; \
57041  int16x8_t __s1_653 = __p1_653; \
57042  int32x4_t __ret_653; \
57043  __ret_653 = vqdmull_s16(__s0_653, splat_laneq_s16(__s1_653, __p2_653)); \
57044  __ret_653; \
57045 })
57046 #else
57047 #define vqdmull_laneq_s16(__p0_654, __p1_654, __p2_654) __extension__ ({ \
57048  int16x4_t __s0_654 = __p0_654; \
57049  int16x8_t __s1_654 = __p1_654; \
57050  int16x4_t __rev0_654; __rev0_654 = __builtin_shufflevector(__s0_654, __s0_654, 3, 2, 1, 0); \
57051  int16x8_t __rev1_654; __rev1_654 = __builtin_shufflevector(__s1_654, __s1_654, 7, 6, 5, 4, 3, 2, 1, 0); \
57052  int32x4_t __ret_654; \
57053  __ret_654 = __noswap_vqdmull_s16(__rev0_654, __noswap_splat_laneq_s16(__rev1_654, __p2_654)); \
57054  __ret_654 = __builtin_shufflevector(__ret_654, __ret_654, 3, 2, 1, 0); \
57055  __ret_654; \
57056 })
57057 #endif
57058 
57059 __ai int16_t vqmovns_s32(int32_t __p0) {
57060  int16_t __ret;
57061  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57062  return __ret;
57063 }
57064 __ai int32_t vqmovnd_s64(int64_t __p0) {
57065  int32_t __ret;
57066  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57067  return __ret;
57068 }
57069 __ai int8_t vqmovnh_s16(int16_t __p0) {
57070  int8_t __ret;
57071  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57072  return __ret;
57073 }
57074 __ai uint16_t vqmovns_u32(uint32_t __p0) {
57075  uint16_t __ret;
57076  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57077  return __ret;
57078 }
57079 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
57080  uint32_t __ret;
57081  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57082  return __ret;
57083 }
57084 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
57085  uint8_t __ret;
57086  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57087  return __ret;
57088 }
57089 #ifdef __LITTLE_ENDIAN__
57090 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57091  uint16x8_t __ret;
57092  __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
57093  return __ret;
57094 }
57095 #else
57096 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57097  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57098  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57099  uint16x8_t __ret;
57100  __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
57101  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57102  return __ret;
57103 }
57104 #endif
57105 
57106 #ifdef __LITTLE_ENDIAN__
57107 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57108  uint32x4_t __ret;
57109  __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
57110  return __ret;
57111 }
57112 #else
57113 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57114  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57115  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57116  uint32x4_t __ret;
57117  __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
57118  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57119  return __ret;
57120 }
57121 #endif
57122 
57123 #ifdef __LITTLE_ENDIAN__
57124 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57125  uint8x16_t __ret;
57126  __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
57127  return __ret;
57128 }
57129 #else
57130 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57131  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57132  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57133  uint8x16_t __ret;
57134  __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
57135  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57136  return __ret;
57137 }
57138 #endif
57139 
57140 #ifdef __LITTLE_ENDIAN__
57141 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57142  int16x8_t __ret;
57143  __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
57144  return __ret;
57145 }
57146 #else
57147 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57148  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57149  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57150  int16x8_t __ret;
57151  __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
57152  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57153  return __ret;
57154 }
57155 #endif
57156 
57157 #ifdef __LITTLE_ENDIAN__
57158 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57159  int32x4_t __ret;
57160  __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
57161  return __ret;
57162 }
57163 #else
57164 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57165  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57166  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57167  int32x4_t __ret;
57168  __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
57169  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57170  return __ret;
57171 }
57172 #endif
57173 
57174 #ifdef __LITTLE_ENDIAN__
57175 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57176  int8x16_t __ret;
57177  __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
57178  return __ret;
57179 }
57180 #else
57181 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57182  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57183  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57184  int8x16_t __ret;
57185  __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
57186  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57187  return __ret;
57188 }
57189 #endif
57190 
57191 __ai int16_t vqmovuns_s32(int32_t __p0) {
57192  int16_t __ret;
57193  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57194  return __ret;
57195 }
57196 __ai int32_t vqmovund_s64(int64_t __p0) {
57197  int32_t __ret;
57198  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57199  return __ret;
57200 }
57201 __ai int8_t vqmovunh_s16(int16_t __p0) {
57202  int8_t __ret;
57203  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57204  return __ret;
57205 }
57206 #ifdef __LITTLE_ENDIAN__
57207 __ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57208  uint16x8_t __ret;
57209  __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
57210  return __ret;
57211 }
57212 #else
57213 __ai uint16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57214  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57215  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57216  uint16x8_t __ret;
57217  __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
57218  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57219  return __ret;
57220 }
57221 #endif
57222 
57223 #ifdef __LITTLE_ENDIAN__
57224 __ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57225  uint32x4_t __ret;
57226  __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
57227  return __ret;
57228 }
57229 #else
57230 __ai uint32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57231  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57232  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57233  uint32x4_t __ret;
57234  __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
57235  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57236  return __ret;
57237 }
57238 #endif
57239 
57240 #ifdef __LITTLE_ENDIAN__
57241 __ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57242  uint8x16_t __ret;
57243  __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
57244  return __ret;
57245 }
57246 #else
57247 __ai uint8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57248  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57249  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57250  uint8x16_t __ret;
57251  __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
57252  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57253  return __ret;
57254 }
57255 #endif
57256 
57257 #ifdef __LITTLE_ENDIAN__
57258 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57259  int64x2_t __ret;
57260  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
57261  return __ret;
57262 }
57263 #else
57264 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57265  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57266  int64x2_t __ret;
57267  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
57268  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57269  return __ret;
57270 }
57271 #endif
57272 
57273 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
57274  int64x1_t __ret;
57275  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57276  return __ret;
57277 }
57278 __ai int8_t vqnegb_s8(int8_t __p0) {
57279  int8_t __ret;
57280  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57281  return __ret;
57282 }
57283 __ai int32_t vqnegs_s32(int32_t __p0) {
57284  int32_t __ret;
57285  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57286  return __ret;
57287 }
57288 __ai int64_t vqnegd_s64(int64_t __p0) {
57289  int64_t __ret;
57290  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57291  return __ret;
57292 }
57293 __ai int16_t vqnegh_s16(int16_t __p0) {
57294  int16_t __ret;
57295  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57296  return __ret;
57297 }
57298 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57299  int32_t __ret;
57300  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57301  return __ret;
57302 }
57303 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57304  int16_t __ret;
57305  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57306  return __ret;
57307 }
57308 #ifdef __LITTLE_ENDIAN__
57309 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
57310  int32x4_t __s0 = __p0; \
57311  int32x2_t __s1 = __p1; \
57312  int32x4_t __ret; \
57313  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 2); \
57314  __ret; \
57315 })
57316 #else
57317 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
57318  int32x4_t __s0 = __p0; \
57319  int32x2_t __s1 = __p1; \
57320  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57321  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
57322  int32x4_t __ret; \
57323  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
57324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57325  __ret; \
57326 })
57327 #endif
57328 
57329 #ifdef __LITTLE_ENDIAN__
57330 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
57331  int16x8_t __s0 = __p0; \
57332  int16x4_t __s1 = __p1; \
57333  int16x8_t __ret; \
57334  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__s0, (int8x8_t)__s1, __p2, 1); \
57335  __ret; \
57336 })
57337 #else
57338 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
57339  int16x8_t __s0 = __p0; \
57340  int16x4_t __s1 = __p1; \
57341  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
57342  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57343  int16x8_t __ret; \
57344  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_lane_v((int8x16_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
57345  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
57346  __ret; \
57347 })
57348 #endif
57349 
57350 #ifdef __LITTLE_ENDIAN__
57351 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
57352  int32x2_t __s0 = __p0; \
57353  int32x2_t __s1 = __p1; \
57354  int32x2_t __ret; \
57355  __ret = (int32x2_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
57356  __ret; \
57357 })
57358 #else
57359 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
57360  int32x2_t __s0 = __p0; \
57361  int32x2_t __s1 = __p1; \
57362  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57363  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
57364  int32x2_t __ret; \
57365  __ret = (int32x2_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
57366  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57367  __ret; \
57368 })
57369 #endif
57370 
57371 #ifdef __LITTLE_ENDIAN__
57372 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
57373  int16x4_t __s0 = __p0; \
57374  int16x4_t __s1 = __p1; \
57375  int16x4_t __ret; \
57376  __ret = (int16x4_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
57377  __ret; \
57378 })
57379 #else
57380 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
57381  int16x4_t __s0 = __p0; \
57382  int16x4_t __s1 = __p1; \
57383  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57384  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57385  int16x4_t __ret; \
57386  __ret = (int16x4_t) __builtin_neon_vqrdmulh_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
57387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57388  __ret; \
57389 })
57390 #endif
57391 
57392 #ifdef __LITTLE_ENDIAN__
57393 #define vqrdmulhs_lane_s32(__p0_655, __p1_655, __p2_655) __extension__ ({ \
57394  int32_t __s0_655 = __p0_655; \
57395  int32x2_t __s1_655 = __p1_655; \
57396  int32_t __ret_655; \
57397  __ret_655 = vqrdmulhs_s32(__s0_655, vget_lane_s32(__s1_655, __p2_655)); \
57398  __ret_655; \
57399 })
57400 #else
57401 #define vqrdmulhs_lane_s32(__p0_656, __p1_656, __p2_656) __extension__ ({ \
57402  int32_t __s0_656 = __p0_656; \
57403  int32x2_t __s1_656 = __p1_656; \
57404  int32x2_t __rev1_656; __rev1_656 = __builtin_shufflevector(__s1_656, __s1_656, 1, 0); \
57405  int32_t __ret_656; \
57406  __ret_656 = vqrdmulhs_s32(__s0_656, __noswap_vget_lane_s32(__rev1_656, __p2_656)); \
57407  __ret_656; \
57408 })
57409 #endif
57410 
57411 #ifdef __LITTLE_ENDIAN__
57412 #define vqrdmulhh_lane_s16(__p0_657, __p1_657, __p2_657) __extension__ ({ \
57413  int16_t __s0_657 = __p0_657; \
57414  int16x4_t __s1_657 = __p1_657; \
57415  int16_t __ret_657; \
57416  __ret_657 = vqrdmulhh_s16(__s0_657, vget_lane_s16(__s1_657, __p2_657)); \
57417  __ret_657; \
57418 })
57419 #else
57420 #define vqrdmulhh_lane_s16(__p0_658, __p1_658, __p2_658) __extension__ ({ \
57421  int16_t __s0_658 = __p0_658; \
57422  int16x4_t __s1_658 = __p1_658; \
57423  int16x4_t __rev1_658; __rev1_658 = __builtin_shufflevector(__s1_658, __s1_658, 3, 2, 1, 0); \
57424  int16_t __ret_658; \
57425  __ret_658 = vqrdmulhh_s16(__s0_658, __noswap_vget_lane_s16(__rev1_658, __p2_658)); \
57426  __ret_658; \
57427 })
57428 #endif
57429 
57430 #ifdef __LITTLE_ENDIAN__
57431 #define vqrdmulhs_laneq_s32(__p0_659, __p1_659, __p2_659) __extension__ ({ \
57432  int32_t __s0_659 = __p0_659; \
57433  int32x4_t __s1_659 = __p1_659; \
57434  int32_t __ret_659; \
57435  __ret_659 = vqrdmulhs_s32(__s0_659, vgetq_lane_s32(__s1_659, __p2_659)); \
57436  __ret_659; \
57437 })
57438 #else
57439 #define vqrdmulhs_laneq_s32(__p0_660, __p1_660, __p2_660) __extension__ ({ \
57440  int32_t __s0_660 = __p0_660; \
57441  int32x4_t __s1_660 = __p1_660; \
57442  int32x4_t __rev1_660; __rev1_660 = __builtin_shufflevector(__s1_660, __s1_660, 3, 2, 1, 0); \
57443  int32_t __ret_660; \
57444  __ret_660 = vqrdmulhs_s32(__s0_660, __noswap_vgetq_lane_s32(__rev1_660, __p2_660)); \
57445  __ret_660; \
57446 })
57447 #endif
57448 
57449 #ifdef __LITTLE_ENDIAN__
57450 #define vqrdmulhh_laneq_s16(__p0_661, __p1_661, __p2_661) __extension__ ({ \
57451  int16_t __s0_661 = __p0_661; \
57452  int16x8_t __s1_661 = __p1_661; \
57453  int16_t __ret_661; \
57454  __ret_661 = vqrdmulhh_s16(__s0_661, vgetq_lane_s16(__s1_661, __p2_661)); \
57455  __ret_661; \
57456 })
57457 #else
57458 #define vqrdmulhh_laneq_s16(__p0_662, __p1_662, __p2_662) __extension__ ({ \
57459  int16_t __s0_662 = __p0_662; \
57460  int16x8_t __s1_662 = __p1_662; \
57461  int16x8_t __rev1_662; __rev1_662 = __builtin_shufflevector(__s1_662, __s1_662, 7, 6, 5, 4, 3, 2, 1, 0); \
57462  int16_t __ret_662; \
57463  __ret_662 = vqrdmulhh_s16(__s0_662, __noswap_vgetq_lane_s16(__rev1_662, __p2_662)); \
57464  __ret_662; \
57465 })
57466 #endif
57467 
57468 #ifdef __LITTLE_ENDIAN__
57469 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57470  int32x4_t __s0 = __p0; \
57471  int32x4_t __s1 = __p1; \
57472  int32x4_t __ret; \
57473  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
57474  __ret; \
57475 })
57476 #else
57477 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57478  int32x4_t __s0 = __p0; \
57479  int32x4_t __s1 = __p1; \
57480  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57481  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57482  int32x4_t __ret; \
57483  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
57484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57485  __ret; \
57486 })
57487 #endif
57488 
57489 #ifdef __LITTLE_ENDIAN__
57490 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57491  int16x8_t __s0 = __p0; \
57492  int16x8_t __s1 = __p1; \
57493  int16x8_t __ret; \
57494  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
57495  __ret; \
57496 })
57497 #else
57498 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57499  int16x8_t __s0 = __p0; \
57500  int16x8_t __s1 = __p1; \
57501  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
57502  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57503  int16x8_t __ret; \
57504  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
57505  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
57506  __ret; \
57507 })
57508 #endif
57509 
57510 #ifdef __LITTLE_ENDIAN__
57511 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57512  int32x2_t __s0 = __p0; \
57513  int32x4_t __s1 = __p1; \
57514  int32x2_t __ret; \
57515  __ret = (int32x2_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 2); \
57516  __ret; \
57517 })
57518 #else
57519 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57520  int32x2_t __s0 = __p0; \
57521  int32x4_t __s1 = __p1; \
57522  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57523  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57524  int32x2_t __ret; \
57525  __ret = (int32x2_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 2); \
57526  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57527  __ret; \
57528 })
57529 #endif
57530 
57531 #ifdef __LITTLE_ENDIAN__
57532 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57533  int16x4_t __s0 = __p0; \
57534  int16x8_t __s1 = __p1; \
57535  int16x4_t __ret; \
57536  __ret = (int16x4_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 1); \
57537  __ret; \
57538 })
57539 #else
57540 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57541  int16x4_t __s0 = __p0; \
57542  int16x8_t __s1 = __p1; \
57543  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57544  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57545  int16x4_t __ret; \
57546  __ret = (int16x4_t) __builtin_neon_vqrdmulh_laneq_v((int8x8_t)__rev0, (int8x16_t)__rev1, __p2, 1); \
57547  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57548  __ret; \
57549 })
57550 #endif
57551 
57552 __ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57553  uint8_t __ret;
57554  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57555  return __ret;
57556 }
57557 __ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57558  uint32_t __ret;
57559  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57560  return __ret;
57561 }
57562 __ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57563  uint64_t __ret;
57564  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57565  return __ret;
57566 }
57567 __ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57568  uint16_t __ret;
57569  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57570  return __ret;
57571 }
57572 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57573  int8_t __ret;
57574  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57575  return __ret;
57576 }
57577 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57578  int32_t __ret;
57579  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57580  return __ret;
57581 }
57582 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57583  int64_t __ret;
57584  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57585  return __ret;
57586 }
57587 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57588  int16_t __ret;
57589  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57590  return __ret;
57591 }
57592 #ifdef __LITTLE_ENDIAN__
57593 #define vqrshrn_high_n_u32(__p0_663, __p1_663, __p2_663) __extension__ ({ \
57594  uint16x4_t __s0_663 = __p0_663; \
57595  uint32x4_t __s1_663 = __p1_663; \
57596  uint16x8_t __ret_663; \
57597  __ret_663 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_663), (uint16x4_t)(vqrshrn_n_u32(__s1_663, __p2_663)))); \
57598  __ret_663; \
57599 })
57600 #else
57601 #define vqrshrn_high_n_u32(__p0_664, __p1_664, __p2_664) __extension__ ({ \
57602  uint16x4_t __s0_664 = __p0_664; \
57603  uint32x4_t __s1_664 = __p1_664; \
57604  uint16x4_t __rev0_664; __rev0_664 = __builtin_shufflevector(__s0_664, __s0_664, 3, 2, 1, 0); \
57605  uint32x4_t __rev1_664; __rev1_664 = __builtin_shufflevector(__s1_664, __s1_664, 3, 2, 1, 0); \
57606  uint16x8_t __ret_664; \
57607  __ret_664 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_664), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_664, __p2_664)))); \
57608  __ret_664 = __builtin_shufflevector(__ret_664, __ret_664, 7, 6, 5, 4, 3, 2, 1, 0); \
57609  __ret_664; \
57610 })
57611 #endif
57612 
57613 #ifdef __LITTLE_ENDIAN__
57614 #define vqrshrn_high_n_u64(__p0_665, __p1_665, __p2_665) __extension__ ({ \
57615  uint32x2_t __s0_665 = __p0_665; \
57616  uint64x2_t __s1_665 = __p1_665; \
57617  uint32x4_t __ret_665; \
57618  __ret_665 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_665), (uint32x2_t)(vqrshrn_n_u64(__s1_665, __p2_665)))); \
57619  __ret_665; \
57620 })
57621 #else
57622 #define vqrshrn_high_n_u64(__p0_666, __p1_666, __p2_666) __extension__ ({ \
57623  uint32x2_t __s0_666 = __p0_666; \
57624  uint64x2_t __s1_666 = __p1_666; \
57625  uint32x2_t __rev0_666; __rev0_666 = __builtin_shufflevector(__s0_666, __s0_666, 1, 0); \
57626  uint64x2_t __rev1_666; __rev1_666 = __builtin_shufflevector(__s1_666, __s1_666, 1, 0); \
57627  uint32x4_t __ret_666; \
57628  __ret_666 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_666), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_666, __p2_666)))); \
57629  __ret_666 = __builtin_shufflevector(__ret_666, __ret_666, 3, 2, 1, 0); \
57630  __ret_666; \
57631 })
57632 #endif
57633 
57634 #ifdef __LITTLE_ENDIAN__
57635 #define vqrshrn_high_n_u16(__p0_667, __p1_667, __p2_667) __extension__ ({ \
57636  uint8x8_t __s0_667 = __p0_667; \
57637  uint16x8_t __s1_667 = __p1_667; \
57638  uint8x16_t __ret_667; \
57639  __ret_667 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_667), (uint8x8_t)(vqrshrn_n_u16(__s1_667, __p2_667)))); \
57640  __ret_667; \
57641 })
57642 #else
57643 #define vqrshrn_high_n_u16(__p0_668, __p1_668, __p2_668) __extension__ ({ \
57644  uint8x8_t __s0_668 = __p0_668; \
57645  uint16x8_t __s1_668 = __p1_668; \
57646  uint8x8_t __rev0_668; __rev0_668 = __builtin_shufflevector(__s0_668, __s0_668, 7, 6, 5, 4, 3, 2, 1, 0); \
57647  uint16x8_t __rev1_668; __rev1_668 = __builtin_shufflevector(__s1_668, __s1_668, 7, 6, 5, 4, 3, 2, 1, 0); \
57648  uint8x16_t __ret_668; \
57649  __ret_668 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_668), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_668, __p2_668)))); \
57650  __ret_668 = __builtin_shufflevector(__ret_668, __ret_668, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57651  __ret_668; \
57652 })
57653 #endif
57654 
57655 #ifdef __LITTLE_ENDIAN__
57656 #define vqrshrn_high_n_s32(__p0_669, __p1_669, __p2_669) __extension__ ({ \
57657  int16x4_t __s0_669 = __p0_669; \
57658  int32x4_t __s1_669 = __p1_669; \
57659  int16x8_t __ret_669; \
57660  __ret_669 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_669), (int16x4_t)(vqrshrn_n_s32(__s1_669, __p2_669)))); \
57661  __ret_669; \
57662 })
57663 #else
57664 #define vqrshrn_high_n_s32(__p0_670, __p1_670, __p2_670) __extension__ ({ \
57665  int16x4_t __s0_670 = __p0_670; \
57666  int32x4_t __s1_670 = __p1_670; \
57667  int16x4_t __rev0_670; __rev0_670 = __builtin_shufflevector(__s0_670, __s0_670, 3, 2, 1, 0); \
57668  int32x4_t __rev1_670; __rev1_670 = __builtin_shufflevector(__s1_670, __s1_670, 3, 2, 1, 0); \
57669  int16x8_t __ret_670; \
57670  __ret_670 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_670), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_670, __p2_670)))); \
57671  __ret_670 = __builtin_shufflevector(__ret_670, __ret_670, 7, 6, 5, 4, 3, 2, 1, 0); \
57672  __ret_670; \
57673 })
57674 #endif
57675 
57676 #ifdef __LITTLE_ENDIAN__
57677 #define vqrshrn_high_n_s64(__p0_671, __p1_671, __p2_671) __extension__ ({ \
57678  int32x2_t __s0_671 = __p0_671; \
57679  int64x2_t __s1_671 = __p1_671; \
57680  int32x4_t __ret_671; \
57681  __ret_671 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_671), (int32x2_t)(vqrshrn_n_s64(__s1_671, __p2_671)))); \
57682  __ret_671; \
57683 })
57684 #else
57685 #define vqrshrn_high_n_s64(__p0_672, __p1_672, __p2_672) __extension__ ({ \
57686  int32x2_t __s0_672 = __p0_672; \
57687  int64x2_t __s1_672 = __p1_672; \
57688  int32x2_t __rev0_672; __rev0_672 = __builtin_shufflevector(__s0_672, __s0_672, 1, 0); \
57689  int64x2_t __rev1_672; __rev1_672 = __builtin_shufflevector(__s1_672, __s1_672, 1, 0); \
57690  int32x4_t __ret_672; \
57691  __ret_672 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_672), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_672, __p2_672)))); \
57692  __ret_672 = __builtin_shufflevector(__ret_672, __ret_672, 3, 2, 1, 0); \
57693  __ret_672; \
57694 })
57695 #endif
57696 
57697 #ifdef __LITTLE_ENDIAN__
57698 #define vqrshrn_high_n_s16(__p0_673, __p1_673, __p2_673) __extension__ ({ \
57699  int8x8_t __s0_673 = __p0_673; \
57700  int16x8_t __s1_673 = __p1_673; \
57701  int8x16_t __ret_673; \
57702  __ret_673 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_673), (int8x8_t)(vqrshrn_n_s16(__s1_673, __p2_673)))); \
57703  __ret_673; \
57704 })
57705 #else
57706 #define vqrshrn_high_n_s16(__p0_674, __p1_674, __p2_674) __extension__ ({ \
57707  int8x8_t __s0_674 = __p0_674; \
57708  int16x8_t __s1_674 = __p1_674; \
57709  int8x8_t __rev0_674; __rev0_674 = __builtin_shufflevector(__s0_674, __s0_674, 7, 6, 5, 4, 3, 2, 1, 0); \
57710  int16x8_t __rev1_674; __rev1_674 = __builtin_shufflevector(__s1_674, __s1_674, 7, 6, 5, 4, 3, 2, 1, 0); \
57711  int8x16_t __ret_674; \
57712  __ret_674 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_674), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_674, __p2_674)))); \
57713  __ret_674 = __builtin_shufflevector(__ret_674, __ret_674, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57714  __ret_674; \
57715 })
57716 #endif
57717 
57718 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57719  uint32_t __s0 = __p0; \
57720  uint16_t __ret; \
57721  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57722  __ret; \
57723 })
57724 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57725  uint64_t __s0 = __p0; \
57726  uint32_t __ret; \
57727  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57728  __ret; \
57729 })
57730 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57731  uint16_t __s0 = __p0; \
57732  uint8_t __ret; \
57733  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57734  __ret; \
57735 })
57736 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
57737  int32_t __s0 = __p0; \
57738  int16_t __ret; \
57739  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
57740  __ret; \
57741 })
57742 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
57743  int64_t __s0 = __p0; \
57744  int32_t __ret; \
57745  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
57746  __ret; \
57747 })
57748 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
57749  int16_t __s0 = __p0; \
57750  int8_t __ret; \
57751  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
57752  __ret; \
57753 })
57754 #ifdef __LITTLE_ENDIAN__
57755 #define vqrshrun_high_n_s32(__p0_675, __p1_675, __p2_675) __extension__ ({ \
57756  int16x4_t __s0_675 = __p0_675; \
57757  int32x4_t __s1_675 = __p1_675; \
57758  int16x8_t __ret_675; \
57759  __ret_675 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_675), (int16x4_t)(vqrshrun_n_s32(__s1_675, __p2_675)))); \
57760  __ret_675; \
57761 })
57762 #else
57763 #define vqrshrun_high_n_s32(__p0_676, __p1_676, __p2_676) __extension__ ({ \
57764  int16x4_t __s0_676 = __p0_676; \
57765  int32x4_t __s1_676 = __p1_676; \
57766  int16x4_t __rev0_676; __rev0_676 = __builtin_shufflevector(__s0_676, __s0_676, 3, 2, 1, 0); \
57767  int32x4_t __rev1_676; __rev1_676 = __builtin_shufflevector(__s1_676, __s1_676, 3, 2, 1, 0); \
57768  int16x8_t __ret_676; \
57769  __ret_676 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_676), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_676, __p2_676)))); \
57770  __ret_676 = __builtin_shufflevector(__ret_676, __ret_676, 7, 6, 5, 4, 3, 2, 1, 0); \
57771  __ret_676; \
57772 })
57773 #endif
57774 
57775 #ifdef __LITTLE_ENDIAN__
57776 #define vqrshrun_high_n_s64(__p0_677, __p1_677, __p2_677) __extension__ ({ \
57777  int32x2_t __s0_677 = __p0_677; \
57778  int64x2_t __s1_677 = __p1_677; \
57779  int32x4_t __ret_677; \
57780  __ret_677 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_677), (int32x2_t)(vqrshrun_n_s64(__s1_677, __p2_677)))); \
57781  __ret_677; \
57782 })
57783 #else
57784 #define vqrshrun_high_n_s64(__p0_678, __p1_678, __p2_678) __extension__ ({ \
57785  int32x2_t __s0_678 = __p0_678; \
57786  int64x2_t __s1_678 = __p1_678; \
57787  int32x2_t __rev0_678; __rev0_678 = __builtin_shufflevector(__s0_678, __s0_678, 1, 0); \
57788  int64x2_t __rev1_678; __rev1_678 = __builtin_shufflevector(__s1_678, __s1_678, 1, 0); \
57789  int32x4_t __ret_678; \
57790  __ret_678 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_678), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_678, __p2_678)))); \
57791  __ret_678 = __builtin_shufflevector(__ret_678, __ret_678, 3, 2, 1, 0); \
57792  __ret_678; \
57793 })
57794 #endif
57795 
57796 #ifdef __LITTLE_ENDIAN__
57797 #define vqrshrun_high_n_s16(__p0_679, __p1_679, __p2_679) __extension__ ({ \
57798  int8x8_t __s0_679 = __p0_679; \
57799  int16x8_t __s1_679 = __p1_679; \
57800  int8x16_t __ret_679; \
57801  __ret_679 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_679), (int8x8_t)(vqrshrun_n_s16(__s1_679, __p2_679)))); \
57802  __ret_679; \
57803 })
57804 #else
57805 #define vqrshrun_high_n_s16(__p0_680, __p1_680, __p2_680) __extension__ ({ \
57806  int8x8_t __s0_680 = __p0_680; \
57807  int16x8_t __s1_680 = __p1_680; \
57808  int8x8_t __rev0_680; __rev0_680 = __builtin_shufflevector(__s0_680, __s0_680, 7, 6, 5, 4, 3, 2, 1, 0); \
57809  int16x8_t __rev1_680; __rev1_680 = __builtin_shufflevector(__s1_680, __s1_680, 7, 6, 5, 4, 3, 2, 1, 0); \
57810  int8x16_t __ret_680; \
57811  __ret_680 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_680), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_680, __p2_680)))); \
57812  __ret_680 = __builtin_shufflevector(__ret_680, __ret_680, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57813  __ret_680; \
57814 })
57815 #endif
57816 
57817 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
57818  int32_t __s0 = __p0; \
57819  int16_t __ret; \
57820  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
57821  __ret; \
57822 })
57823 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
57824  int64_t __s0 = __p0; \
57825  int32_t __ret; \
57826  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
57827  __ret; \
57828 })
57829 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
57830  int16_t __s0 = __p0; \
57831  int8_t __ret; \
57832  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
57833  __ret; \
57834 })
57835 __ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
57836  uint8_t __ret;
57837  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
57838  return __ret;
57839 }
57840 __ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
57841  uint32_t __ret;
57842  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
57843  return __ret;
57844 }
57845 __ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
57846  uint64_t __ret;
57847  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
57848  return __ret;
57849 }
57850 __ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
57851  uint16_t __ret;
57852  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
57853  return __ret;
57854 }
57855 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
57856  int8_t __ret;
57857  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
57858  return __ret;
57859 }
57860 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
57861  int32_t __ret;
57862  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
57863  return __ret;
57864 }
57865 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
57866  int64_t __ret;
57867  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
57868  return __ret;
57869 }
57870 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
57871  int16_t __ret;
57872  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
57873  return __ret;
57874 }
57875 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
57876  uint8_t __s0 = __p0; \
57877  uint8_t __ret; \
57878  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
57879  __ret; \
57880 })
57881 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
57882  uint32_t __s0 = __p0; \
57883  uint32_t __ret; \
57884  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
57885  __ret; \
57886 })
57887 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
57888  uint64_t __s0 = __p0; \
57889  uint64_t __ret; \
57890  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
57891  __ret; \
57892 })
57893 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
57894  uint16_t __s0 = __p0; \
57895  uint16_t __ret; \
57896  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
57897  __ret; \
57898 })
57899 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
57900  int8_t __s0 = __p0; \
57901  int8_t __ret; \
57902  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
57903  __ret; \
57904 })
57905 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
57906  int32_t __s0 = __p0; \
57907  int32_t __ret; \
57908  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
57909  __ret; \
57910 })
57911 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
57912  int64_t __s0 = __p0; \
57913  int64_t __ret; \
57914  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
57915  __ret; \
57916 })
57917 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
57918  int16_t __s0 = __p0; \
57919  int16_t __ret; \
57920  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
57921  __ret; \
57922 })
57923 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
57924  int8_t __s0 = __p0; \
57925  int8_t __ret; \
57926  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
57927  __ret; \
57928 })
57929 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
57930  int32_t __s0 = __p0; \
57931  int32_t __ret; \
57932  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
57933  __ret; \
57934 })
57935 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
57936  int64_t __s0 = __p0; \
57937  int64_t __ret; \
57938  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
57939  __ret; \
57940 })
57941 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
57942  int16_t __s0 = __p0; \
57943  int16_t __ret; \
57944  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
57945  __ret; \
57946 })
57947 #ifdef __LITTLE_ENDIAN__
57948 #define vqshrn_high_n_u32(__p0_681, __p1_681, __p2_681) __extension__ ({ \
57949  uint16x4_t __s0_681 = __p0_681; \
57950  uint32x4_t __s1_681 = __p1_681; \
57951  uint16x8_t __ret_681; \
57952  __ret_681 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_681), (uint16x4_t)(vqshrn_n_u32(__s1_681, __p2_681)))); \
57953  __ret_681; \
57954 })
57955 #else
57956 #define vqshrn_high_n_u32(__p0_682, __p1_682, __p2_682) __extension__ ({ \
57957  uint16x4_t __s0_682 = __p0_682; \
57958  uint32x4_t __s1_682 = __p1_682; \
57959  uint16x4_t __rev0_682; __rev0_682 = __builtin_shufflevector(__s0_682, __s0_682, 3, 2, 1, 0); \
57960  uint32x4_t __rev1_682; __rev1_682 = __builtin_shufflevector(__s1_682, __s1_682, 3, 2, 1, 0); \
57961  uint16x8_t __ret_682; \
57962  __ret_682 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_682), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_682, __p2_682)))); \
57963  __ret_682 = __builtin_shufflevector(__ret_682, __ret_682, 7, 6, 5, 4, 3, 2, 1, 0); \
57964  __ret_682; \
57965 })
57966 #endif
57967 
57968 #ifdef __LITTLE_ENDIAN__
57969 #define vqshrn_high_n_u64(__p0_683, __p1_683, __p2_683) __extension__ ({ \
57970  uint32x2_t __s0_683 = __p0_683; \
57971  uint64x2_t __s1_683 = __p1_683; \
57972  uint32x4_t __ret_683; \
57973  __ret_683 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_683), (uint32x2_t)(vqshrn_n_u64(__s1_683, __p2_683)))); \
57974  __ret_683; \
57975 })
57976 #else
57977 #define vqshrn_high_n_u64(__p0_684, __p1_684, __p2_684) __extension__ ({ \
57978  uint32x2_t __s0_684 = __p0_684; \
57979  uint64x2_t __s1_684 = __p1_684; \
57980  uint32x2_t __rev0_684; __rev0_684 = __builtin_shufflevector(__s0_684, __s0_684, 1, 0); \
57981  uint64x2_t __rev1_684; __rev1_684 = __builtin_shufflevector(__s1_684, __s1_684, 1, 0); \
57982  uint32x4_t __ret_684; \
57983  __ret_684 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_684), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_684, __p2_684)))); \
57984  __ret_684 = __builtin_shufflevector(__ret_684, __ret_684, 3, 2, 1, 0); \
57985  __ret_684; \
57986 })
57987 #endif
57988 
57989 #ifdef __LITTLE_ENDIAN__
57990 #define vqshrn_high_n_u16(__p0_685, __p1_685, __p2_685) __extension__ ({ \
57991  uint8x8_t __s0_685 = __p0_685; \
57992  uint16x8_t __s1_685 = __p1_685; \
57993  uint8x16_t __ret_685; \
57994  __ret_685 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_685), (uint8x8_t)(vqshrn_n_u16(__s1_685, __p2_685)))); \
57995  __ret_685; \
57996 })
57997 #else
57998 #define vqshrn_high_n_u16(__p0_686, __p1_686, __p2_686) __extension__ ({ \
57999  uint8x8_t __s0_686 = __p0_686; \
58000  uint16x8_t __s1_686 = __p1_686; \
58001  uint8x8_t __rev0_686; __rev0_686 = __builtin_shufflevector(__s0_686, __s0_686, 7, 6, 5, 4, 3, 2, 1, 0); \
58002  uint16x8_t __rev1_686; __rev1_686 = __builtin_shufflevector(__s1_686, __s1_686, 7, 6, 5, 4, 3, 2, 1, 0); \
58003  uint8x16_t __ret_686; \
58004  __ret_686 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_686), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_686, __p2_686)))); \
58005  __ret_686 = __builtin_shufflevector(__ret_686, __ret_686, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58006  __ret_686; \
58007 })
58008 #endif
58009 
58010 #ifdef __LITTLE_ENDIAN__
58011 #define vqshrn_high_n_s32(__p0_687, __p1_687, __p2_687) __extension__ ({ \
58012  int16x4_t __s0_687 = __p0_687; \
58013  int32x4_t __s1_687 = __p1_687; \
58014  int16x8_t __ret_687; \
58015  __ret_687 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_687), (int16x4_t)(vqshrn_n_s32(__s1_687, __p2_687)))); \
58016  __ret_687; \
58017 })
58018 #else
58019 #define vqshrn_high_n_s32(__p0_688, __p1_688, __p2_688) __extension__ ({ \
58020  int16x4_t __s0_688 = __p0_688; \
58021  int32x4_t __s1_688 = __p1_688; \
58022  int16x4_t __rev0_688; __rev0_688 = __builtin_shufflevector(__s0_688, __s0_688, 3, 2, 1, 0); \
58023  int32x4_t __rev1_688; __rev1_688 = __builtin_shufflevector(__s1_688, __s1_688, 3, 2, 1, 0); \
58024  int16x8_t __ret_688; \
58025  __ret_688 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_688), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_688, __p2_688)))); \
58026  __ret_688 = __builtin_shufflevector(__ret_688, __ret_688, 7, 6, 5, 4, 3, 2, 1, 0); \
58027  __ret_688; \
58028 })
58029 #endif
58030 
58031 #ifdef __LITTLE_ENDIAN__
58032 #define vqshrn_high_n_s64(__p0_689, __p1_689, __p2_689) __extension__ ({ \
58033  int32x2_t __s0_689 = __p0_689; \
58034  int64x2_t __s1_689 = __p1_689; \
58035  int32x4_t __ret_689; \
58036  __ret_689 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_689), (int32x2_t)(vqshrn_n_s64(__s1_689, __p2_689)))); \
58037  __ret_689; \
58038 })
58039 #else
58040 #define vqshrn_high_n_s64(__p0_690, __p1_690, __p2_690) __extension__ ({ \
58041  int32x2_t __s0_690 = __p0_690; \
58042  int64x2_t __s1_690 = __p1_690; \
58043  int32x2_t __rev0_690; __rev0_690 = __builtin_shufflevector(__s0_690, __s0_690, 1, 0); \
58044  int64x2_t __rev1_690; __rev1_690 = __builtin_shufflevector(__s1_690, __s1_690, 1, 0); \
58045  int32x4_t __ret_690; \
58046  __ret_690 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_690), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_690, __p2_690)))); \
58047  __ret_690 = __builtin_shufflevector(__ret_690, __ret_690, 3, 2, 1, 0); \
58048  __ret_690; \
58049 })
58050 #endif
58051 
58052 #ifdef __LITTLE_ENDIAN__
58053 #define vqshrn_high_n_s16(__p0_691, __p1_691, __p2_691) __extension__ ({ \
58054  int8x8_t __s0_691 = __p0_691; \
58055  int16x8_t __s1_691 = __p1_691; \
58056  int8x16_t __ret_691; \
58057  __ret_691 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_691), (int8x8_t)(vqshrn_n_s16(__s1_691, __p2_691)))); \
58058  __ret_691; \
58059 })
58060 #else
58061 #define vqshrn_high_n_s16(__p0_692, __p1_692, __p2_692) __extension__ ({ \
58062  int8x8_t __s0_692 = __p0_692; \
58063  int16x8_t __s1_692 = __p1_692; \
58064  int8x8_t __rev0_692; __rev0_692 = __builtin_shufflevector(__s0_692, __s0_692, 7, 6, 5, 4, 3, 2, 1, 0); \
58065  int16x8_t __rev1_692; __rev1_692 = __builtin_shufflevector(__s1_692, __s1_692, 7, 6, 5, 4, 3, 2, 1, 0); \
58066  int8x16_t __ret_692; \
58067  __ret_692 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_692), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_692, __p2_692)))); \
58068  __ret_692 = __builtin_shufflevector(__ret_692, __ret_692, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58069  __ret_692; \
58070 })
58071 #endif
58072 
58073 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58074  uint32_t __s0 = __p0; \
58075  uint16_t __ret; \
58076  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58077  __ret; \
58078 })
58079 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58080  uint64_t __s0 = __p0; \
58081  uint32_t __ret; \
58082  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58083  __ret; \
58084 })
58085 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58086  uint16_t __s0 = __p0; \
58087  uint8_t __ret; \
58088  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58089  __ret; \
58090 })
58091 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58092  int32_t __s0 = __p0; \
58093  int16_t __ret; \
58094  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58095  __ret; \
58096 })
58097 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58098  int64_t __s0 = __p0; \
58099  int32_t __ret; \
58100  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58101  __ret; \
58102 })
58103 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58104  int16_t __s0 = __p0; \
58105  int8_t __ret; \
58106  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58107  __ret; \
58108 })
58109 #ifdef __LITTLE_ENDIAN__
58110 #define vqshrun_high_n_s32(__p0_693, __p1_693, __p2_693) __extension__ ({ \
58111  int16x4_t __s0_693 = __p0_693; \
58112  int32x4_t __s1_693 = __p1_693; \
58113  int16x8_t __ret_693; \
58114  __ret_693 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_693), (int16x4_t)(vqshrun_n_s32(__s1_693, __p2_693)))); \
58115  __ret_693; \
58116 })
58117 #else
58118 #define vqshrun_high_n_s32(__p0_694, __p1_694, __p2_694) __extension__ ({ \
58119  int16x4_t __s0_694 = __p0_694; \
58120  int32x4_t __s1_694 = __p1_694; \
58121  int16x4_t __rev0_694; __rev0_694 = __builtin_shufflevector(__s0_694, __s0_694, 3, 2, 1, 0); \
58122  int32x4_t __rev1_694; __rev1_694 = __builtin_shufflevector(__s1_694, __s1_694, 3, 2, 1, 0); \
58123  int16x8_t __ret_694; \
58124  __ret_694 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_694), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_694, __p2_694)))); \
58125  __ret_694 = __builtin_shufflevector(__ret_694, __ret_694, 7, 6, 5, 4, 3, 2, 1, 0); \
58126  __ret_694; \
58127 })
58128 #endif
58129 
58130 #ifdef __LITTLE_ENDIAN__
58131 #define vqshrun_high_n_s64(__p0_695, __p1_695, __p2_695) __extension__ ({ \
58132  int32x2_t __s0_695 = __p0_695; \
58133  int64x2_t __s1_695 = __p1_695; \
58134  int32x4_t __ret_695; \
58135  __ret_695 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_695), (int32x2_t)(vqshrun_n_s64(__s1_695, __p2_695)))); \
58136  __ret_695; \
58137 })
58138 #else
58139 #define vqshrun_high_n_s64(__p0_696, __p1_696, __p2_696) __extension__ ({ \
58140  int32x2_t __s0_696 = __p0_696; \
58141  int64x2_t __s1_696 = __p1_696; \
58142  int32x2_t __rev0_696; __rev0_696 = __builtin_shufflevector(__s0_696, __s0_696, 1, 0); \
58143  int64x2_t __rev1_696; __rev1_696 = __builtin_shufflevector(__s1_696, __s1_696, 1, 0); \
58144  int32x4_t __ret_696; \
58145  __ret_696 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_696), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_696, __p2_696)))); \
58146  __ret_696 = __builtin_shufflevector(__ret_696, __ret_696, 3, 2, 1, 0); \
58147  __ret_696; \
58148 })
58149 #endif
58150 
58151 #ifdef __LITTLE_ENDIAN__
58152 #define vqshrun_high_n_s16(__p0_697, __p1_697, __p2_697) __extension__ ({ \
58153  int8x8_t __s0_697 = __p0_697; \
58154  int16x8_t __s1_697 = __p1_697; \
58155  int8x16_t __ret_697; \
58156  __ret_697 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_697), (int8x8_t)(vqshrun_n_s16(__s1_697, __p2_697)))); \
58157  __ret_697; \
58158 })
58159 #else
58160 #define vqshrun_high_n_s16(__p0_698, __p1_698, __p2_698) __extension__ ({ \
58161  int8x8_t __s0_698 = __p0_698; \
58162  int16x8_t __s1_698 = __p1_698; \
58163  int8x8_t __rev0_698; __rev0_698 = __builtin_shufflevector(__s0_698, __s0_698, 7, 6, 5, 4, 3, 2, 1, 0); \
58164  int16x8_t __rev1_698; __rev1_698 = __builtin_shufflevector(__s1_698, __s1_698, 7, 6, 5, 4, 3, 2, 1, 0); \
58165  int8x16_t __ret_698; \
58166  __ret_698 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_698), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_698, __p2_698)))); \
58167  __ret_698 = __builtin_shufflevector(__ret_698, __ret_698, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58168  __ret_698; \
58169 })
58170 #endif
58171 
58172 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58173  int32_t __s0 = __p0; \
58174  int16_t __ret; \
58175  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58176  __ret; \
58177 })
58178 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58179  int64_t __s0 = __p0; \
58180  int32_t __ret; \
58181  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58182  __ret; \
58183 })
58184 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58185  int16_t __s0 = __p0; \
58186  int8_t __ret; \
58187  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58188  __ret; \
58189 })
58190 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58191  uint8_t __ret;
58192  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58193  return __ret;
58194 }
58195 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58196  uint32_t __ret;
58197  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58198  return __ret;
58199 }
58200 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58201  uint64_t __ret;
58202  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58203  return __ret;
58204 }
58205 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58206  uint16_t __ret;
58207  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58208  return __ret;
58209 }
58210 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58211  int8_t __ret;
58212  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58213  return __ret;
58214 }
58215 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58216  int32_t __ret;
58217  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58218  return __ret;
58219 }
58220 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58221  int64_t __ret;
58222  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58223  return __ret;
58224 }
58225 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58226  int16_t __ret;
58227  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58228  return __ret;
58229 }
58230 #ifdef __LITTLE_ENDIAN__
58231 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58232  poly8x8_t __ret;
58233  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
58234  return __ret;
58235 }
58236 #else
58237 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58238  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58239  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58240  poly8x8_t __ret;
58241  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
58242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58243  return __ret;
58244 }
58245 #endif
58246 
58247 #ifdef __LITTLE_ENDIAN__
58248 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58249  poly8x16_t __ret;
58250  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
58251  return __ret;
58252 }
58253 #else
58254 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58255  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58256  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58257  poly8x16_t __ret;
58258  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
58259  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58260  return __ret;
58261 }
58262 #endif
58263 
58264 #ifdef __LITTLE_ENDIAN__
58265 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58266  uint8x16_t __ret;
58267  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
58268  return __ret;
58269 }
58270 #else
58271 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58272  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58273  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58274  uint8x16_t __ret;
58275  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
58276  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58277  return __ret;
58278 }
58279 #endif
58280 
58281 #ifdef __LITTLE_ENDIAN__
58282 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, uint8x16_t __p1) {
58283  int8x16_t __ret;
58284  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
58285  return __ret;
58286 }
58287 #else
58288 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, uint8x16_t __p1) {
58289  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58290  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58291  int8x16_t __ret;
58292  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
58293  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58294  return __ret;
58295 }
58296 #endif
58297 
58298 #ifdef __LITTLE_ENDIAN__
58299 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58300  uint8x8_t __ret;
58301  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
58302  return __ret;
58303 }
58304 #else
58305 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58306  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58307  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58308  uint8x8_t __ret;
58309  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
58310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58311  return __ret;
58312 }
58313 #endif
58314 
58315 #ifdef __LITTLE_ENDIAN__
58316 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, uint8x8_t __p1) {
58317  int8x8_t __ret;
58318  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
58319  return __ret;
58320 }
58321 #else
58322 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, uint8x8_t __p1) {
58323  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58324  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58325  int8x8_t __ret;
58326  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
58327  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58328  return __ret;
58329 }
58330 #endif
58331 
58332 #ifdef __LITTLE_ENDIAN__
58333 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
58334  poly8x8_t __ret;
58335  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
58336  return __ret;
58337 }
58338 #else
58339 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
58340  poly8x16x2_t __rev0;
58341  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58342  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58343  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58344  poly8x8_t __ret;
58345  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
58346  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58347  return __ret;
58348 }
58349 #endif
58350 
58351 #ifdef __LITTLE_ENDIAN__
58352 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
58353  poly8x16_t __ret;
58354  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
58355  return __ret;
58356 }
58357 #else
58358 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
58359  poly8x16x2_t __rev0;
58360  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58361  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58362  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58363  poly8x16_t __ret;
58364  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
58365  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58366  return __ret;
58367 }
58368 #endif
58369 
58370 #ifdef __LITTLE_ENDIAN__
58371 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
58372  uint8x16_t __ret;
58373  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
58374  return __ret;
58375 }
58376 #else
58377 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
58378  uint8x16x2_t __rev0;
58379  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58380  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58381  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58382  uint8x16_t __ret;
58383  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
58384  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58385  return __ret;
58386 }
58387 #endif
58388 
58389 #ifdef __LITTLE_ENDIAN__
58390 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, uint8x16_t __p1) {
58391  int8x16_t __ret;
58392  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
58393  return __ret;
58394 }
58395 #else
58396 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, uint8x16_t __p1) {
58397  int8x16x2_t __rev0;
58398  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58399  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58400  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58401  int8x16_t __ret;
58402  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
58403  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58404  return __ret;
58405 }
58406 #endif
58407 
58408 #ifdef __LITTLE_ENDIAN__
58409 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
58410  uint8x8_t __ret;
58411  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
58412  return __ret;
58413 }
58414 #else
58415 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
58416  uint8x16x2_t __rev0;
58417  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58418  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58419  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58420  uint8x8_t __ret;
58421  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
58422  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58423  return __ret;
58424 }
58425 #endif
58426 
58427 #ifdef __LITTLE_ENDIAN__
58428 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, uint8x8_t __p1) {
58429  int8x8_t __ret;
58430  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
58431  return __ret;
58432 }
58433 #else
58434 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, uint8x8_t __p1) {
58435  int8x16x2_t __rev0;
58436  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58437  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58438  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58439  int8x8_t __ret;
58440  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
58441  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58442  return __ret;
58443 }
58444 #endif
58445 
58446 #ifdef __LITTLE_ENDIAN__
58447 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
58448  poly8x8_t __ret;
58449  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
58450  return __ret;
58451 }
58452 #else
58453 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
58454  poly8x16x3_t __rev0;
58455  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58456  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58457  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58458  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58459  poly8x8_t __ret;
58460  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
58461  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58462  return __ret;
58463 }
58464 #endif
58465 
58466 #ifdef __LITTLE_ENDIAN__
58467 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
58468  poly8x16_t __ret;
58469  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
58470  return __ret;
58471 }
58472 #else
58473 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
58474  poly8x16x3_t __rev0;
58475  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58476  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58477  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58478  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58479  poly8x16_t __ret;
58480  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
58481  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58482  return __ret;
58483 }
58484 #endif
58485 
58486 #ifdef __LITTLE_ENDIAN__
58487 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
58488  uint8x16_t __ret;
58489  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
58490  return __ret;
58491 }
58492 #else
58493 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
58494  uint8x16x3_t __rev0;
58495  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58496  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58497  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58498  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58499  uint8x16_t __ret;
58500  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
58501  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58502  return __ret;
58503 }
58504 #endif
58505 
58506 #ifdef __LITTLE_ENDIAN__
58507 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, uint8x16_t __p1) {
58508  int8x16_t __ret;
58509  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
58510  return __ret;
58511 }
58512 #else
58513 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, uint8x16_t __p1) {
58514  int8x16x3_t __rev0;
58515  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58516  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58517  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58518  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58519  int8x16_t __ret;
58520  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
58521  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58522  return __ret;
58523 }
58524 #endif
58525 
58526 #ifdef __LITTLE_ENDIAN__
58527 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
58528  uint8x8_t __ret;
58529  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
58530  return __ret;
58531 }
58532 #else
58533 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
58534  uint8x16x3_t __rev0;
58535  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58536  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58537  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58538  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58539  uint8x8_t __ret;
58540  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
58541  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58542  return __ret;
58543 }
58544 #endif
58545 
58546 #ifdef __LITTLE_ENDIAN__
58547 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, uint8x8_t __p1) {
58548  int8x8_t __ret;
58549  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
58550  return __ret;
58551 }
58552 #else
58553 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, uint8x8_t __p1) {
58554  int8x16x3_t __rev0;
58555  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58556  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58557  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58558  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58559  int8x8_t __ret;
58560  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
58561  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58562  return __ret;
58563 }
58564 #endif
58565 
58566 #ifdef __LITTLE_ENDIAN__
58567 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
58568  poly8x8_t __ret;
58569  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
58570  return __ret;
58571 }
58572 #else
58573 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
58574  poly8x16x4_t __rev0;
58575  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58576  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58577  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58578  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58579  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58580  poly8x8_t __ret;
58581  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
58582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58583  return __ret;
58584 }
58585 #endif
58586 
58587 #ifdef __LITTLE_ENDIAN__
58588 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
58589  poly8x16_t __ret;
58590  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
58591  return __ret;
58592 }
58593 #else
58594 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
58595  poly8x16x4_t __rev0;
58596  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58597  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58598  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58599  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58600  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58601  poly8x16_t __ret;
58602  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
58603  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58604  return __ret;
58605 }
58606 #endif
58607 
58608 #ifdef __LITTLE_ENDIAN__
58609 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
58610  uint8x16_t __ret;
58611  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
58612  return __ret;
58613 }
58614 #else
58615 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
58616  uint8x16x4_t __rev0;
58617  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58618  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58619  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58620  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58621  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58622  uint8x16_t __ret;
58623  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
58624  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58625  return __ret;
58626 }
58627 #endif
58628 
58629 #ifdef __LITTLE_ENDIAN__
58630 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, uint8x16_t __p1) {
58631  int8x16_t __ret;
58632  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
58633  return __ret;
58634 }
58635 #else
58636 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, uint8x16_t __p1) {
58637  int8x16x4_t __rev0;
58638  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58639  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58640  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58641  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58642  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58643  int8x16_t __ret;
58644  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
58645  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58646  return __ret;
58647 }
58648 #endif
58649 
58650 #ifdef __LITTLE_ENDIAN__
58651 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
58652  uint8x8_t __ret;
58653  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
58654  return __ret;
58655 }
58656 #else
58657 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
58658  uint8x16x4_t __rev0;
58659  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58660  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58661  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58662  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58663  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58664  uint8x8_t __ret;
58665  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
58666  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58667  return __ret;
58668 }
58669 #endif
58670 
58671 #ifdef __LITTLE_ENDIAN__
58672 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, uint8x8_t __p1) {
58673  int8x8_t __ret;
58674  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
58675  return __ret;
58676 }
58677 #else
58678 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, uint8x8_t __p1) {
58679  int8x16x4_t __rev0;
58680  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58681  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58682  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58683  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58684  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58685  int8x8_t __ret;
58686  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
58687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58688  return __ret;
58689 }
58690 #endif
58691 
58692 #ifdef __LITTLE_ENDIAN__
58693 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
58694  poly8x8_t __ret;
58695  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
58696  return __ret;
58697 }
58698 #else
58699 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
58700  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58701  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58702  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58703  poly8x8_t __ret;
58704  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
58705  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58706  return __ret;
58707 }
58708 #endif
58709 
58710 #ifdef __LITTLE_ENDIAN__
58711 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
58712  poly8x16_t __ret;
58713  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
58714  return __ret;
58715 }
58716 #else
58717 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
58718  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58719  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58720  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58721  poly8x16_t __ret;
58722  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
58723  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58724  return __ret;
58725 }
58726 #endif
58727 
58728 #ifdef __LITTLE_ENDIAN__
58729 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
58730  uint8x16_t __ret;
58731  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
58732  return __ret;
58733 }
58734 #else
58735 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
58736  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58737  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58738  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58739  uint8x16_t __ret;
58740  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
58741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58742  return __ret;
58743 }
58744 #endif
58745 
58746 #ifdef __LITTLE_ENDIAN__
58747 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, uint8x16_t __p2) {
58748  int8x16_t __ret;
58749  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
58750  return __ret;
58751 }
58752 #else
58753 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, uint8x16_t __p2) {
58754  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58755  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58756  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58757  int8x16_t __ret;
58758  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
58759  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58760  return __ret;
58761 }
58762 #endif
58763 
58764 #ifdef __LITTLE_ENDIAN__
58765 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
58766  uint8x8_t __ret;
58767  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
58768  return __ret;
58769 }
58770 #else
58771 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
58772  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58773  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58774  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58775  uint8x8_t __ret;
58776  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
58777  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58778  return __ret;
58779 }
58780 #endif
58781 
58782 #ifdef __LITTLE_ENDIAN__
58783 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, uint8x8_t __p2) {
58784  int8x8_t __ret;
58785  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
58786  return __ret;
58787 }
58788 #else
58789 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, uint8x8_t __p2) {
58790  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58791  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58792  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58793  int8x8_t __ret;
58794  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
58795  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58796  return __ret;
58797 }
58798 #endif
58799 
58800 #ifdef __LITTLE_ENDIAN__
58801 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
58802  poly8x8_t __ret;
58803  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
58804  return __ret;
58805 }
58806 #else
58807 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
58808  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58809  poly8x16x2_t __rev1;
58810  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58811  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58812  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58813  poly8x8_t __ret;
58814  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
58815  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58816  return __ret;
58817 }
58818 #endif
58819 
58820 #ifdef __LITTLE_ENDIAN__
58821 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
58822  poly8x16_t __ret;
58823  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
58824  return __ret;
58825 }
58826 #else
58827 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
58828  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58829  poly8x16x2_t __rev1;
58830  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58831  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58832  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58833  poly8x16_t __ret;
58834  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
58835  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58836  return __ret;
58837 }
58838 #endif
58839 
58840 #ifdef __LITTLE_ENDIAN__
58841 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
58842  uint8x16_t __ret;
58843  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
58844  return __ret;
58845 }
58846 #else
58847 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
58848  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58849  uint8x16x2_t __rev1;
58850  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58851  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58852  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58853  uint8x16_t __ret;
58854  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
58855  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58856  return __ret;
58857 }
58858 #endif
58859 
58860 #ifdef __LITTLE_ENDIAN__
58861 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, uint8x16_t __p2) {
58862  int8x16_t __ret;
58863  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
58864  return __ret;
58865 }
58866 #else
58867 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, uint8x16_t __p2) {
58868  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58869  int8x16x2_t __rev1;
58870  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58871  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58872  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58873  int8x16_t __ret;
58874  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
58875  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58876  return __ret;
58877 }
58878 #endif
58879 
58880 #ifdef __LITTLE_ENDIAN__
58881 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
58882  uint8x8_t __ret;
58883  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
58884  return __ret;
58885 }
58886 #else
58887 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
58888  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58889  uint8x16x2_t __rev1;
58890  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58891  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58892  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58893  uint8x8_t __ret;
58894  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
58895  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58896  return __ret;
58897 }
58898 #endif
58899 
58900 #ifdef __LITTLE_ENDIAN__
58901 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, uint8x8_t __p2) {
58902  int8x8_t __ret;
58903  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
58904  return __ret;
58905 }
58906 #else
58907 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, uint8x8_t __p2) {
58908  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58909  int8x16x2_t __rev1;
58910  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58911  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58912  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58913  int8x8_t __ret;
58914  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
58915  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58916  return __ret;
58917 }
58918 #endif
58919 
58920 #ifdef __LITTLE_ENDIAN__
58921 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
58922  poly8x8_t __ret;
58923  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
58924  return __ret;
58925 }
58926 #else
58927 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
58928  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58929  poly8x16x3_t __rev1;
58930  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58931  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58932  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58933  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58934  poly8x8_t __ret;
58935  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
58936  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58937  return __ret;
58938 }
58939 #endif
58940 
58941 #ifdef __LITTLE_ENDIAN__
58942 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
58943  poly8x16_t __ret;
58944  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
58945  return __ret;
58946 }
58947 #else
58948 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
58949  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58950  poly8x16x3_t __rev1;
58951  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58952  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58953  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58954  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58955  poly8x16_t __ret;
58956  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
58957  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58958  return __ret;
58959 }
58960 #endif
58961 
58962 #ifdef __LITTLE_ENDIAN__
58963 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
58964  uint8x16_t __ret;
58965  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
58966  return __ret;
58967 }
58968 #else
58969 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
58970  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58971  uint8x16x3_t __rev1;
58972  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58973  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58974  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58975  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58976  uint8x16_t __ret;
58977  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
58978  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58979  return __ret;
58980 }
58981 #endif
58982 
58983 #ifdef __LITTLE_ENDIAN__
58984 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, uint8x16_t __p2) {
58985  int8x16_t __ret;
58986  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
58987  return __ret;
58988 }
58989 #else
58990 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, uint8x16_t __p2) {
58991  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58992  int8x16x3_t __rev1;
58993  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58994  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58995  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58996  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58997  int8x16_t __ret;
58998  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
58999  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59000  return __ret;
59001 }
59002 #endif
59003 
59004 #ifdef __LITTLE_ENDIAN__
59005 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59006  uint8x8_t __ret;
59007  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
59008  return __ret;
59009 }
59010 #else
59011 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59012  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59013  uint8x16x3_t __rev1;
59014  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59015  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59016  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59017  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59018  uint8x8_t __ret;
59019  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
59020  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59021  return __ret;
59022 }
59023 #endif
59024 
59025 #ifdef __LITTLE_ENDIAN__
59026 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, uint8x8_t __p2) {
59027  int8x8_t __ret;
59028  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
59029  return __ret;
59030 }
59031 #else
59032 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, uint8x8_t __p2) {
59033  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59034  int8x16x3_t __rev1;
59035  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59036  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59037  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59038  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59039  int8x8_t __ret;
59040  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
59041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59042  return __ret;
59043 }
59044 #endif
59045 
59046 #ifdef __LITTLE_ENDIAN__
59047 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59048  poly8x8_t __ret;
59049  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
59050  return __ret;
59051 }
59052 #else
59053 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59054  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59055  poly8x16x4_t __rev1;
59056  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59057  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59058  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59059  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59060  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59061  poly8x8_t __ret;
59062  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
59063  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59064  return __ret;
59065 }
59066 #endif
59067 
59068 #ifdef __LITTLE_ENDIAN__
59069 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59070  poly8x16_t __ret;
59071  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
59072  return __ret;
59073 }
59074 #else
59075 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59076  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59077  poly8x16x4_t __rev1;
59078  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59079  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59080  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59081  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59082  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59083  poly8x16_t __ret;
59084  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
59085  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59086  return __ret;
59087 }
59088 #endif
59089 
59090 #ifdef __LITTLE_ENDIAN__
59091 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59092  uint8x16_t __ret;
59093  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
59094  return __ret;
59095 }
59096 #else
59097 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59098  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59099  uint8x16x4_t __rev1;
59100  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59101  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59102  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59103  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59104  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59105  uint8x16_t __ret;
59106  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
59107  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59108  return __ret;
59109 }
59110 #endif
59111 
59112 #ifdef __LITTLE_ENDIAN__
59113 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, uint8x16_t __p2) {
59114  int8x16_t __ret;
59115  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
59116  return __ret;
59117 }
59118 #else
59119 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, uint8x16_t __p2) {
59120  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59121  int8x16x4_t __rev1;
59122  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59123  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59124  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59125  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59126  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59127  int8x16_t __ret;
59128  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
59129  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59130  return __ret;
59131 }
59132 #endif
59133 
59134 #ifdef __LITTLE_ENDIAN__
59135 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59136  uint8x8_t __ret;
59137  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
59138  return __ret;
59139 }
59140 #else
59141 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59142  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59143  uint8x16x4_t __rev1;
59144  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59145  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59146  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59147  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59148  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59149  uint8x8_t __ret;
59150  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
59151  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59152  return __ret;
59153 }
59154 #endif
59155 
59156 #ifdef __LITTLE_ENDIAN__
59157 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, uint8x8_t __p2) {
59158  int8x8_t __ret;
59159  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
59160  return __ret;
59161 }
59162 #else
59163 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, uint8x8_t __p2) {
59164  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59165  int8x16x4_t __rev1;
59166  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59167  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59168  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59169  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59170  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59171  int8x8_t __ret;
59172  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
59173  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59174  return __ret;
59175 }
59176 #endif
59177 
59178 #ifdef __LITTLE_ENDIAN__
59179 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59180  uint16x8_t __ret;
59181  __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
59182  return __ret;
59183 }
59184 #else
59185 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59186  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59187  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59188  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59189  uint16x8_t __ret;
59190  __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
59191  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59192  return __ret;
59193 }
59194 #endif
59195 
59196 #ifdef __LITTLE_ENDIAN__
59197 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59198  uint32x4_t __ret;
59199  __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
59200  return __ret;
59201 }
59202 #else
59203 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59204  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59205  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59206  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59207  uint32x4_t __ret;
59208  __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
59209  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59210  return __ret;
59211 }
59212 #endif
59213 
59214 #ifdef __LITTLE_ENDIAN__
59215 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59216  uint8x16_t __ret;
59217  __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
59218  return __ret;
59219 }
59220 #else
59221 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59222  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59223  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59224  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59225  uint8x16_t __ret;
59226  __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
59227  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59228  return __ret;
59229 }
59230 #endif
59231 
59232 #ifdef __LITTLE_ENDIAN__
59233 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59234  int16x8_t __ret;
59235  __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
59236  return __ret;
59237 }
59238 #else
59239 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59240  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59241  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59242  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59243  int16x8_t __ret;
59244  __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
59245  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59246  return __ret;
59247 }
59248 #endif
59249 
59250 #ifdef __LITTLE_ENDIAN__
59251 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59252  int32x4_t __ret;
59253  __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
59254  return __ret;
59255 }
59256 #else
59257 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59258  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59259  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59260  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59261  int32x4_t __ret;
59262  __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
59263  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59264  return __ret;
59265 }
59266 #endif
59267 
59268 #ifdef __LITTLE_ENDIAN__
59269 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59270  int8x16_t __ret;
59271  __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
59272  return __ret;
59273 }
59274 #else
59275 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59276  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59277  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59278  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59279  int8x16_t __ret;
59280  __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
59281  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59282  return __ret;
59283 }
59284 #endif
59285 
59286 #ifdef __LITTLE_ENDIAN__
59287 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59288  poly8x8_t __ret;
59289  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
59290  return __ret;
59291 }
59292 #else
59293 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59294  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59295  poly8x8_t __ret;
59296  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
59297  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59298  return __ret;
59299 }
59300 #endif
59301 
59302 #ifdef __LITTLE_ENDIAN__
59303 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59304  poly8x16_t __ret;
59305  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
59306  return __ret;
59307 }
59308 #else
59309 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59310  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59311  poly8x16_t __ret;
59312  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
59313  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59314  return __ret;
59315 }
59316 #endif
59317 
59318 #ifdef __LITTLE_ENDIAN__
59319 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
59320  uint8x16_t __ret;
59321  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
59322  return __ret;
59323 }
59324 #else
59325 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
59326  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59327  uint8x16_t __ret;
59328  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
59329  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59330  return __ret;
59331 }
59332 #endif
59333 
59334 #ifdef __LITTLE_ENDIAN__
59335 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
59336  int8x16_t __ret;
59337  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
59338  return __ret;
59339 }
59340 #else
59341 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
59342  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59343  int8x16_t __ret;
59344  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
59345  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59346  return __ret;
59347 }
59348 #endif
59349 
59350 #ifdef __LITTLE_ENDIAN__
59351 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
59352  uint8x8_t __ret;
59353  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
59354  return __ret;
59355 }
59356 #else
59357 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
59358  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59359  uint8x8_t __ret;
59360  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
59361  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59362  return __ret;
59363 }
59364 #endif
59365 
59366 #ifdef __LITTLE_ENDIAN__
59367 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
59368  int8x8_t __ret;
59369  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
59370  return __ret;
59371 }
59372 #else
59373 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
59374  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59375  int8x8_t __ret;
59376  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
59377  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59378  return __ret;
59379 }
59380 #endif
59381 
59382 #ifdef __LITTLE_ENDIAN__
59383 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
59384  float64x2_t __ret;
59385  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
59386  return __ret;
59387 }
59388 #else
59389 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
59390  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59391  float64x2_t __ret;
59392  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
59393  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59394  return __ret;
59395 }
59396 #endif
59397 
59398 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
59399  float64x1_t __ret;
59400  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
59401  return __ret;
59402 }
59403 __ai float64_t vrecped_f64(float64_t __p0) {
59404  float64_t __ret;
59405  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
59406  return __ret;
59407 }
59408 __ai float32_t vrecpes_f32(float32_t __p0) {
59409  float32_t __ret;
59410  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
59411  return __ret;
59412 }
59413 #ifdef __LITTLE_ENDIAN__
59414 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
59415  float64x2_t __ret;
59416  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
59417  return __ret;
59418 }
59419 #else
59420 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
59421  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59422  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59423  float64x2_t __ret;
59424  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
59425  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59426  return __ret;
59427 }
59428 #endif
59429 
59430 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
59431  float64x1_t __ret;
59432  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59433  return __ret;
59434 }
59435 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
59436  float64_t __ret;
59437  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
59438  return __ret;
59439 }
59440 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
59441  float32_t __ret;
59442  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
59443  return __ret;
59444 }
59445 __ai float64_t vrecpxd_f64(float64_t __p0) {
59446  float64_t __ret;
59447  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
59448  return __ret;
59449 }
59450 __ai float32_t vrecpxs_f32(float32_t __p0) {
59451  float32_t __ret;
59452  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
59453  return __ret;
59454 }
59455 __ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
59456  uint64_t __ret;
59457  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
59458  return __ret;
59459 }
59460 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
59461  int64_t __ret;
59462  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
59463  return __ret;
59464 }
59465 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
59466  uint64_t __s0 = __p0; \
59467  uint64_t __ret; \
59468  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
59469  __ret; \
59470 })
59471 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
59472  int64_t __s0 = __p0; \
59473  int64_t __ret; \
59474  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
59475  __ret; \
59476 })
59477 #ifdef __LITTLE_ENDIAN__
59478 #define vrshrn_high_n_u32(__p0_699, __p1_699, __p2_699) __extension__ ({ \
59479  uint16x4_t __s0_699 = __p0_699; \
59480  uint32x4_t __s1_699 = __p1_699; \
59481  uint16x8_t __ret_699; \
59482  __ret_699 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_699), (uint16x4_t)(vrshrn_n_u32(__s1_699, __p2_699)))); \
59483  __ret_699; \
59484 })
59485 #else
59486 #define vrshrn_high_n_u32(__p0_700, __p1_700, __p2_700) __extension__ ({ \
59487  uint16x4_t __s0_700 = __p0_700; \
59488  uint32x4_t __s1_700 = __p1_700; \
59489  uint16x4_t __rev0_700; __rev0_700 = __builtin_shufflevector(__s0_700, __s0_700, 3, 2, 1, 0); \
59490  uint32x4_t __rev1_700; __rev1_700 = __builtin_shufflevector(__s1_700, __s1_700, 3, 2, 1, 0); \
59491  uint16x8_t __ret_700; \
59492  __ret_700 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_700), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_700, __p2_700)))); \
59493  __ret_700 = __builtin_shufflevector(__ret_700, __ret_700, 7, 6, 5, 4, 3, 2, 1, 0); \
59494  __ret_700; \
59495 })
59496 #endif
59497 
59498 #ifdef __LITTLE_ENDIAN__
59499 #define vrshrn_high_n_u64(__p0_701, __p1_701, __p2_701) __extension__ ({ \
59500  uint32x2_t __s0_701 = __p0_701; \
59501  uint64x2_t __s1_701 = __p1_701; \
59502  uint32x4_t __ret_701; \
59503  __ret_701 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_701), (uint32x2_t)(vrshrn_n_u64(__s1_701, __p2_701)))); \
59504  __ret_701; \
59505 })
59506 #else
59507 #define vrshrn_high_n_u64(__p0_702, __p1_702, __p2_702) __extension__ ({ \
59508  uint32x2_t __s0_702 = __p0_702; \
59509  uint64x2_t __s1_702 = __p1_702; \
59510  uint32x2_t __rev0_702; __rev0_702 = __builtin_shufflevector(__s0_702, __s0_702, 1, 0); \
59511  uint64x2_t __rev1_702; __rev1_702 = __builtin_shufflevector(__s1_702, __s1_702, 1, 0); \
59512  uint32x4_t __ret_702; \
59513  __ret_702 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_702), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_702, __p2_702)))); \
59514  __ret_702 = __builtin_shufflevector(__ret_702, __ret_702, 3, 2, 1, 0); \
59515  __ret_702; \
59516 })
59517 #endif
59518 
59519 #ifdef __LITTLE_ENDIAN__
59520 #define vrshrn_high_n_u16(__p0_703, __p1_703, __p2_703) __extension__ ({ \
59521  uint8x8_t __s0_703 = __p0_703; \
59522  uint16x8_t __s1_703 = __p1_703; \
59523  uint8x16_t __ret_703; \
59524  __ret_703 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_703), (uint8x8_t)(vrshrn_n_u16(__s1_703, __p2_703)))); \
59525  __ret_703; \
59526 })
59527 #else
59528 #define vrshrn_high_n_u16(__p0_704, __p1_704, __p2_704) __extension__ ({ \
59529  uint8x8_t __s0_704 = __p0_704; \
59530  uint16x8_t __s1_704 = __p1_704; \
59531  uint8x8_t __rev0_704; __rev0_704 = __builtin_shufflevector(__s0_704, __s0_704, 7, 6, 5, 4, 3, 2, 1, 0); \
59532  uint16x8_t __rev1_704; __rev1_704 = __builtin_shufflevector(__s1_704, __s1_704, 7, 6, 5, 4, 3, 2, 1, 0); \
59533  uint8x16_t __ret_704; \
59534  __ret_704 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_704), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_704, __p2_704)))); \
59535  __ret_704 = __builtin_shufflevector(__ret_704, __ret_704, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59536  __ret_704; \
59537 })
59538 #endif
59539 
59540 #ifdef __LITTLE_ENDIAN__
59541 #define vrshrn_high_n_s32(__p0_705, __p1_705, __p2_705) __extension__ ({ \
59542  int16x4_t __s0_705 = __p0_705; \
59543  int32x4_t __s1_705 = __p1_705; \
59544  int16x8_t __ret_705; \
59545  __ret_705 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_705), (int16x4_t)(vrshrn_n_s32(__s1_705, __p2_705)))); \
59546  __ret_705; \
59547 })
59548 #else
59549 #define vrshrn_high_n_s32(__p0_706, __p1_706, __p2_706) __extension__ ({ \
59550  int16x4_t __s0_706 = __p0_706; \
59551  int32x4_t __s1_706 = __p1_706; \
59552  int16x4_t __rev0_706; __rev0_706 = __builtin_shufflevector(__s0_706, __s0_706, 3, 2, 1, 0); \
59553  int32x4_t __rev1_706; __rev1_706 = __builtin_shufflevector(__s1_706, __s1_706, 3, 2, 1, 0); \
59554  int16x8_t __ret_706; \
59555  __ret_706 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_706), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_706, __p2_706)))); \
59556  __ret_706 = __builtin_shufflevector(__ret_706, __ret_706, 7, 6, 5, 4, 3, 2, 1, 0); \
59557  __ret_706; \
59558 })
59559 #endif
59560 
59561 #ifdef __LITTLE_ENDIAN__
59562 #define vrshrn_high_n_s64(__p0_707, __p1_707, __p2_707) __extension__ ({ \
59563  int32x2_t __s0_707 = __p0_707; \
59564  int64x2_t __s1_707 = __p1_707; \
59565  int32x4_t __ret_707; \
59566  __ret_707 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_707), (int32x2_t)(vrshrn_n_s64(__s1_707, __p2_707)))); \
59567  __ret_707; \
59568 })
59569 #else
59570 #define vrshrn_high_n_s64(__p0_708, __p1_708, __p2_708) __extension__ ({ \
59571  int32x2_t __s0_708 = __p0_708; \
59572  int64x2_t __s1_708 = __p1_708; \
59573  int32x2_t __rev0_708; __rev0_708 = __builtin_shufflevector(__s0_708, __s0_708, 1, 0); \
59574  int64x2_t __rev1_708; __rev1_708 = __builtin_shufflevector(__s1_708, __s1_708, 1, 0); \
59575  int32x4_t __ret_708; \
59576  __ret_708 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_708), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_708, __p2_708)))); \
59577  __ret_708 = __builtin_shufflevector(__ret_708, __ret_708, 3, 2, 1, 0); \
59578  __ret_708; \
59579 })
59580 #endif
59581 
59582 #ifdef __LITTLE_ENDIAN__
59583 #define vrshrn_high_n_s16(__p0_709, __p1_709, __p2_709) __extension__ ({ \
59584  int8x8_t __s0_709 = __p0_709; \
59585  int16x8_t __s1_709 = __p1_709; \
59586  int8x16_t __ret_709; \
59587  __ret_709 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_709), (int8x8_t)(vrshrn_n_s16(__s1_709, __p2_709)))); \
59588  __ret_709; \
59589 })
59590 #else
59591 #define vrshrn_high_n_s16(__p0_710, __p1_710, __p2_710) __extension__ ({ \
59592  int8x8_t __s0_710 = __p0_710; \
59593  int16x8_t __s1_710 = __p1_710; \
59594  int8x8_t __rev0_710; __rev0_710 = __builtin_shufflevector(__s0_710, __s0_710, 7, 6, 5, 4, 3, 2, 1, 0); \
59595  int16x8_t __rev1_710; __rev1_710 = __builtin_shufflevector(__s1_710, __s1_710, 7, 6, 5, 4, 3, 2, 1, 0); \
59596  int8x16_t __ret_710; \
59597  __ret_710 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_710), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_710, __p2_710)))); \
59598  __ret_710 = __builtin_shufflevector(__ret_710, __ret_710, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59599  __ret_710; \
59600 })
59601 #endif
59602 
59603 #ifdef __LITTLE_ENDIAN__
59604 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
59605  float64x2_t __ret;
59606  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
59607  return __ret;
59608 }
59609 #else
59610 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
59611  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59612  float64x2_t __ret;
59613  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
59614  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59615  return __ret;
59616 }
59617 #endif
59618 
59619 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
59620  float64x1_t __ret;
59621  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
59622  return __ret;
59623 }
59624 __ai float64_t vrsqrted_f64(float64_t __p0) {
59625  float64_t __ret;
59626  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
59627  return __ret;
59628 }
59629 __ai float32_t vrsqrtes_f32(float32_t __p0) {
59630  float32_t __ret;
59631  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
59632  return __ret;
59633 }
59634 #ifdef __LITTLE_ENDIAN__
59635 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
59636  float64x2_t __ret;
59637  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
59638  return __ret;
59639 }
59640 #else
59641 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
59642  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59643  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59644  float64x2_t __ret;
59645  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
59646  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59647  return __ret;
59648 }
59649 #endif
59650 
59651 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
59652  float64x1_t __ret;
59653  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59654  return __ret;
59655 }
59656 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
59657  float64_t __ret;
59658  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
59659  return __ret;
59660 }
59661 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
59662  float32_t __ret;
59663  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
59664  return __ret;
59665 }
59666 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
59667  uint64_t __s0 = __p0; \
59668  uint64_t __s1 = __p1; \
59669  uint64_t __ret; \
59670  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
59671  __ret; \
59672 })
59673 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
59674  int64_t __s0 = __p0; \
59675  int64_t __s1 = __p1; \
59676  int64_t __ret; \
59677  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
59678  __ret; \
59679 })
59680 #ifdef __LITTLE_ENDIAN__
59681 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59682  uint16x8_t __ret;
59683  __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
59684  return __ret;
59685 }
59686 #else
59687 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59688  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59689  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59690  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59691  uint16x8_t __ret;
59692  __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
59693  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59694  return __ret;
59695 }
59696 #endif
59697 
59698 #ifdef __LITTLE_ENDIAN__
59699 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59700  uint32x4_t __ret;
59701  __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
59702  return __ret;
59703 }
59704 #else
59705 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59706  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59707  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59708  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59709  uint32x4_t __ret;
59710  __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
59711  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59712  return __ret;
59713 }
59714 #endif
59715 
59716 #ifdef __LITTLE_ENDIAN__
59717 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59718  uint8x16_t __ret;
59719  __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
59720  return __ret;
59721 }
59722 #else
59723 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59724  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59725  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59726  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59727  uint8x16_t __ret;
59728  __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
59729  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59730  return __ret;
59731 }
59732 #endif
59733 
59734 #ifdef __LITTLE_ENDIAN__
59735 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59736  int16x8_t __ret;
59737  __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
59738  return __ret;
59739 }
59740 #else
59741 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59742  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59743  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59744  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59745  int16x8_t __ret;
59746  __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
59747  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59748  return __ret;
59749 }
59750 #endif
59751 
59752 #ifdef __LITTLE_ENDIAN__
59753 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59754  int32x4_t __ret;
59755  __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
59756  return __ret;
59757 }
59758 #else
59759 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59760  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59761  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59762  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59763  int32x4_t __ret;
59764  __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
59765  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59766  return __ret;
59767 }
59768 #endif
59769 
59770 #ifdef __LITTLE_ENDIAN__
59771 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59772  int8x16_t __ret;
59773  __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
59774  return __ret;
59775 }
59776 #else
59777 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59778  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59779  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59780  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59781  int8x16_t __ret;
59782  __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
59783  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59784  return __ret;
59785 }
59786 #endif
59787 
59788 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59789  poly64_t __s0 = __p0; \
59790  poly64x1_t __s1 = __p1; \
59791  poly64x1_t __ret; \
59792  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (poly64x1_t)__s1, __p2); \
59793  __ret; \
59794 })
59795 #ifdef __LITTLE_ENDIAN__
59796 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59797  poly64_t __s0 = __p0; \
59798  poly64x2_t __s1 = __p1; \
59799  poly64x2_t __ret; \
59800  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__s1, __p2); \
59801  __ret; \
59802 })
59803 #else
59804 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59805  poly64_t __s0 = __p0; \
59806  poly64x2_t __s1 = __p1; \
59807  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
59808  poly64x2_t __ret; \
59809  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__rev1, __p2); \
59810  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
59811  __ret; \
59812 })
59813 #define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
59814  poly64_t __s0 = __p0; \
59815  poly64x2_t __s1 = __p1; \
59816  poly64x2_t __ret; \
59817  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (poly64x2_t)__s1, __p2); \
59818  __ret; \
59819 })
59820 #endif
59821 
59822 #ifdef __LITTLE_ENDIAN__
59823 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59824  float64_t __s0 = __p0; \
59825  float64x2_t __s1 = __p1; \
59826  float64x2_t __ret; \
59827  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__s1, __p2); \
59828  __ret; \
59829 })
59830 #else
59831 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59832  float64_t __s0 = __p0; \
59833  float64x2_t __s1 = __p1; \
59834  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
59835  float64x2_t __ret; \
59836  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__rev1, __p2); \
59837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
59838  __ret; \
59839 })
59840 #define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59841  float64_t __s0 = __p0; \
59842  float64x2_t __s1 = __p1; \
59843  float64x2_t __ret; \
59844  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (float64x2_t)__s1, __p2); \
59845  __ret; \
59846 })
59847 #endif
59848 
59849 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
59850  float64_t __s0 = __p0; \
59851  float64x1_t __s1 = __p1; \
59852  float64x1_t __ret; \
59853  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (float64x1_t)__s1, __p2); \
59854  __ret; \
59855 })
59856 __ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
59857  uint64_t __ret;
59858  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
59859  return __ret;
59860 }
59861 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
59862  int64_t __ret;
59863  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
59864  return __ret;
59865 }
59866 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
59867  uint64_t __s0 = __p0; \
59868  uint64_t __ret; \
59869  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
59870  __ret; \
59871 })
59872 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
59873  int64_t __s0 = __p0; \
59874  int64_t __ret; \
59875  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
59876  __ret; \
59877 })
59878 #ifdef __LITTLE_ENDIAN__
59879 #define vshll_high_n_u8(__p0_711, __p1_711) __extension__ ({ \
59880  uint8x16_t __s0_711 = __p0_711; \
59881  uint16x8_t __ret_711; \
59882  __ret_711 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_711), __p1_711)); \
59883  __ret_711; \
59884 })
59885 #else
59886 #define vshll_high_n_u8(__p0_712, __p1_712) __extension__ ({ \
59887  uint8x16_t __s0_712 = __p0_712; \
59888  uint8x16_t __rev0_712; __rev0_712 = __builtin_shufflevector(__s0_712, __s0_712, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59889  uint16x8_t __ret_712; \
59890  __ret_712 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_712), __p1_712)); \
59891  __ret_712 = __builtin_shufflevector(__ret_712, __ret_712, 7, 6, 5, 4, 3, 2, 1, 0); \
59892  __ret_712; \
59893 })
59894 #endif
59895 
59896 #ifdef __LITTLE_ENDIAN__
59897 #define vshll_high_n_u32(__p0_713, __p1_713) __extension__ ({ \
59898  uint32x4_t __s0_713 = __p0_713; \
59899  uint64x2_t __ret_713; \
59900  __ret_713 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_713), __p1_713)); \
59901  __ret_713; \
59902 })
59903 #else
59904 #define vshll_high_n_u32(__p0_714, __p1_714) __extension__ ({ \
59905  uint32x4_t __s0_714 = __p0_714; \
59906  uint32x4_t __rev0_714; __rev0_714 = __builtin_shufflevector(__s0_714, __s0_714, 3, 2, 1, 0); \
59907  uint64x2_t __ret_714; \
59908  __ret_714 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_714), __p1_714)); \
59909  __ret_714 = __builtin_shufflevector(__ret_714, __ret_714, 1, 0); \
59910  __ret_714; \
59911 })
59912 #endif
59913 
59914 #ifdef __LITTLE_ENDIAN__
59915 #define vshll_high_n_u16(__p0_715, __p1_715) __extension__ ({ \
59916  uint16x8_t __s0_715 = __p0_715; \
59917  uint32x4_t __ret_715; \
59918  __ret_715 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_715), __p1_715)); \
59919  __ret_715; \
59920 })
59921 #else
59922 #define vshll_high_n_u16(__p0_716, __p1_716) __extension__ ({ \
59923  uint16x8_t __s0_716 = __p0_716; \
59924  uint16x8_t __rev0_716; __rev0_716 = __builtin_shufflevector(__s0_716, __s0_716, 7, 6, 5, 4, 3, 2, 1, 0); \
59925  uint32x4_t __ret_716; \
59926  __ret_716 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_716), __p1_716)); \
59927  __ret_716 = __builtin_shufflevector(__ret_716, __ret_716, 3, 2, 1, 0); \
59928  __ret_716; \
59929 })
59930 #endif
59931 
59932 #ifdef __LITTLE_ENDIAN__
59933 #define vshll_high_n_s8(__p0_717, __p1_717) __extension__ ({ \
59934  int8x16_t __s0_717 = __p0_717; \
59935  int16x8_t __ret_717; \
59936  __ret_717 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_717), __p1_717)); \
59937  __ret_717; \
59938 })
59939 #else
59940 #define vshll_high_n_s8(__p0_718, __p1_718) __extension__ ({ \
59941  int8x16_t __s0_718 = __p0_718; \
59942  int8x16_t __rev0_718; __rev0_718 = __builtin_shufflevector(__s0_718, __s0_718, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59943  int16x8_t __ret_718; \
59944  __ret_718 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_718), __p1_718)); \
59945  __ret_718 = __builtin_shufflevector(__ret_718, __ret_718, 7, 6, 5, 4, 3, 2, 1, 0); \
59946  __ret_718; \
59947 })
59948 #endif
59949 
59950 #ifdef __LITTLE_ENDIAN__
59951 #define vshll_high_n_s32(__p0_719, __p1_719) __extension__ ({ \
59952  int32x4_t __s0_719 = __p0_719; \
59953  int64x2_t __ret_719; \
59954  __ret_719 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_719), __p1_719)); \
59955  __ret_719; \
59956 })
59957 #else
59958 #define vshll_high_n_s32(__p0_720, __p1_720) __extension__ ({ \
59959  int32x4_t __s0_720 = __p0_720; \
59960  int32x4_t __rev0_720; __rev0_720 = __builtin_shufflevector(__s0_720, __s0_720, 3, 2, 1, 0); \
59961  int64x2_t __ret_720; \
59962  __ret_720 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_720), __p1_720)); \
59963  __ret_720 = __builtin_shufflevector(__ret_720, __ret_720, 1, 0); \
59964  __ret_720; \
59965 })
59966 #endif
59967 
59968 #ifdef __LITTLE_ENDIAN__
59969 #define vshll_high_n_s16(__p0_721, __p1_721) __extension__ ({ \
59970  int16x8_t __s0_721 = __p0_721; \
59971  int32x4_t __ret_721; \
59972  __ret_721 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_721), __p1_721)); \
59973  __ret_721; \
59974 })
59975 #else
59976 #define vshll_high_n_s16(__p0_722, __p1_722) __extension__ ({ \
59977  int16x8_t __s0_722 = __p0_722; \
59978  int16x8_t __rev0_722; __rev0_722 = __builtin_shufflevector(__s0_722, __s0_722, 7, 6, 5, 4, 3, 2, 1, 0); \
59979  int32x4_t __ret_722; \
59980  __ret_722 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_722), __p1_722)); \
59981  __ret_722 = __builtin_shufflevector(__ret_722, __ret_722, 3, 2, 1, 0); \
59982  __ret_722; \
59983 })
59984 #endif
59985 
59986 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
59987  uint64_t __s0 = __p0; \
59988  uint64_t __ret; \
59989  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
59990  __ret; \
59991 })
59992 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
59993  int64_t __s0 = __p0; \
59994  int64_t __ret; \
59995  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
59996  __ret; \
59997 })
59998 #ifdef __LITTLE_ENDIAN__
59999 #define vshrn_high_n_u32(__p0_723, __p1_723, __p2_723) __extension__ ({ \
60000  uint16x4_t __s0_723 = __p0_723; \
60001  uint32x4_t __s1_723 = __p1_723; \
60002  uint16x8_t __ret_723; \
60003  __ret_723 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_723), (uint16x4_t)(vshrn_n_u32(__s1_723, __p2_723)))); \
60004  __ret_723; \
60005 })
60006 #else
60007 #define vshrn_high_n_u32(__p0_724, __p1_724, __p2_724) __extension__ ({ \
60008  uint16x4_t __s0_724 = __p0_724; \
60009  uint32x4_t __s1_724 = __p1_724; \
60010  uint16x4_t __rev0_724; __rev0_724 = __builtin_shufflevector(__s0_724, __s0_724, 3, 2, 1, 0); \
60011  uint32x4_t __rev1_724; __rev1_724 = __builtin_shufflevector(__s1_724, __s1_724, 3, 2, 1, 0); \
60012  uint16x8_t __ret_724; \
60013  __ret_724 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_724), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_724, __p2_724)))); \
60014  __ret_724 = __builtin_shufflevector(__ret_724, __ret_724, 7, 6, 5, 4, 3, 2, 1, 0); \
60015  __ret_724; \
60016 })
60017 #endif
60018 
60019 #ifdef __LITTLE_ENDIAN__
60020 #define vshrn_high_n_u64(__p0_725, __p1_725, __p2_725) __extension__ ({ \
60021  uint32x2_t __s0_725 = __p0_725; \
60022  uint64x2_t __s1_725 = __p1_725; \
60023  uint32x4_t __ret_725; \
60024  __ret_725 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_725), (uint32x2_t)(vshrn_n_u64(__s1_725, __p2_725)))); \
60025  __ret_725; \
60026 })
60027 #else
60028 #define vshrn_high_n_u64(__p0_726, __p1_726, __p2_726) __extension__ ({ \
60029  uint32x2_t __s0_726 = __p0_726; \
60030  uint64x2_t __s1_726 = __p1_726; \
60031  uint32x2_t __rev0_726; __rev0_726 = __builtin_shufflevector(__s0_726, __s0_726, 1, 0); \
60032  uint64x2_t __rev1_726; __rev1_726 = __builtin_shufflevector(__s1_726, __s1_726, 1, 0); \
60033  uint32x4_t __ret_726; \
60034  __ret_726 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_726), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_726, __p2_726)))); \
60035  __ret_726 = __builtin_shufflevector(__ret_726, __ret_726, 3, 2, 1, 0); \
60036  __ret_726; \
60037 })
60038 #endif
60039 
60040 #ifdef __LITTLE_ENDIAN__
60041 #define vshrn_high_n_u16(__p0_727, __p1_727, __p2_727) __extension__ ({ \
60042  uint8x8_t __s0_727 = __p0_727; \
60043  uint16x8_t __s1_727 = __p1_727; \
60044  uint8x16_t __ret_727; \
60045  __ret_727 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_727), (uint8x8_t)(vshrn_n_u16(__s1_727, __p2_727)))); \
60046  __ret_727; \
60047 })
60048 #else
60049 #define vshrn_high_n_u16(__p0_728, __p1_728, __p2_728) __extension__ ({ \
60050  uint8x8_t __s0_728 = __p0_728; \
60051  uint16x8_t __s1_728 = __p1_728; \
60052  uint8x8_t __rev0_728; __rev0_728 = __builtin_shufflevector(__s0_728, __s0_728, 7, 6, 5, 4, 3, 2, 1, 0); \
60053  uint16x8_t __rev1_728; __rev1_728 = __builtin_shufflevector(__s1_728, __s1_728, 7, 6, 5, 4, 3, 2, 1, 0); \
60054  uint8x16_t __ret_728; \
60055  __ret_728 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_728), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_728, __p2_728)))); \
60056  __ret_728 = __builtin_shufflevector(__ret_728, __ret_728, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60057  __ret_728; \
60058 })
60059 #endif
60060 
60061 #ifdef __LITTLE_ENDIAN__
60062 #define vshrn_high_n_s32(__p0_729, __p1_729, __p2_729) __extension__ ({ \
60063  int16x4_t __s0_729 = __p0_729; \
60064  int32x4_t __s1_729 = __p1_729; \
60065  int16x8_t __ret_729; \
60066  __ret_729 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_729), (int16x4_t)(vshrn_n_s32(__s1_729, __p2_729)))); \
60067  __ret_729; \
60068 })
60069 #else
60070 #define vshrn_high_n_s32(__p0_730, __p1_730, __p2_730) __extension__ ({ \
60071  int16x4_t __s0_730 = __p0_730; \
60072  int32x4_t __s1_730 = __p1_730; \
60073  int16x4_t __rev0_730; __rev0_730 = __builtin_shufflevector(__s0_730, __s0_730, 3, 2, 1, 0); \
60074  int32x4_t __rev1_730; __rev1_730 = __builtin_shufflevector(__s1_730, __s1_730, 3, 2, 1, 0); \
60075  int16x8_t __ret_730; \
60076  __ret_730 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_730), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_730, __p2_730)))); \
60077  __ret_730 = __builtin_shufflevector(__ret_730, __ret_730, 7, 6, 5, 4, 3, 2, 1, 0); \
60078  __ret_730; \
60079 })
60080 #endif
60081 
60082 #ifdef __LITTLE_ENDIAN__
60083 #define vshrn_high_n_s64(__p0_731, __p1_731, __p2_731) __extension__ ({ \
60084  int32x2_t __s0_731 = __p0_731; \
60085  int64x2_t __s1_731 = __p1_731; \
60086  int32x4_t __ret_731; \
60087  __ret_731 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_731), (int32x2_t)(vshrn_n_s64(__s1_731, __p2_731)))); \
60088  __ret_731; \
60089 })
60090 #else
60091 #define vshrn_high_n_s64(__p0_732, __p1_732, __p2_732) __extension__ ({ \
60092  int32x2_t __s0_732 = __p0_732; \
60093  int64x2_t __s1_732 = __p1_732; \
60094  int32x2_t __rev0_732; __rev0_732 = __builtin_shufflevector(__s0_732, __s0_732, 1, 0); \
60095  int64x2_t __rev1_732; __rev1_732 = __builtin_shufflevector(__s1_732, __s1_732, 1, 0); \
60096  int32x4_t __ret_732; \
60097  __ret_732 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_732), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_732, __p2_732)))); \
60098  __ret_732 = __builtin_shufflevector(__ret_732, __ret_732, 3, 2, 1, 0); \
60099  __ret_732; \
60100 })
60101 #endif
60102 
60103 #ifdef __LITTLE_ENDIAN__
60104 #define vshrn_high_n_s16(__p0_733, __p1_733, __p2_733) __extension__ ({ \
60105  int8x8_t __s0_733 = __p0_733; \
60106  int16x8_t __s1_733 = __p1_733; \
60107  int8x16_t __ret_733; \
60108  __ret_733 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_733), (int8x8_t)(vshrn_n_s16(__s1_733, __p2_733)))); \
60109  __ret_733; \
60110 })
60111 #else
60112 #define vshrn_high_n_s16(__p0_734, __p1_734, __p2_734) __extension__ ({ \
60113  int8x8_t __s0_734 = __p0_734; \
60114  int16x8_t __s1_734 = __p1_734; \
60115  int8x8_t __rev0_734; __rev0_734 = __builtin_shufflevector(__s0_734, __s0_734, 7, 6, 5, 4, 3, 2, 1, 0); \
60116  int16x8_t __rev1_734; __rev1_734 = __builtin_shufflevector(__s1_734, __s1_734, 7, 6, 5, 4, 3, 2, 1, 0); \
60117  int8x16_t __ret_734; \
60118  __ret_734 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_734), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_734, __p2_734)))); \
60119  __ret_734 = __builtin_shufflevector(__ret_734, __ret_734, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60120  __ret_734; \
60121 })
60122 #endif
60123 
60124 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60125  uint64_t __s0 = __p0; \
60126  uint64_t __s1 = __p1; \
60127  uint64_t __ret; \
60128  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
60129  __ret; \
60130 })
60131 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60132  int64_t __s0 = __p0; \
60133  int64_t __s1 = __p1; \
60134  int64_t __ret; \
60135  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
60136  __ret; \
60137 })
60138 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
60139  poly64x1_t __s0 = __p0; \
60140  poly64x1_t __s1 = __p1; \
60141  poly64x1_t __ret; \
60142  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60143  __ret; \
60144 })
60145 #ifdef __LITTLE_ENDIAN__
60146 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60147  poly64x2_t __s0 = __p0; \
60148  poly64x2_t __s1 = __p1; \
60149  poly64x2_t __ret; \
60150  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
60151  __ret; \
60152 })
60153 #else
60154 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60155  poly64x2_t __s0 = __p0; \
60156  poly64x2_t __s1 = __p1; \
60157  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
60158  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60159  poly64x2_t __ret; \
60160  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
60161  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60162  __ret; \
60163 })
60164 #endif
60165 
60166 __ai uint8_t vsqaddb_u8(uint8_t __p0, int8_t __p1) {
60167  uint8_t __ret;
60168  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
60169  return __ret;
60170 }
60171 __ai uint32_t vsqadds_u32(uint32_t __p0, int32_t __p1) {
60172  uint32_t __ret;
60173  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
60174  return __ret;
60175 }
60176 __ai uint64_t vsqaddd_u64(uint64_t __p0, int64_t __p1) {
60177  uint64_t __ret;
60178  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
60179  return __ret;
60180 }
60181 __ai uint16_t vsqaddh_u16(uint16_t __p0, int16_t __p1) {
60182  uint16_t __ret;
60183  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
60184  return __ret;
60185 }
60186 #ifdef __LITTLE_ENDIAN__
60187 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, int8x16_t __p1) {
60188  uint8x16_t __ret;
60189  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
60190  return __ret;
60191 }
60192 #else
60193 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, int8x16_t __p1) {
60194  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60195  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60196  uint8x16_t __ret;
60197  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
60198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60199  return __ret;
60200 }
60201 #endif
60202 
60203 #ifdef __LITTLE_ENDIAN__
60204 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, int32x4_t __p1) {
60205  uint32x4_t __ret;
60206  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
60207  return __ret;
60208 }
60209 #else
60210 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, int32x4_t __p1) {
60211  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60212  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60213  uint32x4_t __ret;
60214  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
60215  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60216  return __ret;
60217 }
60218 #endif
60219 
60220 #ifdef __LITTLE_ENDIAN__
60221 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, int64x2_t __p1) {
60222  uint64x2_t __ret;
60223  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
60224  return __ret;
60225 }
60226 #else
60227 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, int64x2_t __p1) {
60228  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60229  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60230  uint64x2_t __ret;
60231  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
60232  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60233  return __ret;
60234 }
60235 #endif
60236 
60237 #ifdef __LITTLE_ENDIAN__
60238 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, int16x8_t __p1) {
60239  uint16x8_t __ret;
60240  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
60241  return __ret;
60242 }
60243 #else
60244 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, int16x8_t __p1) {
60245  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60246  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60247  uint16x8_t __ret;
60248  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
60249  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60250  return __ret;
60251 }
60252 #endif
60253 
60254 #ifdef __LITTLE_ENDIAN__
60255 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, int8x8_t __p1) {
60256  uint8x8_t __ret;
60257  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
60258  return __ret;
60259 }
60260 #else
60261 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, int8x8_t __p1) {
60262  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60263  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60264  uint8x8_t __ret;
60265  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
60266  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60267  return __ret;
60268 }
60269 #endif
60270 
60271 #ifdef __LITTLE_ENDIAN__
60272 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, int32x2_t __p1) {
60273  uint32x2_t __ret;
60274  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
60275  return __ret;
60276 }
60277 #else
60278 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, int32x2_t __p1) {
60279  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60280  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60281  uint32x2_t __ret;
60282  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
60283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60284  return __ret;
60285 }
60286 #endif
60287 
60288 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, int64x1_t __p1) {
60289  uint64x1_t __ret;
60290  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60291  return __ret;
60292 }
60293 #ifdef __LITTLE_ENDIAN__
60294 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, int16x4_t __p1) {
60295  uint16x4_t __ret;
60296  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
60297  return __ret;
60298 }
60299 #else
60300 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, int16x4_t __p1) {
60301  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60302  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60303  uint16x4_t __ret;
60304  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
60305  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60306  return __ret;
60307 }
60308 #endif
60309 
60310 #ifdef __LITTLE_ENDIAN__
60311 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
60312  float64x2_t __ret;
60313  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
60314  return __ret;
60315 }
60316 #else
60317 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
60318  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60319  float64x2_t __ret;
60320  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
60321  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60322  return __ret;
60323 }
60324 #endif
60325 
60326 #ifdef __LITTLE_ENDIAN__
60327 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
60328  float32x4_t __ret;
60329  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
60330  return __ret;
60331 }
60332 #else
60333 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
60334  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60335  float32x4_t __ret;
60336  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
60337  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60338  return __ret;
60339 }
60340 #endif
60341 
60342 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
60343  float64x1_t __ret;
60344  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
60345  return __ret;
60346 }
60347 #ifdef __LITTLE_ENDIAN__
60348 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
60349  float32x2_t __ret;
60350  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
60351  return __ret;
60352 }
60353 #else
60354 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
60355  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60356  float32x2_t __ret;
60357  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
60358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60359  return __ret;
60360 }
60361 #endif
60362 
60363 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60364  uint64_t __s0 = __p0; \
60365  uint64_t __s1 = __p1; \
60366  uint64_t __ret; \
60367  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
60368  __ret; \
60369 })
60370 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60371  int64_t __s0 = __p0; \
60372  int64_t __s1 = __p1; \
60373  int64_t __ret; \
60374  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
60375  __ret; \
60376 })
60377 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60378  uint64_t __s0 = __p0; \
60379  uint64_t __s1 = __p1; \
60380  uint64_t __ret; \
60381  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
60382  __ret; \
60383 })
60384 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60385  int64_t __s0 = __p0; \
60386  int64_t __s1 = __p1; \
60387  int64_t __ret; \
60388  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
60389  __ret; \
60390 })
60391 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
60392  poly64x1_t __s0 = __p0; \
60393  poly64x1_t __s1 = __p1; \
60394  poly64x1_t __ret; \
60395  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60396  __ret; \
60397 })
60398 #ifdef __LITTLE_ENDIAN__
60399 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60400  poly64x2_t __s0 = __p0; \
60401  poly64x2_t __s1 = __p1; \
60402  poly64x2_t __ret; \
60403  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
60404  __ret; \
60405 })
60406 #else
60407 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60408  poly64x2_t __s0 = __p0; \
60409  poly64x2_t __s1 = __p1; \
60410  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
60411  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60412  poly64x2_t __ret; \
60413  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
60414  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60415  __ret; \
60416 })
60417 #endif
60418 
60419 #define vst1_p64(__p0, __p1) __extension__ ({ \
60420  poly64x1_t __s1 = __p1; \
60421  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
60422 })
60423 #ifdef __LITTLE_ENDIAN__
60424 #define vst1q_p64(__p0, __p1) __extension__ ({ \
60425  poly64x2_t __s1 = __p1; \
60426  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
60427 })
60428 #else
60429 #define vst1q_p64(__p0, __p1) __extension__ ({ \
60430  poly64x2_t __s1 = __p1; \
60431  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60432  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
60433 })
60434 #endif
60435 
60436 #ifdef __LITTLE_ENDIAN__
60437 #define vst1q_f64(__p0, __p1) __extension__ ({ \
60438  float64x2_t __s1 = __p1; \
60439  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
60440 })
60441 #else
60442 #define vst1q_f64(__p0, __p1) __extension__ ({ \
60443  float64x2_t __s1 = __p1; \
60444  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60445  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
60446 })
60447 #endif
60448 
60449 #define vst1_f64(__p0, __p1) __extension__ ({ \
60450  float64x1_t __s1 = __p1; \
60451  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
60452 })
60453 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60454  poly64x1_t __s1 = __p1; \
60455  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
60456 })
60457 #ifdef __LITTLE_ENDIAN__
60458 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60459  poly64x2_t __s1 = __p1; \
60460  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
60461 })
60462 #else
60463 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60464  poly64x2_t __s1 = __p1; \
60465  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60466  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
60467 })
60468 #endif
60469 
60470 #ifdef __LITTLE_ENDIAN__
60471 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60472  float64x2_t __s1 = __p1; \
60473  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
60474 })
60475 #else
60476 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60477  float64x2_t __s1 = __p1; \
60478  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60479  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
60480 })
60481 #endif
60482 
60483 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60484  float64x1_t __s1 = __p1; \
60485  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
60486 })
60487 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
60488  poly64x1x2_t __s1 = __p1; \
60489  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
60490 })
60491 #ifdef __LITTLE_ENDIAN__
60492 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
60493  poly64x2x2_t __s1 = __p1; \
60494  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
60495 })
60496 #else
60497 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
60498  poly64x2x2_t __s1 = __p1; \
60499  poly64x2x2_t __rev1; \
60500  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60501  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60502  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
60503 })
60504 #endif
60505 
60506 #ifdef __LITTLE_ENDIAN__
60507 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
60508  float64x2x2_t __s1 = __p1; \
60509  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 42); \
60510 })
60511 #else
60512 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
60513  float64x2x2_t __s1 = __p1; \
60514  float64x2x2_t __rev1; \
60515  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60516  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60517  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 42); \
60518 })
60519 #endif
60520 
60521 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
60522  float64x1x2_t __s1 = __p1; \
60523  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 10); \
60524 })
60525 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
60526  poly64x1x3_t __s1 = __p1; \
60527  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
60528 })
60529 #ifdef __LITTLE_ENDIAN__
60530 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
60531  poly64x2x3_t __s1 = __p1; \
60532  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
60533 })
60534 #else
60535 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
60536  poly64x2x3_t __s1 = __p1; \
60537  poly64x2x3_t __rev1; \
60538  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60539  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60540  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60541  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
60542 })
60543 #endif
60544 
60545 #ifdef __LITTLE_ENDIAN__
60546 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
60547  float64x2x3_t __s1 = __p1; \
60548  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 42); \
60549 })
60550 #else
60551 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
60552  float64x2x3_t __s1 = __p1; \
60553  float64x2x3_t __rev1; \
60554  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60555  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60556  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60557  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 42); \
60558 })
60559 #endif
60560 
60561 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
60562  float64x1x3_t __s1 = __p1; \
60563  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 10); \
60564 })
60565 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
60566  poly64x1x4_t __s1 = __p1; \
60567  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
60568 })
60569 #ifdef __LITTLE_ENDIAN__
60570 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
60571  poly64x2x4_t __s1 = __p1; \
60572  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
60573 })
60574 #else
60575 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
60576  poly64x2x4_t __s1 = __p1; \
60577  poly64x2x4_t __rev1; \
60578  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60579  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60580  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60581  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
60582  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
60583 })
60584 #endif
60585 
60586 #ifdef __LITTLE_ENDIAN__
60587 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
60588  float64x2x4_t __s1 = __p1; \
60589  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 42); \
60590 })
60591 #else
60592 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
60593  float64x2x4_t __s1 = __p1; \
60594  float64x2x4_t __rev1; \
60595  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60596  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60597  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60598  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
60599  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 42); \
60600 })
60601 #endif
60602 
60603 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
60604  float64x1x4_t __s1 = __p1; \
60605  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 10); \
60606 })
60607 #define vst2_p64(__p0, __p1) __extension__ ({ \
60608  poly64x1x2_t __s1 = __p1; \
60609  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
60610 })
60611 #ifdef __LITTLE_ENDIAN__
60612 #define vst2q_p64(__p0, __p1) __extension__ ({ \
60613  poly64x2x2_t __s1 = __p1; \
60614  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
60615 })
60616 #else
60617 #define vst2q_p64(__p0, __p1) __extension__ ({ \
60618  poly64x2x2_t __s1 = __p1; \
60619  poly64x2x2_t __rev1; \
60620  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60621  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60622  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
60623 })
60624 #endif
60625 
60626 #ifdef __LITTLE_ENDIAN__
60627 #define vst2q_u64(__p0, __p1) __extension__ ({ \
60628  uint64x2x2_t __s1 = __p1; \
60629  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
60630 })
60631 #else
60632 #define vst2q_u64(__p0, __p1) __extension__ ({ \
60633  uint64x2x2_t __s1 = __p1; \
60634  uint64x2x2_t __rev1; \
60635  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60636  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60637  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
60638 })
60639 #endif
60640 
60641 #ifdef __LITTLE_ENDIAN__
60642 #define vst2q_f64(__p0, __p1) __extension__ ({ \
60643  float64x2x2_t __s1 = __p1; \
60644  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 42); \
60645 })
60646 #else
60647 #define vst2q_f64(__p0, __p1) __extension__ ({ \
60648  float64x2x2_t __s1 = __p1; \
60649  float64x2x2_t __rev1; \
60650  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60651  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60652  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 42); \
60653 })
60654 #endif
60655 
60656 #ifdef __LITTLE_ENDIAN__
60657 #define vst2q_s64(__p0, __p1) __extension__ ({ \
60658  int64x2x2_t __s1 = __p1; \
60659  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 35); \
60660 })
60661 #else
60662 #define vst2q_s64(__p0, __p1) __extension__ ({ \
60663  int64x2x2_t __s1 = __p1; \
60664  int64x2x2_t __rev1; \
60665  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60666  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60667  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 35); \
60668 })
60669 #endif
60670 
60671 #define vst2_f64(__p0, __p1) __extension__ ({ \
60672  float64x1x2_t __s1 = __p1; \
60673  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 10); \
60674 })
60675 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60676  poly64x1x2_t __s1 = __p1; \
60677  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
60678 })
60679 #ifdef __LITTLE_ENDIAN__
60680 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
60681  poly8x16x2_t __s1 = __p1; \
60682  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
60683 })
60684 #else
60685 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
60686  poly8x16x2_t __s1 = __p1; \
60687  poly8x16x2_t __rev1; \
60688  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60689  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60690  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
60691 })
60692 #endif
60693 
60694 #ifdef __LITTLE_ENDIAN__
60695 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60696  poly64x2x2_t __s1 = __p1; \
60697  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
60698 })
60699 #else
60700 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60701  poly64x2x2_t __s1 = __p1; \
60702  poly64x2x2_t __rev1; \
60703  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60704  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60705  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
60706 })
60707 #endif
60708 
60709 #ifdef __LITTLE_ENDIAN__
60710 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
60711  uint8x16x2_t __s1 = __p1; \
60712  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
60713 })
60714 #else
60715 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
60716  uint8x16x2_t __s1 = __p1; \
60717  uint8x16x2_t __rev1; \
60718  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60719  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60720  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
60721 })
60722 #endif
60723 
60724 #ifdef __LITTLE_ENDIAN__
60725 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60726  uint64x2x2_t __s1 = __p1; \
60727  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
60728 })
60729 #else
60730 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60731  uint64x2x2_t __s1 = __p1; \
60732  uint64x2x2_t __rev1; \
60733  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60734  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60735  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
60736 })
60737 #endif
60738 
60739 #ifdef __LITTLE_ENDIAN__
60740 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
60741  int8x16x2_t __s1 = __p1; \
60742  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
60743 })
60744 #else
60745 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
60746  int8x16x2_t __s1 = __p1; \
60747  int8x16x2_t __rev1; \
60748  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60749  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60750  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
60751 })
60752 #endif
60753 
60754 #ifdef __LITTLE_ENDIAN__
60755 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60756  float64x2x2_t __s1 = __p1; \
60757  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 42); \
60758 })
60759 #else
60760 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60761  float64x2x2_t __s1 = __p1; \
60762  float64x2x2_t __rev1; \
60763  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60764  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60765  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 42); \
60766 })
60767 #endif
60768 
60769 #ifdef __LITTLE_ENDIAN__
60770 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60771  int64x2x2_t __s1 = __p1; \
60772  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 35); \
60773 })
60774 #else
60775 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60776  int64x2x2_t __s1 = __p1; \
60777  int64x2x2_t __rev1; \
60778  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60779  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60780  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 35); \
60781 })
60782 #endif
60783 
60784 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60785  uint64x1x2_t __s1 = __p1; \
60786  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
60787 })
60788 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60789  float64x1x2_t __s1 = __p1; \
60790  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 10); \
60791 })
60792 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60793  int64x1x2_t __s1 = __p1; \
60794  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 3); \
60795 })
60796 #define vst3_p64(__p0, __p1) __extension__ ({ \
60797  poly64x1x3_t __s1 = __p1; \
60798  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
60799 })
60800 #ifdef __LITTLE_ENDIAN__
60801 #define vst3q_p64(__p0, __p1) __extension__ ({ \
60802  poly64x2x3_t __s1 = __p1; \
60803  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
60804 })
60805 #else
60806 #define vst3q_p64(__p0, __p1) __extension__ ({ \
60807  poly64x2x3_t __s1 = __p1; \
60808  poly64x2x3_t __rev1; \
60809  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60810  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60811  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60812  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
60813 })
60814 #endif
60815 
60816 #ifdef __LITTLE_ENDIAN__
60817 #define vst3q_u64(__p0, __p1) __extension__ ({ \
60818  uint64x2x3_t __s1 = __p1; \
60819  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
60820 })
60821 #else
60822 #define vst3q_u64(__p0, __p1) __extension__ ({ \
60823  uint64x2x3_t __s1 = __p1; \
60824  uint64x2x3_t __rev1; \
60825  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60826  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60827  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60828  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
60829 })
60830 #endif
60831 
60832 #ifdef __LITTLE_ENDIAN__
60833 #define vst3q_f64(__p0, __p1) __extension__ ({ \
60834  float64x2x3_t __s1 = __p1; \
60835  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 42); \
60836 })
60837 #else
60838 #define vst3q_f64(__p0, __p1) __extension__ ({ \
60839  float64x2x3_t __s1 = __p1; \
60840  float64x2x3_t __rev1; \
60841  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60842  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60843  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60844  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 42); \
60845 })
60846 #endif
60847 
60848 #ifdef __LITTLE_ENDIAN__
60849 #define vst3q_s64(__p0, __p1) __extension__ ({ \
60850  int64x2x3_t __s1 = __p1; \
60851  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 35); \
60852 })
60853 #else
60854 #define vst3q_s64(__p0, __p1) __extension__ ({ \
60855  int64x2x3_t __s1 = __p1; \
60856  int64x2x3_t __rev1; \
60857  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60858  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60859  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60860  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 35); \
60861 })
60862 #endif
60863 
60864 #define vst3_f64(__p0, __p1) __extension__ ({ \
60865  float64x1x3_t __s1 = __p1; \
60866  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 10); \
60867 })
60868 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60869  poly64x1x3_t __s1 = __p1; \
60870  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
60871 })
60872 #ifdef __LITTLE_ENDIAN__
60873 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
60874  poly8x16x3_t __s1 = __p1; \
60875  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
60876 })
60877 #else
60878 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
60879  poly8x16x3_t __s1 = __p1; \
60880  poly8x16x3_t __rev1; \
60881  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60882  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60883  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60884  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
60885 })
60886 #endif
60887 
60888 #ifdef __LITTLE_ENDIAN__
60889 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60890  poly64x2x3_t __s1 = __p1; \
60891  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
60892 })
60893 #else
60894 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60895  poly64x2x3_t __s1 = __p1; \
60896  poly64x2x3_t __rev1; \
60897  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60898  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60899  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60900  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
60901 })
60902 #endif
60903 
60904 #ifdef __LITTLE_ENDIAN__
60905 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
60906  uint8x16x3_t __s1 = __p1; \
60907  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
60908 })
60909 #else
60910 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
60911  uint8x16x3_t __s1 = __p1; \
60912  uint8x16x3_t __rev1; \
60913  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60914  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60915  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60916  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
60917 })
60918 #endif
60919 
60920 #ifdef __LITTLE_ENDIAN__
60921 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60922  uint64x2x3_t __s1 = __p1; \
60923  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
60924 })
60925 #else
60926 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60927  uint64x2x3_t __s1 = __p1; \
60928  uint64x2x3_t __rev1; \
60929  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60930  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60931  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60932  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
60933 })
60934 #endif
60935 
60936 #ifdef __LITTLE_ENDIAN__
60937 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
60938  int8x16x3_t __s1 = __p1; \
60939  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
60940 })
60941 #else
60942 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
60943  int8x16x3_t __s1 = __p1; \
60944  int8x16x3_t __rev1; \
60945  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60946  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60947  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60948  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
60949 })
60950 #endif
60951 
60952 #ifdef __LITTLE_ENDIAN__
60953 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60954  float64x2x3_t __s1 = __p1; \
60955  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 42); \
60956 })
60957 #else
60958 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60959  float64x2x3_t __s1 = __p1; \
60960  float64x2x3_t __rev1; \
60961  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60962  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60963  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60964  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 42); \
60965 })
60966 #endif
60967 
60968 #ifdef __LITTLE_ENDIAN__
60969 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60970  int64x2x3_t __s1 = __p1; \
60971  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 35); \
60972 })
60973 #else
60974 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60975  int64x2x3_t __s1 = __p1; \
60976  int64x2x3_t __rev1; \
60977  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
60978  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
60979  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
60980  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 35); \
60981 })
60982 #endif
60983 
60984 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
60985  uint64x1x3_t __s1 = __p1; \
60986  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
60987 })
60988 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60989  float64x1x3_t __s1 = __p1; \
60990  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 10); \
60991 })
60992 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
60993  int64x1x3_t __s1 = __p1; \
60994  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 3); \
60995 })
60996 #define vst4_p64(__p0, __p1) __extension__ ({ \
60997  poly64x1x4_t __s1 = __p1; \
60998  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
60999 })
61000 #ifdef __LITTLE_ENDIAN__
61001 #define vst4q_p64(__p0, __p1) __extension__ ({ \
61002  poly64x2x4_t __s1 = __p1; \
61003  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
61004 })
61005 #else
61006 #define vst4q_p64(__p0, __p1) __extension__ ({ \
61007  poly64x2x4_t __s1 = __p1; \
61008  poly64x2x4_t __rev1; \
61009  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61010  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61011  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61012  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61013  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
61014 })
61015 #endif
61016 
61017 #ifdef __LITTLE_ENDIAN__
61018 #define vst4q_u64(__p0, __p1) __extension__ ({ \
61019  uint64x2x4_t __s1 = __p1; \
61020  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
61021 })
61022 #else
61023 #define vst4q_u64(__p0, __p1) __extension__ ({ \
61024  uint64x2x4_t __s1 = __p1; \
61025  uint64x2x4_t __rev1; \
61026  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61027  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61028  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61029  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61030  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
61031 })
61032 #endif
61033 
61034 #ifdef __LITTLE_ENDIAN__
61035 #define vst4q_f64(__p0, __p1) __extension__ ({ \
61036  float64x2x4_t __s1 = __p1; \
61037  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 42); \
61038 })
61039 #else
61040 #define vst4q_f64(__p0, __p1) __extension__ ({ \
61041  float64x2x4_t __s1 = __p1; \
61042  float64x2x4_t __rev1; \
61043  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61044  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61045  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61046  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61047  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 42); \
61048 })
61049 #endif
61050 
61051 #ifdef __LITTLE_ENDIAN__
61052 #define vst4q_s64(__p0, __p1) __extension__ ({ \
61053  int64x2x4_t __s1 = __p1; \
61054  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 35); \
61055 })
61056 #else
61057 #define vst4q_s64(__p0, __p1) __extension__ ({ \
61058  int64x2x4_t __s1 = __p1; \
61059  int64x2x4_t __rev1; \
61060  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61061  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61062  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61063  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61064  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 35); \
61065 })
61066 #endif
61067 
61068 #define vst4_f64(__p0, __p1) __extension__ ({ \
61069  float64x1x4_t __s1 = __p1; \
61070  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 10); \
61071 })
61072 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61073  poly64x1x4_t __s1 = __p1; \
61074  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
61075 })
61076 #ifdef __LITTLE_ENDIAN__
61077 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
61078  poly8x16x4_t __s1 = __p1; \
61079  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
61080 })
61081 #else
61082 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
61083  poly8x16x4_t __s1 = __p1; \
61084  poly8x16x4_t __rev1; \
61085  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61086  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61087  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61088  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61089  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
61090 })
61091 #endif
61092 
61093 #ifdef __LITTLE_ENDIAN__
61094 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61095  poly64x2x4_t __s1 = __p1; \
61096  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
61097 })
61098 #else
61099 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61100  poly64x2x4_t __s1 = __p1; \
61101  poly64x2x4_t __rev1; \
61102  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61103  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61104  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61105  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61106  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
61107 })
61108 #endif
61109 
61110 #ifdef __LITTLE_ENDIAN__
61111 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
61112  uint8x16x4_t __s1 = __p1; \
61113  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
61114 })
61115 #else
61116 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
61117  uint8x16x4_t __s1 = __p1; \
61118  uint8x16x4_t __rev1; \
61119  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61120  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61121  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61122  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61123  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
61124 })
61125 #endif
61126 
61127 #ifdef __LITTLE_ENDIAN__
61128 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
61129  uint64x2x4_t __s1 = __p1; \
61130  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
61131 })
61132 #else
61133 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
61134  uint64x2x4_t __s1 = __p1; \
61135  uint64x2x4_t __rev1; \
61136  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61137  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61138  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61139  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61140  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
61141 })
61142 #endif
61143 
61144 #ifdef __LITTLE_ENDIAN__
61145 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
61146  int8x16x4_t __s1 = __p1; \
61147  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
61148 })
61149 #else
61150 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
61151  int8x16x4_t __s1 = __p1; \
61152  int8x16x4_t __rev1; \
61153  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61154  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61155  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61156  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61157  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
61158 })
61159 #endif
61160 
61161 #ifdef __LITTLE_ENDIAN__
61162 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61163  float64x2x4_t __s1 = __p1; \
61164  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 42); \
61165 })
61166 #else
61167 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61168  float64x2x4_t __s1 = __p1; \
61169  float64x2x4_t __rev1; \
61170  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61171  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61172  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61173  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61174  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 42); \
61175 })
61176 #endif
61177 
61178 #ifdef __LITTLE_ENDIAN__
61179 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
61180  int64x2x4_t __s1 = __p1; \
61181  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 35); \
61182 })
61183 #else
61184 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
61185  int64x2x4_t __s1 = __p1; \
61186  int64x2x4_t __rev1; \
61187  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61188  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61189  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61190  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
61191  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 35); \
61192 })
61193 #endif
61194 
61195 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
61196  uint64x1x4_t __s1 = __p1; \
61197  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
61198 })
61199 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61200  float64x1x4_t __s1 = __p1; \
61201  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 10); \
61202 })
61203 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
61204  int64x1x4_t __s1 = __p1; \
61205  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 3); \
61206 })
61207 #define vstrq_p128(__p0, __p1) __extension__ ({ \
61208  poly128_t __s1 = __p1; \
61209  __builtin_neon_vstrq_p128(__p0, __s1); \
61210 })
61211 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
61212  uint64_t __ret;
61213  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
61214  return __ret;
61215 }
61216 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
61217  int64_t __ret;
61218  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
61219  return __ret;
61220 }
61221 #ifdef __LITTLE_ENDIAN__
61222 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
61223  float64x2_t __ret;
61224  __ret = __p0 - __p1;
61225  return __ret;
61226 }
61227 #else
61228 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
61229  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61230  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61231  float64x2_t __ret;
61232  __ret = __rev0 - __rev1;
61233  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61234  return __ret;
61235 }
61236 #endif
61237 
61238 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
61239  float64x1_t __ret;
61240  __ret = __p0 - __p1;
61241  return __ret;
61242 }
61243 #ifdef __LITTLE_ENDIAN__
61244 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
61245  uint16x8_t __ret;
61246  __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
61247  return __ret;
61248 }
61249 #else
61250 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
61251  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61252  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61253  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
61254  uint16x8_t __ret;
61255  __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
61256  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61257  return __ret;
61258 }
61259 #endif
61260 
61261 #ifdef __LITTLE_ENDIAN__
61262 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
61263  uint32x4_t __ret;
61264  __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
61265  return __ret;
61266 }
61267 #else
61268 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
61269  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61270  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61271  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
61272  uint32x4_t __ret;
61273  __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
61274  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61275  return __ret;
61276 }
61277 #endif
61278 
61279 #ifdef __LITTLE_ENDIAN__
61280 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
61281  uint8x16_t __ret;
61282  __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
61283  return __ret;
61284 }
61285 #else
61286 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
61287  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61288  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61289  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
61290  uint8x16_t __ret;
61291  __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
61292  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61293  return __ret;
61294 }
61295 #endif
61296 
61297 #ifdef __LITTLE_ENDIAN__
61298 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
61299  int16x8_t __ret;
61300  __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
61301  return __ret;
61302 }
61303 #else
61304 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
61305  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61306  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61307  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
61308  int16x8_t __ret;
61309  __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
61310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61311  return __ret;
61312 }
61313 #endif
61314 
61315 #ifdef __LITTLE_ENDIAN__
61316 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
61317  int32x4_t __ret;
61318  __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
61319  return __ret;
61320 }
61321 #else
61322 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
61323  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61324  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61325  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
61326  int32x4_t __ret;
61327  __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
61328  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61329  return __ret;
61330 }
61331 #endif
61332 
61333 #ifdef __LITTLE_ENDIAN__
61334 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
61335  int8x16_t __ret;
61336  __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
61337  return __ret;
61338 }
61339 #else
61340 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
61341  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61342  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61343  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
61344  int8x16_t __ret;
61345  __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
61346  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61347  return __ret;
61348 }
61349 #endif
61350 
61351 #ifdef __LITTLE_ENDIAN__
61352 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
61353  uint16x8_t __ret;
61354  __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
61355  return __ret;
61356 }
61357 #else
61358 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
61359  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61360  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61361  uint16x8_t __ret;
61362  __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
61363  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61364  return __ret;
61365 }
61366 #endif
61367 
61368 #ifdef __LITTLE_ENDIAN__
61369 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
61370  uint64x2_t __ret;
61371  __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
61372  return __ret;
61373 }
61374 #else
61375 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
61376  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61377  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61378  uint64x2_t __ret;
61379  __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
61380  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61381  return __ret;
61382 }
61383 #endif
61384 
61385 #ifdef __LITTLE_ENDIAN__
61386 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
61387  uint32x4_t __ret;
61388  __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
61389  return __ret;
61390 }
61391 #else
61392 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
61393  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61394  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61395  uint32x4_t __ret;
61396  __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
61397  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61398  return __ret;
61399 }
61400 #endif
61401 
61402 #ifdef __LITTLE_ENDIAN__
61403 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
61404  int16x8_t __ret;
61405  __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
61406  return __ret;
61407 }
61408 #else
61409 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
61410  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61411  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61412  int16x8_t __ret;
61413  __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
61414  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61415  return __ret;
61416 }
61417 #endif
61418 
61419 #ifdef __LITTLE_ENDIAN__
61420 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
61421  int64x2_t __ret;
61422  __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
61423  return __ret;
61424 }
61425 #else
61426 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
61427  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61428  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61429  int64x2_t __ret;
61430  __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
61431  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61432  return __ret;
61433 }
61434 #endif
61435 
61436 #ifdef __LITTLE_ENDIAN__
61437 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
61438  int32x4_t __ret;
61439  __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
61440  return __ret;
61441 }
61442 #else
61443 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
61444  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61445  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61446  int32x4_t __ret;
61447  __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
61448  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61449  return __ret;
61450 }
61451 #endif
61452 
61453 #ifdef __LITTLE_ENDIAN__
61454 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
61455  uint16x8_t __ret;
61456  __ret = __p0 - vmovl_high_u8(__p1);
61457  return __ret;
61458 }
61459 #else
61460 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
61461  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61462  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61463  uint16x8_t __ret;
61464  __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
61465  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61466  return __ret;
61467 }
61468 #endif
61469 
61470 #ifdef __LITTLE_ENDIAN__
61471 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
61472  uint64x2_t __ret;
61473  __ret = __p0 - vmovl_high_u32(__p1);
61474  return __ret;
61475 }
61476 #else
61477 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
61478  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61479  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61480  uint64x2_t __ret;
61481  __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
61482  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61483  return __ret;
61484 }
61485 #endif
61486 
61487 #ifdef __LITTLE_ENDIAN__
61488 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
61489  uint32x4_t __ret;
61490  __ret = __p0 - vmovl_high_u16(__p1);
61491  return __ret;
61492 }
61493 #else
61494 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
61495  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61496  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61497  uint32x4_t __ret;
61498  __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
61499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61500  return __ret;
61501 }
61502 #endif
61503 
61504 #ifdef __LITTLE_ENDIAN__
61505 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
61506  int16x8_t __ret;
61507  __ret = __p0 - vmovl_high_s8(__p1);
61508  return __ret;
61509 }
61510 #else
61511 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
61512  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61513  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61514  int16x8_t __ret;
61515  __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
61516  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61517  return __ret;
61518 }
61519 #endif
61520 
61521 #ifdef __LITTLE_ENDIAN__
61522 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
61523  int64x2_t __ret;
61524  __ret = __p0 - vmovl_high_s32(__p1);
61525  return __ret;
61526 }
61527 #else
61528 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
61529  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61530  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61531  int64x2_t __ret;
61532  __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
61533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61534  return __ret;
61535 }
61536 #endif
61537 
61538 #ifdef __LITTLE_ENDIAN__
61539 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
61540  int32x4_t __ret;
61541  __ret = __p0 - vmovl_high_s16(__p1);
61542  return __ret;
61543 }
61544 #else
61545 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
61546  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61547  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61548  int32x4_t __ret;
61549  __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
61550  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61551  return __ret;
61552 }
61553 #endif
61554 
61555 #ifdef __LITTLE_ENDIAN__
61556 #define vsudotq_laneq_s32(__p0_735, __p1_735, __p2_735, __p3_735) __extension__ ({ \
61557  int32x4_t __s0_735 = __p0_735; \
61558  int8x16_t __s1_735 = __p1_735; \
61559  uint8x16_t __s2_735 = __p2_735; \
61560  int32x4_t __ret_735; \
61561 uint8x16_t __reint_735 = __s2_735; \
61562  __ret_735 = vusdotq_s32(__s0_735, (uint8x16_t)(splatq_laneq_s32(*(int32x4_t *) &__reint_735, __p3_735)), __s1_735); \
61563  __ret_735; \
61564 })
61565 #else
61566 #define vsudotq_laneq_s32(__p0_736, __p1_736, __p2_736, __p3_736) __extension__ ({ \
61567  int32x4_t __s0_736 = __p0_736; \
61568  int8x16_t __s1_736 = __p1_736; \
61569  uint8x16_t __s2_736 = __p2_736; \
61570  int32x4_t __rev0_736; __rev0_736 = __builtin_shufflevector(__s0_736, __s0_736, 3, 2, 1, 0); \
61571  int8x16_t __rev1_736; __rev1_736 = __builtin_shufflevector(__s1_736, __s1_736, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61572  uint8x16_t __rev2_736; __rev2_736 = __builtin_shufflevector(__s2_736, __s2_736, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61573  int32x4_t __ret_736; \
61574 uint8x16_t __reint_736 = __rev2_736; \
61575  __ret_736 = __noswap_vusdotq_s32(__rev0_736, (uint8x16_t)(__noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_736, __p3_736)), __rev1_736); \
61576  __ret_736 = __builtin_shufflevector(__ret_736, __ret_736, 3, 2, 1, 0); \
61577  __ret_736; \
61578 })
61579 #endif
61580 
61581 #ifdef __LITTLE_ENDIAN__
61582 #define vsudot_laneq_s32(__p0_737, __p1_737, __p2_737, __p3_737) __extension__ ({ \
61583  int32x2_t __s0_737 = __p0_737; \
61584  int8x8_t __s1_737 = __p1_737; \
61585  uint8x16_t __s2_737 = __p2_737; \
61586  int32x2_t __ret_737; \
61587 uint8x16_t __reint_737 = __s2_737; \
61588  __ret_737 = vusdot_s32(__s0_737, (uint8x8_t)(splat_laneq_s32(*(int32x4_t *) &__reint_737, __p3_737)), __s1_737); \
61589  __ret_737; \
61590 })
61591 #else
61592 #define vsudot_laneq_s32(__p0_738, __p1_738, __p2_738, __p3_738) __extension__ ({ \
61593  int32x2_t __s0_738 = __p0_738; \
61594  int8x8_t __s1_738 = __p1_738; \
61595  uint8x16_t __s2_738 = __p2_738; \
61596  int32x2_t __rev0_738; __rev0_738 = __builtin_shufflevector(__s0_738, __s0_738, 1, 0); \
61597  int8x8_t __rev1_738; __rev1_738 = __builtin_shufflevector(__s1_738, __s1_738, 7, 6, 5, 4, 3, 2, 1, 0); \
61598  uint8x16_t __rev2_738; __rev2_738 = __builtin_shufflevector(__s2_738, __s2_738, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61599  int32x2_t __ret_738; \
61600 uint8x16_t __reint_738 = __rev2_738; \
61601  __ret_738 = __noswap_vusdot_s32(__rev0_738, (uint8x8_t)(__noswap_splat_laneq_s32(*(int32x4_t *) &__reint_738, __p3_738)), __rev1_738); \
61602  __ret_738 = __builtin_shufflevector(__ret_738, __ret_738, 1, 0); \
61603  __ret_738; \
61604 })
61605 #endif
61606 
61607 #ifdef __LITTLE_ENDIAN__
61608 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61609  poly8x8_t __ret;
61610  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61611  return __ret;
61612 }
61613 #else
61614 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
61615  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61616  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61617  poly8x8_t __ret;
61618  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61620  return __ret;
61621 }
61622 #endif
61623 
61624 #ifdef __LITTLE_ENDIAN__
61625 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61626  poly16x4_t __ret;
61627  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61628  return __ret;
61629 }
61630 #else
61631 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
61632  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61633  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61634  poly16x4_t __ret;
61635  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61636  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61637  return __ret;
61638 }
61639 #endif
61640 
61641 #ifdef __LITTLE_ENDIAN__
61642 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61643  poly8x16_t __ret;
61644  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61645  return __ret;
61646 }
61647 #else
61648 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
61649  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61650  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61651  poly8x16_t __ret;
61652  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61653  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61654  return __ret;
61655 }
61656 #endif
61657 
61658 #ifdef __LITTLE_ENDIAN__
61659 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61660  poly64x2_t __ret;
61661  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61662  return __ret;
61663 }
61664 #else
61665 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
61666  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61667  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61668  poly64x2_t __ret;
61669  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61671  return __ret;
61672 }
61673 #endif
61674 
61675 #ifdef __LITTLE_ENDIAN__
61676 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61677  poly16x8_t __ret;
61678  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61679  return __ret;
61680 }
61681 #else
61682 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
61683  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61684  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61685  poly16x8_t __ret;
61686  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61688  return __ret;
61689 }
61690 #endif
61691 
61692 #ifdef __LITTLE_ENDIAN__
61693 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61694  uint8x16_t __ret;
61695  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61696  return __ret;
61697 }
61698 #else
61699 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
61700  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61701  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61702  uint8x16_t __ret;
61703  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61704  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61705  return __ret;
61706 }
61707 #endif
61708 
61709 #ifdef __LITTLE_ENDIAN__
61710 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61711  uint32x4_t __ret;
61712  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61713  return __ret;
61714 }
61715 #else
61716 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
61717  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61718  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61719  uint32x4_t __ret;
61720  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61721  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61722  return __ret;
61723 }
61724 #endif
61725 
61726 #ifdef __LITTLE_ENDIAN__
61727 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61728  uint64x2_t __ret;
61729  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61730  return __ret;
61731 }
61732 #else
61733 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
61734  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61735  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61736  uint64x2_t __ret;
61737  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61738  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61739  return __ret;
61740 }
61741 #endif
61742 
61743 #ifdef __LITTLE_ENDIAN__
61744 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61745  uint16x8_t __ret;
61746  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61747  return __ret;
61748 }
61749 #else
61750 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
61751  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61752  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61753  uint16x8_t __ret;
61754  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61755  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61756  return __ret;
61757 }
61758 #endif
61759 
61760 #ifdef __LITTLE_ENDIAN__
61761 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
61762  int8x16_t __ret;
61763  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61764  return __ret;
61765 }
61766 #else
61767 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
61768  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61769  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61770  int8x16_t __ret;
61771  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
61772  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61773  return __ret;
61774 }
61775 #endif
61776 
61777 #ifdef __LITTLE_ENDIAN__
61778 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
61779  float64x2_t __ret;
61780  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61781  return __ret;
61782 }
61783 #else
61784 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
61785  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61786  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61787  float64x2_t __ret;
61788  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61790  return __ret;
61791 }
61792 #endif
61793 
61794 #ifdef __LITTLE_ENDIAN__
61795 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
61796  float32x4_t __ret;
61797  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61798  return __ret;
61799 }
61800 #else
61801 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
61802  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61803  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61804  float32x4_t __ret;
61805  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61806  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61807  return __ret;
61808 }
61809 #endif
61810 
61811 #ifdef __LITTLE_ENDIAN__
61812 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
61813  int32x4_t __ret;
61814  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61815  return __ret;
61816 }
61817 #else
61818 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
61819  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61820  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61821  int32x4_t __ret;
61822  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61824  return __ret;
61825 }
61826 #endif
61827 
61828 #ifdef __LITTLE_ENDIAN__
61829 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
61830  int64x2_t __ret;
61831  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61832  return __ret;
61833 }
61834 #else
61835 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
61836  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61837  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61838  int64x2_t __ret;
61839  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61840  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61841  return __ret;
61842 }
61843 #endif
61844 
61845 #ifdef __LITTLE_ENDIAN__
61846 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
61847  int16x8_t __ret;
61848  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61849  return __ret;
61850 }
61851 #else
61852 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
61853  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61854  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61855  int16x8_t __ret;
61856  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61858  return __ret;
61859 }
61860 #endif
61861 
61862 #ifdef __LITTLE_ENDIAN__
61863 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
61864  uint8x8_t __ret;
61865  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61866  return __ret;
61867 }
61868 #else
61869 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
61870  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61871  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61872  uint8x8_t __ret;
61873  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61874  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61875  return __ret;
61876 }
61877 #endif
61878 
61879 #ifdef __LITTLE_ENDIAN__
61880 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
61881  uint32x2_t __ret;
61882  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61883  return __ret;
61884 }
61885 #else
61886 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
61887  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61888  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61889  uint32x2_t __ret;
61890  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61891  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61892  return __ret;
61893 }
61894 #endif
61895 
61896 #ifdef __LITTLE_ENDIAN__
61897 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
61898  uint16x4_t __ret;
61899  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61900  return __ret;
61901 }
61902 #else
61903 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
61904  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61905  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61906  uint16x4_t __ret;
61907  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61908  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61909  return __ret;
61910 }
61911 #endif
61912 
61913 #ifdef __LITTLE_ENDIAN__
61914 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
61915  int8x8_t __ret;
61916  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
61917  return __ret;
61918 }
61919 #else
61920 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
61921  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61922  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61923  int8x8_t __ret;
61924  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
61925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61926  return __ret;
61927 }
61928 #endif
61929 
61930 #ifdef __LITTLE_ENDIAN__
61931 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
61932  float32x2_t __ret;
61933  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61934  return __ret;
61935 }
61936 #else
61937 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
61938  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61939  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61940  float32x2_t __ret;
61941  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61942  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61943  return __ret;
61944 }
61945 #endif
61946 
61947 #ifdef __LITTLE_ENDIAN__
61948 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
61949  int32x2_t __ret;
61950  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
61951  return __ret;
61952 }
61953 #else
61954 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
61955  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61956  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61957  int32x2_t __ret;
61958  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
61959  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61960  return __ret;
61961 }
61962 #endif
61963 
61964 #ifdef __LITTLE_ENDIAN__
61965 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
61966  int16x4_t __ret;
61967  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
61968  return __ret;
61969 }
61970 #else
61971 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
61972  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61973  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61974  int16x4_t __ret;
61975  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
61976  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61977  return __ret;
61978 }
61979 #endif
61980 
61981 #ifdef __LITTLE_ENDIAN__
61982 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
61983  poly8x8_t __ret;
61984  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
61985  return __ret;
61986 }
61987 #else
61988 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
61989  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61990  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61991  poly8x8_t __ret;
61992  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
61993  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61994  return __ret;
61995 }
61996 #endif
61997 
61998 #ifdef __LITTLE_ENDIAN__
61999 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
62000  poly16x4_t __ret;
62001  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62002  return __ret;
62003 }
62004 #else
62005 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
62006  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62007  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62008  poly16x4_t __ret;
62009  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62010  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62011  return __ret;
62012 }
62013 #endif
62014 
62015 #ifdef __LITTLE_ENDIAN__
62016 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62017  poly8x16_t __ret;
62018  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62019  return __ret;
62020 }
62021 #else
62022 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62023  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62024  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62025  poly8x16_t __ret;
62026  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62027  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62028  return __ret;
62029 }
62030 #endif
62031 
62032 #ifdef __LITTLE_ENDIAN__
62033 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62034  poly64x2_t __ret;
62035  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62036  return __ret;
62037 }
62038 #else
62039 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62040  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62041  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62042  poly64x2_t __ret;
62043  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62044  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62045  return __ret;
62046 }
62047 #endif
62048 
62049 #ifdef __LITTLE_ENDIAN__
62050 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62051  poly16x8_t __ret;
62052  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
62053  return __ret;
62054 }
62055 #else
62056 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62057  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62058  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62059  poly16x8_t __ret;
62060  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
62061  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62062  return __ret;
62063 }
62064 #endif
62065 
62066 #ifdef __LITTLE_ENDIAN__
62067 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62068  uint8x16_t __ret;
62069  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62070  return __ret;
62071 }
62072 #else
62073 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62074  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62075  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62076  uint8x16_t __ret;
62077  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62078  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62079  return __ret;
62080 }
62081 #endif
62082 
62083 #ifdef __LITTLE_ENDIAN__
62084 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62085  uint32x4_t __ret;
62086  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62087  return __ret;
62088 }
62089 #else
62090 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62091  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62092  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62093  uint32x4_t __ret;
62094  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62095  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62096  return __ret;
62097 }
62098 #endif
62099 
62100 #ifdef __LITTLE_ENDIAN__
62101 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62102  uint64x2_t __ret;
62103  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62104  return __ret;
62105 }
62106 #else
62107 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62108  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62109  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62110  uint64x2_t __ret;
62111  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62112  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62113  return __ret;
62114 }
62115 #endif
62116 
62117 #ifdef __LITTLE_ENDIAN__
62118 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62119  uint16x8_t __ret;
62120  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
62121  return __ret;
62122 }
62123 #else
62124 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62125  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62126  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62127  uint16x8_t __ret;
62128  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
62129  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62130  return __ret;
62131 }
62132 #endif
62133 
62134 #ifdef __LITTLE_ENDIAN__
62135 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
62136  int8x16_t __ret;
62137  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62138  return __ret;
62139 }
62140 #else
62141 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
62142  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62143  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62144  int8x16_t __ret;
62145  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
62146  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62147  return __ret;
62148 }
62149 #endif
62150 
62151 #ifdef __LITTLE_ENDIAN__
62152 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
62153  float64x2_t __ret;
62154  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62155  return __ret;
62156 }
62157 #else
62158 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
62159  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62160  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62161  float64x2_t __ret;
62162  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62163  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62164  return __ret;
62165 }
62166 #endif
62167 
62168 #ifdef __LITTLE_ENDIAN__
62169 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
62170  float32x4_t __ret;
62171  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62172  return __ret;
62173 }
62174 #else
62175 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
62176  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62177  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62178  float32x4_t __ret;
62179  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62181  return __ret;
62182 }
62183 #endif
62184 
62185 #ifdef __LITTLE_ENDIAN__
62186 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
62187  int32x4_t __ret;
62188  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62189  return __ret;
62190 }
62191 #else
62192 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
62193  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62194  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62195  int32x4_t __ret;
62196  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62197  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62198  return __ret;
62199 }
62200 #endif
62201 
62202 #ifdef __LITTLE_ENDIAN__
62203 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
62204  int64x2_t __ret;
62205  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62206  return __ret;
62207 }
62208 #else
62209 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
62210  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62211  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62212  int64x2_t __ret;
62213  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62214  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62215  return __ret;
62216 }
62217 #endif
62218 
62219 #ifdef __LITTLE_ENDIAN__
62220 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
62221  int16x8_t __ret;
62222  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
62223  return __ret;
62224 }
62225 #else
62226 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
62227  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62228  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62229  int16x8_t __ret;
62230  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
62231  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62232  return __ret;
62233 }
62234 #endif
62235 
62236 #ifdef __LITTLE_ENDIAN__
62237 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
62238  uint8x8_t __ret;
62239  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
62240  return __ret;
62241 }
62242 #else
62243 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
62244  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62245  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62246  uint8x8_t __ret;
62247  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
62248  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62249  return __ret;
62250 }
62251 #endif
62252 
62253 #ifdef __LITTLE_ENDIAN__
62254 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
62255  uint32x2_t __ret;
62256  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62257  return __ret;
62258 }
62259 #else
62260 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
62261  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62262  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62263  uint32x2_t __ret;
62264  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62266  return __ret;
62267 }
62268 #endif
62269 
62270 #ifdef __LITTLE_ENDIAN__
62271 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
62272  uint16x4_t __ret;
62273  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62274  return __ret;
62275 }
62276 #else
62277 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
62278  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62279  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62280  uint16x4_t __ret;
62281  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62282  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62283  return __ret;
62284 }
62285 #endif
62286 
62287 #ifdef __LITTLE_ENDIAN__
62288 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
62289  int8x8_t __ret;
62290  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
62291  return __ret;
62292 }
62293 #else
62294 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
62295  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62296  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62297  int8x8_t __ret;
62298  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
62299  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62300  return __ret;
62301 }
62302 #endif
62303 
62304 #ifdef __LITTLE_ENDIAN__
62305 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
62306  float32x2_t __ret;
62307  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62308  return __ret;
62309 }
62310 #else
62311 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
62312  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62313  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62314  float32x2_t __ret;
62315  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62316  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62317  return __ret;
62318 }
62319 #endif
62320 
62321 #ifdef __LITTLE_ENDIAN__
62322 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
62323  int32x2_t __ret;
62324  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
62325  return __ret;
62326 }
62327 #else
62328 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
62329  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62330  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62331  int32x2_t __ret;
62332  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
62333  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62334  return __ret;
62335 }
62336 #endif
62337 
62338 #ifdef __LITTLE_ENDIAN__
62339 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
62340  int16x4_t __ret;
62341  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
62342  return __ret;
62343 }
62344 #else
62345 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
62346  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62347  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62348  int16x4_t __ret;
62349  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
62350  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62351  return __ret;
62352 }
62353 #endif
62354 
62355 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
62356  uint64x1_t __ret;
62357  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
62358  return __ret;
62359 }
62360 #ifdef __LITTLE_ENDIAN__
62361 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
62362  uint64x2_t __ret;
62363  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
62364  return __ret;
62365 }
62366 #else
62367 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
62368  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62369  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62370  uint64x2_t __ret;
62371  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
62372  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62373  return __ret;
62374 }
62375 #endif
62376 
62377 #ifdef __LITTLE_ENDIAN__
62378 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
62379  uint64x2_t __ret;
62380  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
62381  return __ret;
62382 }
62383 #else
62384 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
62385  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62386  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62387  uint64x2_t __ret;
62388  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
62389  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62390  return __ret;
62391 }
62392 #endif
62393 
62394 #ifdef __LITTLE_ENDIAN__
62395 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
62396  uint64x2_t __ret;
62397  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
62398  return __ret;
62399 }
62400 #else
62401 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
62402  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62403  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62404  uint64x2_t __ret;
62405  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
62406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62407  return __ret;
62408 }
62409 #endif
62410 
62411 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
62412  uint64x1_t __ret;
62413  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
62414  return __ret;
62415 }
62416 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
62417  uint64x1_t __ret;
62418  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
62419  return __ret;
62420 }
62421 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
62422  uint64_t __ret;
62423  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
62424  return __ret;
62425 }
62426 __ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
62427  int64_t __ret;
62428  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
62429  return __ret;
62430 }
62431 __ai int8_t vuqaddb_s8(int8_t __p0, uint8_t __p1) {
62432  int8_t __ret;
62433  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
62434  return __ret;
62435 }
62436 __ai int32_t vuqadds_s32(int32_t __p0, uint32_t __p1) {
62437  int32_t __ret;
62438  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
62439  return __ret;
62440 }
62441 __ai int64_t vuqaddd_s64(int64_t __p0, uint64_t __p1) {
62442  int64_t __ret;
62443  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
62444  return __ret;
62445 }
62446 __ai int16_t vuqaddh_s16(int16_t __p0, uint16_t __p1) {
62447  int16_t __ret;
62448  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
62449  return __ret;
62450 }
62451 #ifdef __LITTLE_ENDIAN__
62452 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, uint8x16_t __p1) {
62453  int8x16_t __ret;
62454  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
62455  return __ret;
62456 }
62457 #else
62458 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, uint8x16_t __p1) {
62459  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62460  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62461  int8x16_t __ret;
62462  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
62463  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62464  return __ret;
62465 }
62466 #endif
62467 
62468 #ifdef __LITTLE_ENDIAN__
62469 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, uint32x4_t __p1) {
62470  int32x4_t __ret;
62471  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
62472  return __ret;
62473 }
62474 #else
62475 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, uint32x4_t __p1) {
62476  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62477  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62478  int32x4_t __ret;
62479  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
62480  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62481  return __ret;
62482 }
62483 #endif
62484 
62485 #ifdef __LITTLE_ENDIAN__
62486 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, uint64x2_t __p1) {
62487  int64x2_t __ret;
62488  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
62489  return __ret;
62490 }
62491 #else
62492 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, uint64x2_t __p1) {
62493  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62494  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62495  int64x2_t __ret;
62496  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
62497  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62498  return __ret;
62499 }
62500 #endif
62501 
62502 #ifdef __LITTLE_ENDIAN__
62503 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, uint16x8_t __p1) {
62504  int16x8_t __ret;
62505  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
62506  return __ret;
62507 }
62508 #else
62509 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, uint16x8_t __p1) {
62510  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62511  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62512  int16x8_t __ret;
62513  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
62514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62515  return __ret;
62516 }
62517 #endif
62518 
62519 #ifdef __LITTLE_ENDIAN__
62520 __ai int8x8_t vuqadd_s8(int8x8_t __p0, uint8x8_t __p1) {
62521  int8x8_t __ret;
62522  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
62523  return __ret;
62524 }
62525 #else
62526 __ai int8x8_t vuqadd_s8(int8x8_t __p0, uint8x8_t __p1) {
62527  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62528  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62529  int8x8_t __ret;
62530  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
62531  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62532  return __ret;
62533 }
62534 #endif
62535 
62536 #ifdef __LITTLE_ENDIAN__
62537 __ai int32x2_t vuqadd_s32(int32x2_t __p0, uint32x2_t __p1) {
62538  int32x2_t __ret;
62539  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
62540  return __ret;
62541 }
62542 #else
62543 __ai int32x2_t vuqadd_s32(int32x2_t __p0, uint32x2_t __p1) {
62544  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62545  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62546  int32x2_t __ret;
62547  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
62548  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62549  return __ret;
62550 }
62551 #endif
62552 
62553 __ai int64x1_t vuqadd_s64(int64x1_t __p0, uint64x1_t __p1) {
62554  int64x1_t __ret;
62555  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
62556  return __ret;
62557 }
62558 #ifdef __LITTLE_ENDIAN__
62559 __ai int16x4_t vuqadd_s16(int16x4_t __p0, uint16x4_t __p1) {
62560  int16x4_t __ret;
62561  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
62562  return __ret;
62563 }
62564 #else
62565 __ai int16x4_t vuqadd_s16(int16x4_t __p0, uint16x4_t __p1) {
62566  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62567  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62568  int16x4_t __ret;
62569  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
62570  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62571  return __ret;
62572 }
62573 #endif
62574 
62575 #ifdef __LITTLE_ENDIAN__
62576 #define vusdotq_laneq_s32(__p0_739, __p1_739, __p2_739, __p3_739) __extension__ ({ \
62577  int32x4_t __s0_739 = __p0_739; \
62578  uint8x16_t __s1_739 = __p1_739; \
62579  int8x16_t __s2_739 = __p2_739; \
62580  int32x4_t __ret_739; \
62581 int8x16_t __reint_739 = __s2_739; \
62582  __ret_739 = vusdotq_s32(__s0_739, __s1_739, (int8x16_t)(splatq_laneq_s32(*(int32x4_t *) &__reint_739, __p3_739))); \
62583  __ret_739; \
62584 })
62585 #else
62586 #define vusdotq_laneq_s32(__p0_740, __p1_740, __p2_740, __p3_740) __extension__ ({ \
62587  int32x4_t __s0_740 = __p0_740; \
62588  uint8x16_t __s1_740 = __p1_740; \
62589  int8x16_t __s2_740 = __p2_740; \
62590  int32x4_t __rev0_740; __rev0_740 = __builtin_shufflevector(__s0_740, __s0_740, 3, 2, 1, 0); \
62591  uint8x16_t __rev1_740; __rev1_740 = __builtin_shufflevector(__s1_740, __s1_740, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62592  int8x16_t __rev2_740; __rev2_740 = __builtin_shufflevector(__s2_740, __s2_740, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62593  int32x4_t __ret_740; \
62594 int8x16_t __reint_740 = __rev2_740; \
62595  __ret_740 = __noswap_vusdotq_s32(__rev0_740, __rev1_740, (int8x16_t)(__noswap_splatq_laneq_s32(*(int32x4_t *) &__reint_740, __p3_740))); \
62596  __ret_740 = __builtin_shufflevector(__ret_740, __ret_740, 3, 2, 1, 0); \
62597  __ret_740; \
62598 })
62599 #endif
62600 
62601 #ifdef __LITTLE_ENDIAN__
62602 #define vusdot_laneq_s32(__p0_741, __p1_741, __p2_741, __p3_741) __extension__ ({ \
62603  int32x2_t __s0_741 = __p0_741; \
62604  uint8x8_t __s1_741 = __p1_741; \
62605  int8x16_t __s2_741 = __p2_741; \
62606  int32x2_t __ret_741; \
62607 int8x16_t __reint_741 = __s2_741; \
62608  __ret_741 = vusdot_s32(__s0_741, __s1_741, (int8x8_t)(splat_laneq_s32(*(int32x4_t *) &__reint_741, __p3_741))); \
62609  __ret_741; \
62610 })
62611 #else
62612 #define vusdot_laneq_s32(__p0_742, __p1_742, __p2_742, __p3_742) __extension__ ({ \
62613  int32x2_t __s0_742 = __p0_742; \
62614  uint8x8_t __s1_742 = __p1_742; \
62615  int8x16_t __s2_742 = __p2_742; \
62616  int32x2_t __rev0_742; __rev0_742 = __builtin_shufflevector(__s0_742, __s0_742, 1, 0); \
62617  uint8x8_t __rev1_742; __rev1_742 = __builtin_shufflevector(__s1_742, __s1_742, 7, 6, 5, 4, 3, 2, 1, 0); \
62618  int8x16_t __rev2_742; __rev2_742 = __builtin_shufflevector(__s2_742, __s2_742, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62619  int32x2_t __ret_742; \
62620 int8x16_t __reint_742 = __rev2_742; \
62621  __ret_742 = __noswap_vusdot_s32(__rev0_742, __rev1_742, (int8x8_t)(__noswap_splat_laneq_s32(*(int32x4_t *) &__reint_742, __p3_742))); \
62622  __ret_742 = __builtin_shufflevector(__ret_742, __ret_742, 1, 0); \
62623  __ret_742; \
62624 })
62625 #endif
62626 
62627 #ifdef __LITTLE_ENDIAN__
62628 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
62629  poly8x8_t __ret;
62630  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62631  return __ret;
62632 }
62633 #else
62634 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
62635  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62636  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62637  poly8x8_t __ret;
62638  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62639  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62640  return __ret;
62641 }
62642 #endif
62643 
62644 #ifdef __LITTLE_ENDIAN__
62645 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
62646  poly16x4_t __ret;
62647  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62648  return __ret;
62649 }
62650 #else
62651 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
62652  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62653  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62654  poly16x4_t __ret;
62655  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62656  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62657  return __ret;
62658 }
62659 #endif
62660 
62661 #ifdef __LITTLE_ENDIAN__
62662 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62663  poly8x16_t __ret;
62664  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62665  return __ret;
62666 }
62667 #else
62668 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
62669  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62670  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62671  poly8x16_t __ret;
62672  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62673  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62674  return __ret;
62675 }
62676 #endif
62677 
62678 #ifdef __LITTLE_ENDIAN__
62679 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62680  poly64x2_t __ret;
62681  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62682  return __ret;
62683 }
62684 #else
62685 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
62686  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62687  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62688  poly64x2_t __ret;
62689  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62690  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62691  return __ret;
62692 }
62693 #endif
62694 
62695 #ifdef __LITTLE_ENDIAN__
62696 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62697  poly16x8_t __ret;
62698  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62699  return __ret;
62700 }
62701 #else
62702 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
62703  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62704  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62705  poly16x8_t __ret;
62706  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62707  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62708  return __ret;
62709 }
62710 #endif
62711 
62712 #ifdef __LITTLE_ENDIAN__
62713 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62714  uint8x16_t __ret;
62715  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62716  return __ret;
62717 }
62718 #else
62719 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
62720  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62721  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62722  uint8x16_t __ret;
62723  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62724  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62725  return __ret;
62726 }
62727 #endif
62728 
62729 #ifdef __LITTLE_ENDIAN__
62730 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62731  uint32x4_t __ret;
62732  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62733  return __ret;
62734 }
62735 #else
62736 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
62737  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62738  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62739  uint32x4_t __ret;
62740  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62742  return __ret;
62743 }
62744 #endif
62745 
62746 #ifdef __LITTLE_ENDIAN__
62747 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62748  uint64x2_t __ret;
62749  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62750  return __ret;
62751 }
62752 #else
62753 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
62754  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62755  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62756  uint64x2_t __ret;
62757  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62758  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62759  return __ret;
62760 }
62761 #endif
62762 
62763 #ifdef __LITTLE_ENDIAN__
62764 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62765  uint16x8_t __ret;
62766  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62767  return __ret;
62768 }
62769 #else
62770 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
62771  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62772  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62773  uint16x8_t __ret;
62774  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62775  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62776  return __ret;
62777 }
62778 #endif
62779 
62780 #ifdef __LITTLE_ENDIAN__
62781 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
62782  int8x16_t __ret;
62783  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62784  return __ret;
62785 }
62786 #else
62787 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
62788  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62789  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62790  int8x16_t __ret;
62791  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
62792  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
62793  return __ret;
62794 }
62795 #endif
62796 
62797 #ifdef __LITTLE_ENDIAN__
62798 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
62799  float64x2_t __ret;
62800  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62801  return __ret;
62802 }
62803 #else
62804 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
62805  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62806  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62807  float64x2_t __ret;
62808  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62809  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62810  return __ret;
62811 }
62812 #endif
62813 
62814 #ifdef __LITTLE_ENDIAN__
62815 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
62816  float32x4_t __ret;
62817  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62818  return __ret;
62819 }
62820 #else
62821 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
62822  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62823  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62824  float32x4_t __ret;
62825  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62826  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62827  return __ret;
62828 }
62829 #endif
62830 
62831 #ifdef __LITTLE_ENDIAN__
62832 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
62833  int32x4_t __ret;
62834  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62835  return __ret;
62836 }
62837 #else
62838 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
62839  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62840  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62841  int32x4_t __ret;
62842  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62843  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62844  return __ret;
62845 }
62846 #endif
62847 
62848 #ifdef __LITTLE_ENDIAN__
62849 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
62850  int64x2_t __ret;
62851  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62852  return __ret;
62853 }
62854 #else
62855 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
62856  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62857  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62858  int64x2_t __ret;
62859  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62860  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62861  return __ret;
62862 }
62863 #endif
62864 
62865 #ifdef __LITTLE_ENDIAN__
62866 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
62867  int16x8_t __ret;
62868  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62869  return __ret;
62870 }
62871 #else
62872 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
62873  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62874  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62875  int16x8_t __ret;
62876  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62877  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62878  return __ret;
62879 }
62880 #endif
62881 
62882 #ifdef __LITTLE_ENDIAN__
62883 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
62884  uint8x8_t __ret;
62885  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62886  return __ret;
62887 }
62888 #else
62889 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
62890  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62891  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62892  uint8x8_t __ret;
62893  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62894  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62895  return __ret;
62896 }
62897 #endif
62898 
62899 #ifdef __LITTLE_ENDIAN__
62900 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
62901  uint32x2_t __ret;
62902  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62903  return __ret;
62904 }
62905 #else
62906 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
62907  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62908  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62909  uint32x2_t __ret;
62910  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62911  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62912  return __ret;
62913 }
62914 #endif
62915 
62916 #ifdef __LITTLE_ENDIAN__
62917 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
62918  uint16x4_t __ret;
62919  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62920  return __ret;
62921 }
62922 #else
62923 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
62924  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62925  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62926  uint16x4_t __ret;
62927  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62928  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62929  return __ret;
62930 }
62931 #endif
62932 
62933 #ifdef __LITTLE_ENDIAN__
62934 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
62935  int8x8_t __ret;
62936  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
62937  return __ret;
62938 }
62939 #else
62940 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
62941  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
62942  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
62943  int8x8_t __ret;
62944  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
62945  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
62946  return __ret;
62947 }
62948 #endif
62949 
62950 #ifdef __LITTLE_ENDIAN__
62951 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
62952  float32x2_t __ret;
62953  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62954  return __ret;
62955 }
62956 #else
62957 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
62958  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62959  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62960  float32x2_t __ret;
62961  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62962  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62963  return __ret;
62964 }
62965 #endif
62966 
62967 #ifdef __LITTLE_ENDIAN__
62968 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
62969  int32x2_t __ret;
62970  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
62971  return __ret;
62972 }
62973 #else
62974 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
62975  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
62976  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
62977  int32x2_t __ret;
62978  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
62979  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
62980  return __ret;
62981 }
62982 #endif
62983 
62984 #ifdef __LITTLE_ENDIAN__
62985 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
62986  int16x4_t __ret;
62987  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
62988  return __ret;
62989 }
62990 #else
62991 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
62992  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
62993  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
62994  int16x4_t __ret;
62995  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
62996  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
62997  return __ret;
62998 }
62999 #endif
63000 
63001 #ifdef __LITTLE_ENDIAN__
63002 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63003  poly8x8_t __ret;
63004  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63005  return __ret;
63006 }
63007 #else
63008 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63009  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63010  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63011  poly8x8_t __ret;
63012  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63014  return __ret;
63015 }
63016 #endif
63017 
63018 #ifdef __LITTLE_ENDIAN__
63019 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63020  poly16x4_t __ret;
63021  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63022  return __ret;
63023 }
63024 #else
63025 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63026  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63027  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63028  poly16x4_t __ret;
63029  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63030  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63031  return __ret;
63032 }
63033 #endif
63034 
63035 #ifdef __LITTLE_ENDIAN__
63036 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63037  poly8x16_t __ret;
63038  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63039  return __ret;
63040 }
63041 #else
63042 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63043  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63044  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63045  poly8x16_t __ret;
63046  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63047  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63048  return __ret;
63049 }
63050 #endif
63051 
63052 #ifdef __LITTLE_ENDIAN__
63053 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63054  poly64x2_t __ret;
63055  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63056  return __ret;
63057 }
63058 #else
63059 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63060  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63061  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63062  poly64x2_t __ret;
63063  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63064  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63065  return __ret;
63066 }
63067 #endif
63068 
63069 #ifdef __LITTLE_ENDIAN__
63070 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63071  poly16x8_t __ret;
63072  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63073  return __ret;
63074 }
63075 #else
63076 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63077  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63078  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63079  poly16x8_t __ret;
63080  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63081  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63082  return __ret;
63083 }
63084 #endif
63085 
63086 #ifdef __LITTLE_ENDIAN__
63087 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63088  uint8x16_t __ret;
63089  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63090  return __ret;
63091 }
63092 #else
63093 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63094  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63095  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63096  uint8x16_t __ret;
63097  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63098  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63099  return __ret;
63100 }
63101 #endif
63102 
63103 #ifdef __LITTLE_ENDIAN__
63104 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63105  uint32x4_t __ret;
63106  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63107  return __ret;
63108 }
63109 #else
63110 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63111  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63112  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63113  uint32x4_t __ret;
63114  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63115  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63116  return __ret;
63117 }
63118 #endif
63119 
63120 #ifdef __LITTLE_ENDIAN__
63121 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63122  uint64x2_t __ret;
63123  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63124  return __ret;
63125 }
63126 #else
63127 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63128  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63129  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63130  uint64x2_t __ret;
63131  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63132  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63133  return __ret;
63134 }
63135 #endif
63136 
63137 #ifdef __LITTLE_ENDIAN__
63138 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63139  uint16x8_t __ret;
63140  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63141  return __ret;
63142 }
63143 #else
63144 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63145  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63146  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63147  uint16x8_t __ret;
63148  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63149  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63150  return __ret;
63151 }
63152 #endif
63153 
63154 #ifdef __LITTLE_ENDIAN__
63155 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
63156  int8x16_t __ret;
63157  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63158  return __ret;
63159 }
63160 #else
63161 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
63162  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63163  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63164  int8x16_t __ret;
63165  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
63166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63167  return __ret;
63168 }
63169 #endif
63170 
63171 #ifdef __LITTLE_ENDIAN__
63172 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
63173  float64x2_t __ret;
63174  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63175  return __ret;
63176 }
63177 #else
63178 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
63179  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63180  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63181  float64x2_t __ret;
63182  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63183  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63184  return __ret;
63185 }
63186 #endif
63187 
63188 #ifdef __LITTLE_ENDIAN__
63189 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
63190  float32x4_t __ret;
63191  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63192  return __ret;
63193 }
63194 #else
63195 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
63196  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63197  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63198  float32x4_t __ret;
63199  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63201  return __ret;
63202 }
63203 #endif
63204 
63205 #ifdef __LITTLE_ENDIAN__
63206 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
63207  int32x4_t __ret;
63208  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63209  return __ret;
63210 }
63211 #else
63212 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
63213  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63214  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63215  int32x4_t __ret;
63216  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63217  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63218  return __ret;
63219 }
63220 #endif
63221 
63222 #ifdef __LITTLE_ENDIAN__
63223 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
63224  int64x2_t __ret;
63225  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63226  return __ret;
63227 }
63228 #else
63229 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
63230  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63231  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63232  int64x2_t __ret;
63233  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63235  return __ret;
63236 }
63237 #endif
63238 
63239 #ifdef __LITTLE_ENDIAN__
63240 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
63241  int16x8_t __ret;
63242  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63243  return __ret;
63244 }
63245 #else
63246 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
63247  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63248  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63249  int16x8_t __ret;
63250  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63251  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63252  return __ret;
63253 }
63254 #endif
63255 
63256 #ifdef __LITTLE_ENDIAN__
63257 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
63258  uint8x8_t __ret;
63259  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63260  return __ret;
63261 }
63262 #else
63263 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
63264  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63265  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63266  uint8x8_t __ret;
63267  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63268  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63269  return __ret;
63270 }
63271 #endif
63272 
63273 #ifdef __LITTLE_ENDIAN__
63274 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
63275  uint32x2_t __ret;
63276  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63277  return __ret;
63278 }
63279 #else
63280 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
63281  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63282  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63283  uint32x2_t __ret;
63284  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63285  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63286  return __ret;
63287 }
63288 #endif
63289 
63290 #ifdef __LITTLE_ENDIAN__
63291 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
63292  uint16x4_t __ret;
63293  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63294  return __ret;
63295 }
63296 #else
63297 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
63298  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63299  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63300  uint16x4_t __ret;
63301  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63302  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63303  return __ret;
63304 }
63305 #endif
63306 
63307 #ifdef __LITTLE_ENDIAN__
63308 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
63309  int8x8_t __ret;
63310  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
63311  return __ret;
63312 }
63313 #else
63314 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
63315  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63316  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63317  int8x8_t __ret;
63318  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
63319  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63320  return __ret;
63321 }
63322 #endif
63323 
63324 #ifdef __LITTLE_ENDIAN__
63325 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
63326  float32x2_t __ret;
63327  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63328  return __ret;
63329 }
63330 #else
63331 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
63332  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63333  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63334  float32x2_t __ret;
63335  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63337  return __ret;
63338 }
63339 #endif
63340 
63341 #ifdef __LITTLE_ENDIAN__
63342 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
63343  int32x2_t __ret;
63344  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63345  return __ret;
63346 }
63347 #else
63348 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
63349  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63350  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63351  int32x2_t __ret;
63352  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63353  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63354  return __ret;
63355 }
63356 #endif
63357 
63358 #ifdef __LITTLE_ENDIAN__
63359 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
63360  int16x4_t __ret;
63361  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
63362  return __ret;
63363 }
63364 #else
63365 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
63366  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63367  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63368  int16x4_t __ret;
63369  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
63370  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63371  return __ret;
63372 }
63373 #endif
63374 
63375 #ifdef __LITTLE_ENDIAN__
63376 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
63377  poly8x8_t __ret;
63378  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63379  return __ret;
63380 }
63381 #else
63382 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
63383  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63384  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63385  poly8x8_t __ret;
63386  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63387  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63388  return __ret;
63389 }
63390 #endif
63391 
63392 #ifdef __LITTLE_ENDIAN__
63393 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
63394  poly16x4_t __ret;
63395  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63396  return __ret;
63397 }
63398 #else
63399 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
63400  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63401  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63402  poly16x4_t __ret;
63403  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63404  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63405  return __ret;
63406 }
63407 #endif
63408 
63409 #ifdef __LITTLE_ENDIAN__
63410 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63411  poly8x16_t __ret;
63412  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63413  return __ret;
63414 }
63415 #else
63416 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63417  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63418  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63419  poly8x16_t __ret;
63420  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63421  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63422  return __ret;
63423 }
63424 #endif
63425 
63426 #ifdef __LITTLE_ENDIAN__
63427 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63428  poly64x2_t __ret;
63429  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63430  return __ret;
63431 }
63432 #else
63433 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63434  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63435  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63436  poly64x2_t __ret;
63437  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63438  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63439  return __ret;
63440 }
63441 #endif
63442 
63443 #ifdef __LITTLE_ENDIAN__
63444 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63445  poly16x8_t __ret;
63446  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63447  return __ret;
63448 }
63449 #else
63450 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63451  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63452  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63453  poly16x8_t __ret;
63454  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63456  return __ret;
63457 }
63458 #endif
63459 
63460 #ifdef __LITTLE_ENDIAN__
63461 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63462  uint8x16_t __ret;
63463  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63464  return __ret;
63465 }
63466 #else
63467 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63468  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63469  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63470  uint8x16_t __ret;
63471  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63472  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63473  return __ret;
63474 }
63475 #endif
63476 
63477 #ifdef __LITTLE_ENDIAN__
63478 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63479  uint32x4_t __ret;
63480  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63481  return __ret;
63482 }
63483 #else
63484 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63485  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63486  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63487  uint32x4_t __ret;
63488  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63489  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63490  return __ret;
63491 }
63492 #endif
63493 
63494 #ifdef __LITTLE_ENDIAN__
63495 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63496  uint64x2_t __ret;
63497  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63498  return __ret;
63499 }
63500 #else
63501 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63502  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63503  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63504  uint64x2_t __ret;
63505  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63506  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63507  return __ret;
63508 }
63509 #endif
63510 
63511 #ifdef __LITTLE_ENDIAN__
63512 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63513  uint16x8_t __ret;
63514  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63515  return __ret;
63516 }
63517 #else
63518 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63519  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63520  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63521  uint16x8_t __ret;
63522  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63523  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63524  return __ret;
63525 }
63526 #endif
63527 
63528 #ifdef __LITTLE_ENDIAN__
63529 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
63530  int8x16_t __ret;
63531  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63532  return __ret;
63533 }
63534 #else
63535 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
63536  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63537  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63538  int8x16_t __ret;
63539  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
63540  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63541  return __ret;
63542 }
63543 #endif
63544 
63545 #ifdef __LITTLE_ENDIAN__
63546 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
63547  float64x2_t __ret;
63548  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63549  return __ret;
63550 }
63551 #else
63552 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
63553  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63554  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63555  float64x2_t __ret;
63556  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63558  return __ret;
63559 }
63560 #endif
63561 
63562 #ifdef __LITTLE_ENDIAN__
63563 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
63564  float32x4_t __ret;
63565  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63566  return __ret;
63567 }
63568 #else
63569 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
63570  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63571  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63572  float32x4_t __ret;
63573  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63575  return __ret;
63576 }
63577 #endif
63578 
63579 #ifdef __LITTLE_ENDIAN__
63580 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
63581  int32x4_t __ret;
63582  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63583  return __ret;
63584 }
63585 #else
63586 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
63587  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63588  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63589  int32x4_t __ret;
63590  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63591  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63592  return __ret;
63593 }
63594 #endif
63595 
63596 #ifdef __LITTLE_ENDIAN__
63597 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
63598  int64x2_t __ret;
63599  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63600  return __ret;
63601 }
63602 #else
63603 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
63604  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63605  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63606  int64x2_t __ret;
63607  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63608  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63609  return __ret;
63610 }
63611 #endif
63612 
63613 #ifdef __LITTLE_ENDIAN__
63614 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
63615  int16x8_t __ret;
63616  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63617  return __ret;
63618 }
63619 #else
63620 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
63621  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63622  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63623  int16x8_t __ret;
63624  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63625  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63626  return __ret;
63627 }
63628 #endif
63629 
63630 #ifdef __LITTLE_ENDIAN__
63631 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
63632  uint8x8_t __ret;
63633  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63634  return __ret;
63635 }
63636 #else
63637 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
63638  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63639  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63640  uint8x8_t __ret;
63641  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63642  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63643  return __ret;
63644 }
63645 #endif
63646 
63647 #ifdef __LITTLE_ENDIAN__
63648 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
63649  uint32x2_t __ret;
63650  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63651  return __ret;
63652 }
63653 #else
63654 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
63655  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63656  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63657  uint32x2_t __ret;
63658  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63659  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63660  return __ret;
63661 }
63662 #endif
63663 
63664 #ifdef __LITTLE_ENDIAN__
63665 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
63666  uint16x4_t __ret;
63667  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63668  return __ret;
63669 }
63670 #else
63671 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
63672  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63673  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63674  uint16x4_t __ret;
63675  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63676  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63677  return __ret;
63678 }
63679 #endif
63680 
63681 #ifdef __LITTLE_ENDIAN__
63682 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
63683  int8x8_t __ret;
63684  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
63685  return __ret;
63686 }
63687 #else
63688 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
63689  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63690  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63691  int8x8_t __ret;
63692  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
63693  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63694  return __ret;
63695 }
63696 #endif
63697 
63698 #ifdef __LITTLE_ENDIAN__
63699 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
63700  float32x2_t __ret;
63701  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63702  return __ret;
63703 }
63704 #else
63705 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
63706  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63707  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63708  float32x2_t __ret;
63709  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63710  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63711  return __ret;
63712 }
63713 #endif
63714 
63715 #ifdef __LITTLE_ENDIAN__
63716 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
63717  int32x2_t __ret;
63718  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63719  return __ret;
63720 }
63721 #else
63722 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
63723  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63724  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63725  int32x2_t __ret;
63726  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63728  return __ret;
63729 }
63730 #endif
63731 
63732 #ifdef __LITTLE_ENDIAN__
63733 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
63734  int16x4_t __ret;
63735  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
63736  return __ret;
63737 }
63738 #else
63739 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
63740  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63741  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63742  int16x4_t __ret;
63743  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
63744  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63745  return __ret;
63746 }
63747 #endif
63748 
63749 #ifdef __LITTLE_ENDIAN__
63750 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63751  poly8x8_t __ret;
63752  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
63753  return __ret;
63754 }
63755 #else
63756 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63757  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63758  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63759  poly8x8_t __ret;
63760  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
63761  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63762  return __ret;
63763 }
63764 #endif
63765 
63766 #ifdef __LITTLE_ENDIAN__
63767 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63768  poly16x4_t __ret;
63769  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
63770  return __ret;
63771 }
63772 #else
63773 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63774  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63775  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63776  poly16x4_t __ret;
63777  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
63778  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63779  return __ret;
63780 }
63781 #endif
63782 
63783 #ifdef __LITTLE_ENDIAN__
63784 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63785  poly8x16_t __ret;
63786  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63787  return __ret;
63788 }
63789 #else
63790 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63791  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63792  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63793  poly8x16_t __ret;
63794  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63795  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63796  return __ret;
63797 }
63798 #endif
63799 
63800 #ifdef __LITTLE_ENDIAN__
63801 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63802  poly64x2_t __ret;
63803  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63804  return __ret;
63805 }
63806 #else
63807 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63808  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63809  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63810  poly64x2_t __ret;
63811  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63812  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63813  return __ret;
63814 }
63815 #endif
63816 
63817 #ifdef __LITTLE_ENDIAN__
63818 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63819  poly16x8_t __ret;
63820  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
63821  return __ret;
63822 }
63823 #else
63824 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63825  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63826  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63827  poly16x8_t __ret;
63828  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
63829  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63830  return __ret;
63831 }
63832 #endif
63833 
63834 #ifdef __LITTLE_ENDIAN__
63835 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63836  uint8x16_t __ret;
63837  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63838  return __ret;
63839 }
63840 #else
63841 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63842  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63843  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63844  uint8x16_t __ret;
63845  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63846  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63847  return __ret;
63848 }
63849 #endif
63850 
63851 #ifdef __LITTLE_ENDIAN__
63852 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63853  uint32x4_t __ret;
63854  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
63855  return __ret;
63856 }
63857 #else
63858 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63859  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63860  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63861  uint32x4_t __ret;
63862  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
63863  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63864  return __ret;
63865 }
63866 #endif
63867 
63868 #ifdef __LITTLE_ENDIAN__
63869 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63870  uint64x2_t __ret;
63871  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63872  return __ret;
63873 }
63874 #else
63875 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63876  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63877  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63878  uint64x2_t __ret;
63879  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63880  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63881  return __ret;
63882 }
63883 #endif
63884 
63885 #ifdef __LITTLE_ENDIAN__
63886 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63887  uint16x8_t __ret;
63888  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
63889  return __ret;
63890 }
63891 #else
63892 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63893  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63894  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63895  uint16x8_t __ret;
63896  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
63897  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63898  return __ret;
63899 }
63900 #endif
63901 
63902 #ifdef __LITTLE_ENDIAN__
63903 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
63904  int8x16_t __ret;
63905  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63906  return __ret;
63907 }
63908 #else
63909 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
63910  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63911  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63912  int8x16_t __ret;
63913  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
63914  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63915  return __ret;
63916 }
63917 #endif
63918 
63919 #ifdef __LITTLE_ENDIAN__
63920 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
63921  float64x2_t __ret;
63922  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63923  return __ret;
63924 }
63925 #else
63926 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
63927  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63928  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63929  float64x2_t __ret;
63930  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63931  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63932  return __ret;
63933 }
63934 #endif
63935 
63936 #ifdef __LITTLE_ENDIAN__
63937 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
63938  float32x4_t __ret;
63939  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
63940  return __ret;
63941 }
63942 #else
63943 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
63944  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63945  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63946  float32x4_t __ret;
63947  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
63948  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63949  return __ret;
63950 }
63951 #endif
63952 
63953 #ifdef __LITTLE_ENDIAN__
63954 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
63955  int32x4_t __ret;
63956  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
63957  return __ret;
63958 }
63959 #else
63960 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
63961  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63962  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63963  int32x4_t __ret;
63964  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
63965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63966  return __ret;
63967 }
63968 #endif
63969 
63970 #ifdef __LITTLE_ENDIAN__
63971 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
63972  int64x2_t __ret;
63973  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63974  return __ret;
63975 }
63976 #else
63977 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
63978  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63979  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63980  int64x2_t __ret;
63981  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63983  return __ret;
63984 }
63985 #endif
63986 
63987 #ifdef __LITTLE_ENDIAN__
63988 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
63989  int16x8_t __ret;
63990  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
63991  return __ret;
63992 }
63993 #else
63994 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
63995  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63996  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63997  int16x8_t __ret;
63998  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
63999  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64000  return __ret;
64001 }
64002 #endif
64003 
64004 #ifdef __LITTLE_ENDIAN__
64005 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64006  uint8x8_t __ret;
64007  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
64008  return __ret;
64009 }
64010 #else
64011 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64012  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64013  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64014  uint8x8_t __ret;
64015  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
64016  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64017  return __ret;
64018 }
64019 #endif
64020 
64021 #ifdef __LITTLE_ENDIAN__
64022 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64023  uint32x2_t __ret;
64024  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64025  return __ret;
64026 }
64027 #else
64028 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64029  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64030  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64031  uint32x2_t __ret;
64032  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64034  return __ret;
64035 }
64036 #endif
64037 
64038 #ifdef __LITTLE_ENDIAN__
64039 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64040  uint16x4_t __ret;
64041  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
64042  return __ret;
64043 }
64044 #else
64045 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64046  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64047  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64048  uint16x4_t __ret;
64049  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
64050  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64051  return __ret;
64052 }
64053 #endif
64054 
64055 #ifdef __LITTLE_ENDIAN__
64056 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
64057  int8x8_t __ret;
64058  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
64059  return __ret;
64060 }
64061 #else
64062 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
64063  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64064  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64065  int8x8_t __ret;
64066  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
64067  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64068  return __ret;
64069 }
64070 #endif
64071 
64072 #ifdef __LITTLE_ENDIAN__
64073 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
64074  float32x2_t __ret;
64075  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64076  return __ret;
64077 }
64078 #else
64079 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
64080  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64081  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64082  float32x2_t __ret;
64083  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64084  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64085  return __ret;
64086 }
64087 #endif
64088 
64089 #ifdef __LITTLE_ENDIAN__
64090 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
64091  int32x2_t __ret;
64092  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64093  return __ret;
64094 }
64095 #else
64096 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
64097  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64098  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64099  int32x2_t __ret;
64100  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64102  return __ret;
64103 }
64104 #endif
64105 
64106 #ifdef __LITTLE_ENDIAN__
64107 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
64108  int16x4_t __ret;
64109  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
64110  return __ret;
64111 }
64112 #else
64113 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
64114  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64115  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64116  int16x4_t __ret;
64117  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
64118  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64119  return __ret;
64120 }
64121 #endif
64122 
64123 #endif
64124 #ifdef __LITTLE_ENDIAN__
64125 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64126  uint8x16_t __ret;
64127  __ret = __p0 + vabdq_u8(__p1, __p2);
64128  return __ret;
64129 }
64130 #else
64131 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
64132  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64133  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64134  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64135  uint8x16_t __ret;
64136  __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
64137  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64138  return __ret;
64139 }
64140 #endif
64141 
64142 #ifdef __LITTLE_ENDIAN__
64143 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64144  uint32x4_t __ret;
64145  __ret = __p0 + vabdq_u32(__p1, __p2);
64146  return __ret;
64147 }
64148 #else
64149 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
64150  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64151  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64152  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64153  uint32x4_t __ret;
64154  __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
64155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64156  return __ret;
64157 }
64158 #endif
64159 
64160 #ifdef __LITTLE_ENDIAN__
64161 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64162  uint16x8_t __ret;
64163  __ret = __p0 + vabdq_u16(__p1, __p2);
64164  return __ret;
64165 }
64166 #else
64167 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
64168  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64169  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64170  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64171  uint16x8_t __ret;
64172  __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
64173  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64174  return __ret;
64175 }
64176 #endif
64177 
64178 #ifdef __LITTLE_ENDIAN__
64179 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
64180  int8x16_t __ret;
64181  __ret = __p0 + vabdq_s8(__p1, __p2);
64182  return __ret;
64183 }
64184 #else
64185 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
64186  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64187  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64188  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64189  int8x16_t __ret;
64190  __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
64191  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64192  return __ret;
64193 }
64194 #endif
64195 
64196 #ifdef __LITTLE_ENDIAN__
64197 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
64198  int32x4_t __ret;
64199  __ret = __p0 + vabdq_s32(__p1, __p2);
64200  return __ret;
64201 }
64202 #else
64203 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
64204  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64205  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64206  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64207  int32x4_t __ret;
64208  __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
64209  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64210  return __ret;
64211 }
64212 #endif
64213 
64214 #ifdef __LITTLE_ENDIAN__
64215 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
64216  int16x8_t __ret;
64217  __ret = __p0 + vabdq_s16(__p1, __p2);
64218  return __ret;
64219 }
64220 #else
64221 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
64222  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64223  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64224  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64225  int16x8_t __ret;
64226  __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
64227  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64228  return __ret;
64229 }
64230 #endif
64231 
64232 #ifdef __LITTLE_ENDIAN__
64233 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
64234  uint8x8_t __ret;
64235  __ret = __p0 + vabd_u8(__p1, __p2);
64236  return __ret;
64237 }
64238 #else
64239 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
64240  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64241  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64242  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64243  uint8x8_t __ret;
64244  __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
64245  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64246  return __ret;
64247 }
64248 #endif
64249 
64250 #ifdef __LITTLE_ENDIAN__
64251 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
64252  uint32x2_t __ret;
64253  __ret = __p0 + vabd_u32(__p1, __p2);
64254  return __ret;
64255 }
64256 #else
64257 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
64258  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64259  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64260  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64261  uint32x2_t __ret;
64262  __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
64263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64264  return __ret;
64265 }
64266 #endif
64267 
64268 #ifdef __LITTLE_ENDIAN__
64269 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
64270  uint16x4_t __ret;
64271  __ret = __p0 + vabd_u16(__p1, __p2);
64272  return __ret;
64273 }
64274 #else
64275 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
64276  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64277  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64278  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64279  uint16x4_t __ret;
64280  __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
64281  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64282  return __ret;
64283 }
64284 #endif
64285 
64286 #ifdef __LITTLE_ENDIAN__
64287 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
64288  int8x8_t __ret;
64289  __ret = __p0 + vabd_s8(__p1, __p2);
64290  return __ret;
64291 }
64292 #else
64293 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
64294  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64295  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64296  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64297  int8x8_t __ret;
64298  __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
64299  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64300  return __ret;
64301 }
64302 #endif
64303 
64304 #ifdef __LITTLE_ENDIAN__
64305 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
64306  int32x2_t __ret;
64307  __ret = __p0 + vabd_s32(__p1, __p2);
64308  return __ret;
64309 }
64310 #else
64311 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
64312  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64313  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64314  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64315  int32x2_t __ret;
64316  __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
64317  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64318  return __ret;
64319 }
64320 #endif
64321 
64322 #ifdef __LITTLE_ENDIAN__
64323 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
64324  int16x4_t __ret;
64325  __ret = __p0 + vabd_s16(__p1, __p2);
64326  return __ret;
64327 }
64328 #else
64329 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
64330  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64331  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64332  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64333  int16x4_t __ret;
64334  __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
64335  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64336  return __ret;
64337 }
64338 #endif
64339 
64340 #ifdef __LITTLE_ENDIAN__
64341 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
64342  uint16x8_t __ret;
64343  __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
64344  return __ret;
64345 }
64346 #else
64347 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
64348  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64349  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64350  uint16x8_t __ret;
64351  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
64352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64353  return __ret;
64354 }
64355 __ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
64356  uint16x8_t __ret;
64357  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
64358  return __ret;
64359 }
64360 #endif
64361 
64362 #ifdef __LITTLE_ENDIAN__
64363 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
64364  uint64x2_t __ret;
64365  __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
64366  return __ret;
64367 }
64368 #else
64369 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
64370  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64371  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64372  uint64x2_t __ret;
64373  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
64374  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64375  return __ret;
64376 }
64377 __ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
64378  uint64x2_t __ret;
64379  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
64380  return __ret;
64381 }
64382 #endif
64383 
64384 #ifdef __LITTLE_ENDIAN__
64385 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
64386  uint32x4_t __ret;
64387  __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
64388  return __ret;
64389 }
64390 #else
64391 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
64392  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64393  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64394  uint32x4_t __ret;
64395  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
64396  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64397  return __ret;
64398 }
64399 __ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
64400  uint32x4_t __ret;
64401  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
64402  return __ret;
64403 }
64404 #endif
64405 
64406 #ifdef __LITTLE_ENDIAN__
64407 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
64408  int16x8_t __ret;
64409  __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
64410  return __ret;
64411 }
64412 #else
64413 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
64414  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64415  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64416  int16x8_t __ret;
64417  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
64418  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64419  return __ret;
64420 }
64421 __ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
64422  int16x8_t __ret;
64423  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
64424  return __ret;
64425 }
64426 #endif
64427 
64428 #ifdef __LITTLE_ENDIAN__
64429 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
64430  int64x2_t __ret;
64431  __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
64432  return __ret;
64433 }
64434 #else
64435 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
64436  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64437  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64438  int64x2_t __ret;
64439  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
64440  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64441  return __ret;
64442 }
64443 __ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
64444  int64x2_t __ret;
64445  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
64446  return __ret;
64447 }
64448 #endif
64449 
64450 #ifdef __LITTLE_ENDIAN__
64451 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
64452  int32x4_t __ret;
64453  __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
64454  return __ret;
64455 }
64456 #else
64457 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
64458  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64459  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64460  int32x4_t __ret;
64461  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
64462  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64463  return __ret;
64464 }
64465 __ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
64466  int32x4_t __ret;
64467  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
64468  return __ret;
64469 }
64470 #endif
64471 
64472 #ifdef __LITTLE_ENDIAN__
64473 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
64474  uint16x8_t __ret;
64475  __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
64476  return __ret;
64477 }
64478 #else
64479 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
64480  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64481  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64482  uint16x8_t __ret;
64483  __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
64484  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64485  return __ret;
64486 }
64487 #endif
64488 
64489 #ifdef __LITTLE_ENDIAN__
64490 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
64491  uint64x2_t __ret;
64492  __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
64493  return __ret;
64494 }
64495 #else
64496 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
64497  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64498  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64499  uint64x2_t __ret;
64500  __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
64501  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64502  return __ret;
64503 }
64504 #endif
64505 
64506 #ifdef __LITTLE_ENDIAN__
64507 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
64508  uint32x4_t __ret;
64509  __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
64510  return __ret;
64511 }
64512 #else
64513 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
64514  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64515  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64516  uint32x4_t __ret;
64517  __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
64518  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64519  return __ret;
64520 }
64521 #endif
64522 
64523 #ifdef __LITTLE_ENDIAN__
64524 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
64525  int16x8_t __ret;
64526  __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
64527  return __ret;
64528 }
64529 #else
64530 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
64531  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64532  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64533  int16x8_t __ret;
64534  __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
64535  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64536  return __ret;
64537 }
64538 #endif
64539 
64540 #ifdef __LITTLE_ENDIAN__
64541 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
64542  int64x2_t __ret;
64543  __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
64544  return __ret;
64545 }
64546 #else
64547 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
64548  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64549  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64550  int64x2_t __ret;
64551  __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
64552  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64553  return __ret;
64554 }
64555 #endif
64556 
64557 #ifdef __LITTLE_ENDIAN__
64558 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
64559  int32x4_t __ret;
64560  __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
64561  return __ret;
64562 }
64563 #else
64564 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
64565  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64566  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64567  int32x4_t __ret;
64568  __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
64569  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64570  return __ret;
64571 }
64572 #endif
64573 
64574 #ifdef __LITTLE_ENDIAN__
64575 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
64576  uint16x8_t __ret;
64577  __ret = __p0 + vmovl_u8(__p1);
64578  return __ret;
64579 }
64580 #else
64581 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
64582  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64583  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64584  uint16x8_t __ret;
64585  __ret = __rev0 + __noswap_vmovl_u8(__rev1);
64586  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64587  return __ret;
64588 }
64589 #endif
64590 
64591 #ifdef __LITTLE_ENDIAN__
64592 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
64593  uint64x2_t __ret;
64594  __ret = __p0 + vmovl_u32(__p1);
64595  return __ret;
64596 }
64597 #else
64598 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
64599  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64600  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64601  uint64x2_t __ret;
64602  __ret = __rev0 + __noswap_vmovl_u32(__rev1);
64603  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64604  return __ret;
64605 }
64606 #endif
64607 
64608 #ifdef __LITTLE_ENDIAN__
64609 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
64610  uint32x4_t __ret;
64611  __ret = __p0 + vmovl_u16(__p1);
64612  return __ret;
64613 }
64614 #else
64615 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
64616  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64617  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64618  uint32x4_t __ret;
64619  __ret = __rev0 + __noswap_vmovl_u16(__rev1);
64620  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64621  return __ret;
64622 }
64623 #endif
64624 
64625 #ifdef __LITTLE_ENDIAN__
64626 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
64627  int16x8_t __ret;
64628  __ret = __p0 + vmovl_s8(__p1);
64629  return __ret;
64630 }
64631 #else
64632 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
64633  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64634  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64635  int16x8_t __ret;
64636  __ret = __rev0 + __noswap_vmovl_s8(__rev1);
64637  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64638  return __ret;
64639 }
64640 #endif
64641 
64642 #ifdef __LITTLE_ENDIAN__
64643 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
64644  int64x2_t __ret;
64645  __ret = __p0 + vmovl_s32(__p1);
64646  return __ret;
64647 }
64648 #else
64649 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
64650  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64651  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64652  int64x2_t __ret;
64653  __ret = __rev0 + __noswap_vmovl_s32(__rev1);
64654  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64655  return __ret;
64656 }
64657 #endif
64658 
64659 #ifdef __LITTLE_ENDIAN__
64660 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
64661  int32x4_t __ret;
64662  __ret = __p0 + vmovl_s16(__p1);
64663  return __ret;
64664 }
64665 #else
64666 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
64667  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64668  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64669  int32x4_t __ret;
64670  __ret = __rev0 + __noswap_vmovl_s16(__rev1);
64671  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64672  return __ret;
64673 }
64674 #endif
64675 
64676 #ifdef __LITTLE_ENDIAN__
64677 #define vget_lane_f16(__p0_743, __p1_743) __extension__ ({ \
64678  float16x4_t __s0_743 = __p0_743; \
64679  float16_t __ret_743; \
64680 float16x4_t __reint_743 = __s0_743; \
64681 int16_t __reint1_743 = vget_lane_s16(*(int16x4_t *) &__reint_743, __p1_743); \
64682  __ret_743 = *(float16_t *) &__reint1_743; \
64683  __ret_743; \
64684 })
64685 #else
64686 #define vget_lane_f16(__p0_744, __p1_744) __extension__ ({ \
64687  float16x4_t __s0_744 = __p0_744; \
64688  float16x4_t __rev0_744; __rev0_744 = __builtin_shufflevector(__s0_744, __s0_744, 3, 2, 1, 0); \
64689  float16_t __ret_744; \
64690 float16x4_t __reint_744 = __rev0_744; \
64691 int16_t __reint1_744 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_744, __p1_744); \
64692  __ret_744 = *(float16_t *) &__reint1_744; \
64693  __ret_744; \
64694 })
64695 #define __noswap_vget_lane_f16(__p0_745, __p1_745) __extension__ ({ \
64696  float16x4_t __s0_745 = __p0_745; \
64697  float16_t __ret_745; \
64698 float16x4_t __reint_745 = __s0_745; \
64699 int16_t __reint1_745 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_745, __p1_745); \
64700  __ret_745 = *(float16_t *) &__reint1_745; \
64701  __ret_745; \
64702 })
64703 #endif
64704 
64705 #ifdef __LITTLE_ENDIAN__
64706 #define vgetq_lane_f16(__p0_746, __p1_746) __extension__ ({ \
64707  float16x8_t __s0_746 = __p0_746; \
64708  float16_t __ret_746; \
64709 float16x8_t __reint_746 = __s0_746; \
64710 int16_t __reint1_746 = vgetq_lane_s16(*(int16x8_t *) &__reint_746, __p1_746); \
64711  __ret_746 = *(float16_t *) &__reint1_746; \
64712  __ret_746; \
64713 })
64714 #else
64715 #define vgetq_lane_f16(__p0_747, __p1_747) __extension__ ({ \
64716  float16x8_t __s0_747 = __p0_747; \
64717  float16x8_t __rev0_747; __rev0_747 = __builtin_shufflevector(__s0_747, __s0_747, 7, 6, 5, 4, 3, 2, 1, 0); \
64718  float16_t __ret_747; \
64719 float16x8_t __reint_747 = __rev0_747; \
64720 int16_t __reint1_747 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_747, __p1_747); \
64721  __ret_747 = *(float16_t *) &__reint1_747; \
64722  __ret_747; \
64723 })
64724 #define __noswap_vgetq_lane_f16(__p0_748, __p1_748) __extension__ ({ \
64725  float16x8_t __s0_748 = __p0_748; \
64726  float16_t __ret_748; \
64727 float16x8_t __reint_748 = __s0_748; \
64728 int16_t __reint1_748 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_748, __p1_748); \
64729  __ret_748 = *(float16_t *) &__reint1_748; \
64730  __ret_748; \
64731 })
64732 #endif
64733 
64734 #ifdef __LITTLE_ENDIAN__
64735 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
64736  uint16x8_t __ret;
64737  __ret = __p0 + vmull_u8(__p1, __p2);
64738  return __ret;
64739 }
64740 #else
64741 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
64742  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64743  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64744  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64745  uint16x8_t __ret;
64746  __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
64747  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64748  return __ret;
64749 }
64750 __ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
64751  uint16x8_t __ret;
64752  __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
64753  return __ret;
64754 }
64755 #endif
64756 
64757 #ifdef __LITTLE_ENDIAN__
64758 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
64759  uint64x2_t __ret;
64760  __ret = __p0 + vmull_u32(__p1, __p2);
64761  return __ret;
64762 }
64763 #else
64764 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
64765  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64766  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64767  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64768  uint64x2_t __ret;
64769  __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
64770  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64771  return __ret;
64772 }
64773 __ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
64774  uint64x2_t __ret;
64775  __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
64776  return __ret;
64777 }
64778 #endif
64779 
64780 #ifdef __LITTLE_ENDIAN__
64781 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
64782  uint32x4_t __ret;
64783  __ret = __p0 + vmull_u16(__p1, __p2);
64784  return __ret;
64785 }
64786 #else
64787 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
64788  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64789  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64790  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64791  uint32x4_t __ret;
64792  __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
64793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64794  return __ret;
64795 }
64796 __ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
64797  uint32x4_t __ret;
64798  __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
64799  return __ret;
64800 }
64801 #endif
64802 
64803 #ifdef __LITTLE_ENDIAN__
64804 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
64805  int16x8_t __ret;
64806  __ret = __p0 + vmull_s8(__p1, __p2);
64807  return __ret;
64808 }
64809 #else
64810 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
64811  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64812  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64813  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
64814  int16x8_t __ret;
64815  __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
64816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64817  return __ret;
64818 }
64819 __ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
64820  int16x8_t __ret;
64821  __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
64822  return __ret;
64823 }
64824 #endif
64825 
64826 #ifdef __LITTLE_ENDIAN__
64827 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
64828  int64x2_t __ret;
64829  __ret = __p0 + vmull_s32(__p1, __p2);
64830  return __ret;
64831 }
64832 #else
64833 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
64834  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64835  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64836  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
64837  int64x2_t __ret;
64838  __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
64839  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64840  return __ret;
64841 }
64842 __ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
64843  int64x2_t __ret;
64844  __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
64845  return __ret;
64846 }
64847 #endif
64848 
64849 #ifdef __LITTLE_ENDIAN__
64850 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
64851  int32x4_t __ret;
64852  __ret = __p0 + vmull_s16(__p1, __p2);
64853  return __ret;
64854 }
64855 #else
64856 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
64857  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64858  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64859  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
64860  int32x4_t __ret;
64861  __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
64862  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64863  return __ret;
64864 }
64865 __ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
64866  int32x4_t __ret;
64867  __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
64868  return __ret;
64869 }
64870 #endif
64871 
64872 #ifdef __LITTLE_ENDIAN__
64873 #define vmlal_lane_u32(__p0_749, __p1_749, __p2_749, __p3_749) __extension__ ({ \
64874  uint64x2_t __s0_749 = __p0_749; \
64875  uint32x2_t __s1_749 = __p1_749; \
64876  uint32x2_t __s2_749 = __p2_749; \
64877  uint64x2_t __ret_749; \
64878  __ret_749 = __s0_749 + vmull_u32(__s1_749, splat_lane_u32(__s2_749, __p3_749)); \
64879  __ret_749; \
64880 })
64881 #else
64882 #define vmlal_lane_u32(__p0_750, __p1_750, __p2_750, __p3_750) __extension__ ({ \
64883  uint64x2_t __s0_750 = __p0_750; \
64884  uint32x2_t __s1_750 = __p1_750; \
64885  uint32x2_t __s2_750 = __p2_750; \
64886  uint64x2_t __rev0_750; __rev0_750 = __builtin_shufflevector(__s0_750, __s0_750, 1, 0); \
64887  uint32x2_t __rev1_750; __rev1_750 = __builtin_shufflevector(__s1_750, __s1_750, 1, 0); \
64888  uint32x2_t __rev2_750; __rev2_750 = __builtin_shufflevector(__s2_750, __s2_750, 1, 0); \
64889  uint64x2_t __ret_750; \
64890  __ret_750 = __rev0_750 + __noswap_vmull_u32(__rev1_750, __noswap_splat_lane_u32(__rev2_750, __p3_750)); \
64891  __ret_750 = __builtin_shufflevector(__ret_750, __ret_750, 1, 0); \
64892  __ret_750; \
64893 })
64894 #endif
64895 
64896 #ifdef __LITTLE_ENDIAN__
64897 #define vmlal_lane_u16(__p0_751, __p1_751, __p2_751, __p3_751) __extension__ ({ \
64898  uint32x4_t __s0_751 = __p0_751; \
64899  uint16x4_t __s1_751 = __p1_751; \
64900  uint16x4_t __s2_751 = __p2_751; \
64901  uint32x4_t __ret_751; \
64902  __ret_751 = __s0_751 + vmull_u16(__s1_751, splat_lane_u16(__s2_751, __p3_751)); \
64903  __ret_751; \
64904 })
64905 #else
64906 #define vmlal_lane_u16(__p0_752, __p1_752, __p2_752, __p3_752) __extension__ ({ \
64907  uint32x4_t __s0_752 = __p0_752; \
64908  uint16x4_t __s1_752 = __p1_752; \
64909  uint16x4_t __s2_752 = __p2_752; \
64910  uint32x4_t __rev0_752; __rev0_752 = __builtin_shufflevector(__s0_752, __s0_752, 3, 2, 1, 0); \
64911  uint16x4_t __rev1_752; __rev1_752 = __builtin_shufflevector(__s1_752, __s1_752, 3, 2, 1, 0); \
64912  uint16x4_t __rev2_752; __rev2_752 = __builtin_shufflevector(__s2_752, __s2_752, 3, 2, 1, 0); \
64913  uint32x4_t __ret_752; \
64914  __ret_752 = __rev0_752 + __noswap_vmull_u16(__rev1_752, __noswap_splat_lane_u16(__rev2_752, __p3_752)); \
64915  __ret_752 = __builtin_shufflevector(__ret_752, __ret_752, 3, 2, 1, 0); \
64916  __ret_752; \
64917 })
64918 #endif
64919 
64920 #ifdef __LITTLE_ENDIAN__
64921 #define vmlal_lane_s32(__p0_753, __p1_753, __p2_753, __p3_753) __extension__ ({ \
64922  int64x2_t __s0_753 = __p0_753; \
64923  int32x2_t __s1_753 = __p1_753; \
64924  int32x2_t __s2_753 = __p2_753; \
64925  int64x2_t __ret_753; \
64926  __ret_753 = __s0_753 + vmull_s32(__s1_753, splat_lane_s32(__s2_753, __p3_753)); \
64927  __ret_753; \
64928 })
64929 #else
64930 #define vmlal_lane_s32(__p0_754, __p1_754, __p2_754, __p3_754) __extension__ ({ \
64931  int64x2_t __s0_754 = __p0_754; \
64932  int32x2_t __s1_754 = __p1_754; \
64933  int32x2_t __s2_754 = __p2_754; \
64934  int64x2_t __rev0_754; __rev0_754 = __builtin_shufflevector(__s0_754, __s0_754, 1, 0); \
64935  int32x2_t __rev1_754; __rev1_754 = __builtin_shufflevector(__s1_754, __s1_754, 1, 0); \
64936  int32x2_t __rev2_754; __rev2_754 = __builtin_shufflevector(__s2_754, __s2_754, 1, 0); \
64937  int64x2_t __ret_754; \
64938  __ret_754 = __rev0_754 + __noswap_vmull_s32(__rev1_754, __noswap_splat_lane_s32(__rev2_754, __p3_754)); \
64939  __ret_754 = __builtin_shufflevector(__ret_754, __ret_754, 1, 0); \
64940  __ret_754; \
64941 })
64942 #endif
64943 
64944 #ifdef __LITTLE_ENDIAN__
64945 #define vmlal_lane_s16(__p0_755, __p1_755, __p2_755, __p3_755) __extension__ ({ \
64946  int32x4_t __s0_755 = __p0_755; \
64947  int16x4_t __s1_755 = __p1_755; \
64948  int16x4_t __s2_755 = __p2_755; \
64949  int32x4_t __ret_755; \
64950  __ret_755 = __s0_755 + vmull_s16(__s1_755, splat_lane_s16(__s2_755, __p3_755)); \
64951  __ret_755; \
64952 })
64953 #else
64954 #define vmlal_lane_s16(__p0_756, __p1_756, __p2_756, __p3_756) __extension__ ({ \
64955  int32x4_t __s0_756 = __p0_756; \
64956  int16x4_t __s1_756 = __p1_756; \
64957  int16x4_t __s2_756 = __p2_756; \
64958  int32x4_t __rev0_756; __rev0_756 = __builtin_shufflevector(__s0_756, __s0_756, 3, 2, 1, 0); \
64959  int16x4_t __rev1_756; __rev1_756 = __builtin_shufflevector(__s1_756, __s1_756, 3, 2, 1, 0); \
64960  int16x4_t __rev2_756; __rev2_756 = __builtin_shufflevector(__s2_756, __s2_756, 3, 2, 1, 0); \
64961  int32x4_t __ret_756; \
64962  __ret_756 = __rev0_756 + __noswap_vmull_s16(__rev1_756, __noswap_splat_lane_s16(__rev2_756, __p3_756)); \
64963  __ret_756 = __builtin_shufflevector(__ret_756, __ret_756, 3, 2, 1, 0); \
64964  __ret_756; \
64965 })
64966 #endif
64967 
64968 #ifdef __LITTLE_ENDIAN__
64969 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
64970  uint64x2_t __ret;
64971  __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
64972  return __ret;
64973 }
64974 #else
64975 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
64976  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64977  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64978  uint64x2_t __ret;
64979  __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
64980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64981  return __ret;
64982 }
64983 __ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
64984  uint64x2_t __ret;
64985  __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
64986  return __ret;
64987 }
64988 #endif
64989 
64990 #ifdef __LITTLE_ENDIAN__
64991 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
64992  uint32x4_t __ret;
64993  __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
64994  return __ret;
64995 }
64996 #else
64997 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
64998  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64999  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65000  uint32x4_t __ret;
65001  __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
65002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65003  return __ret;
65004 }
65005 __ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
65006  uint32x4_t __ret;
65007  __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
65008  return __ret;
65009 }
65010 #endif
65011 
65012 #ifdef __LITTLE_ENDIAN__
65013 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65014  int64x2_t __ret;
65015  __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
65016  return __ret;
65017 }
65018 #else
65019 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65020  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65021  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65022  int64x2_t __ret;
65023  __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
65024  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65025  return __ret;
65026 }
65027 __ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65028  int64x2_t __ret;
65029  __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
65030  return __ret;
65031 }
65032 #endif
65033 
65034 #ifdef __LITTLE_ENDIAN__
65035 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65036  int32x4_t __ret;
65037  __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
65038  return __ret;
65039 }
65040 #else
65041 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65042  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65043  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65044  int32x4_t __ret;
65045  __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
65046  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65047  return __ret;
65048 }
65049 __ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65050  int32x4_t __ret;
65051  __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
65052  return __ret;
65053 }
65054 #endif
65055 
65056 #ifdef __LITTLE_ENDIAN__
65057 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
65058  uint16x8_t __ret;
65059  __ret = __p0 - vmull_u8(__p1, __p2);
65060  return __ret;
65061 }
65062 #else
65063 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
65064  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65065  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65066  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
65067  uint16x8_t __ret;
65068  __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
65069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65070  return __ret;
65071 }
65072 __ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
65073  uint16x8_t __ret;
65074  __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
65075  return __ret;
65076 }
65077 #endif
65078 
65079 #ifdef __LITTLE_ENDIAN__
65080 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
65081  uint64x2_t __ret;
65082  __ret = __p0 - vmull_u32(__p1, __p2);
65083  return __ret;
65084 }
65085 #else
65086 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
65087  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65088  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65089  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65090  uint64x2_t __ret;
65091  __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
65092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65093  return __ret;
65094 }
65095 __ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
65096  uint64x2_t __ret;
65097  __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
65098  return __ret;
65099 }
65100 #endif
65101 
65102 #ifdef __LITTLE_ENDIAN__
65103 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
65104  uint32x4_t __ret;
65105  __ret = __p0 - vmull_u16(__p1, __p2);
65106  return __ret;
65107 }
65108 #else
65109 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
65110  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65111  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65112  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65113  uint32x4_t __ret;
65114  __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
65115  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65116  return __ret;
65117 }
65118 __ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
65119  uint32x4_t __ret;
65120  __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
65121  return __ret;
65122 }
65123 #endif
65124 
65125 #ifdef __LITTLE_ENDIAN__
65126 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
65127  int16x8_t __ret;
65128  __ret = __p0 - vmull_s8(__p1, __p2);
65129  return __ret;
65130 }
65131 #else
65132 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
65133  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65134  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65135  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
65136  int16x8_t __ret;
65137  __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
65138  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65139  return __ret;
65140 }
65141 __ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
65142  int16x8_t __ret;
65143  __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
65144  return __ret;
65145 }
65146 #endif
65147 
65148 #ifdef __LITTLE_ENDIAN__
65149 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
65150  int64x2_t __ret;
65151  __ret = __p0 - vmull_s32(__p1, __p2);
65152  return __ret;
65153 }
65154 #else
65155 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
65156  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65157  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65158  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
65159  int64x2_t __ret;
65160  __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
65161  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65162  return __ret;
65163 }
65164 __ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
65165  int64x2_t __ret;
65166  __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
65167  return __ret;
65168 }
65169 #endif
65170 
65171 #ifdef __LITTLE_ENDIAN__
65172 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
65173  int32x4_t __ret;
65174  __ret = __p0 - vmull_s16(__p1, __p2);
65175  return __ret;
65176 }
65177 #else
65178 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
65179  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65180  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65181  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
65182  int32x4_t __ret;
65183  __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
65184  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65185  return __ret;
65186 }
65187 __ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
65188  int32x4_t __ret;
65189  __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
65190  return __ret;
65191 }
65192 #endif
65193 
65194 #ifdef __LITTLE_ENDIAN__
65195 #define vmlsl_lane_u32(__p0_757, __p1_757, __p2_757, __p3_757) __extension__ ({ \
65196  uint64x2_t __s0_757 = __p0_757; \
65197  uint32x2_t __s1_757 = __p1_757; \
65198  uint32x2_t __s2_757 = __p2_757; \
65199  uint64x2_t __ret_757; \
65200  __ret_757 = __s0_757 - vmull_u32(__s1_757, splat_lane_u32(__s2_757, __p3_757)); \
65201  __ret_757; \
65202 })
65203 #else
65204 #define vmlsl_lane_u32(__p0_758, __p1_758, __p2_758, __p3_758) __extension__ ({ \
65205  uint64x2_t __s0_758 = __p0_758; \
65206  uint32x2_t __s1_758 = __p1_758; \
65207  uint32x2_t __s2_758 = __p2_758; \
65208  uint64x2_t __rev0_758; __rev0_758 = __builtin_shufflevector(__s0_758, __s0_758, 1, 0); \
65209  uint32x2_t __rev1_758; __rev1_758 = __builtin_shufflevector(__s1_758, __s1_758, 1, 0); \
65210  uint32x2_t __rev2_758; __rev2_758 = __builtin_shufflevector(__s2_758, __s2_758, 1, 0); \
65211  uint64x2_t __ret_758; \
65212  __ret_758 = __rev0_758 - __noswap_vmull_u32(__rev1_758, __noswap_splat_lane_u32(__rev2_758, __p3_758)); \
65213  __ret_758 = __builtin_shufflevector(__ret_758, __ret_758, 1, 0); \
65214  __ret_758; \
65215 })
65216 #endif
65217 
65218 #ifdef __LITTLE_ENDIAN__
65219 #define vmlsl_lane_u16(__p0_759, __p1_759, __p2_759, __p3_759) __extension__ ({ \
65220  uint32x4_t __s0_759 = __p0_759; \
65221  uint16x4_t __s1_759 = __p1_759; \
65222  uint16x4_t __s2_759 = __p2_759; \
65223  uint32x4_t __ret_759; \
65224  __ret_759 = __s0_759 - vmull_u16(__s1_759, splat_lane_u16(__s2_759, __p3_759)); \
65225  __ret_759; \
65226 })
65227 #else
65228 #define vmlsl_lane_u16(__p0_760, __p1_760, __p2_760, __p3_760) __extension__ ({ \
65229  uint32x4_t __s0_760 = __p0_760; \
65230  uint16x4_t __s1_760 = __p1_760; \
65231  uint16x4_t __s2_760 = __p2_760; \
65232  uint32x4_t __rev0_760; __rev0_760 = __builtin_shufflevector(__s0_760, __s0_760, 3, 2, 1, 0); \
65233  uint16x4_t __rev1_760; __rev1_760 = __builtin_shufflevector(__s1_760, __s1_760, 3, 2, 1, 0); \
65234  uint16x4_t __rev2_760; __rev2_760 = __builtin_shufflevector(__s2_760, __s2_760, 3, 2, 1, 0); \
65235  uint32x4_t __ret_760; \
65236  __ret_760 = __rev0_760 - __noswap_vmull_u16(__rev1_760, __noswap_splat_lane_u16(__rev2_760, __p3_760)); \
65237  __ret_760 = __builtin_shufflevector(__ret_760, __ret_760, 3, 2, 1, 0); \
65238  __ret_760; \
65239 })
65240 #endif
65241 
65242 #ifdef __LITTLE_ENDIAN__
65243 #define vmlsl_lane_s32(__p0_761, __p1_761, __p2_761, __p3_761) __extension__ ({ \
65244  int64x2_t __s0_761 = __p0_761; \
65245  int32x2_t __s1_761 = __p1_761; \
65246  int32x2_t __s2_761 = __p2_761; \
65247  int64x2_t __ret_761; \
65248  __ret_761 = __s0_761 - vmull_s32(__s1_761, splat_lane_s32(__s2_761, __p3_761)); \
65249  __ret_761; \
65250 })
65251 #else
65252 #define vmlsl_lane_s32(__p0_762, __p1_762, __p2_762, __p3_762) __extension__ ({ \
65253  int64x2_t __s0_762 = __p0_762; \
65254  int32x2_t __s1_762 = __p1_762; \
65255  int32x2_t __s2_762 = __p2_762; \
65256  int64x2_t __rev0_762; __rev0_762 = __builtin_shufflevector(__s0_762, __s0_762, 1, 0); \
65257  int32x2_t __rev1_762; __rev1_762 = __builtin_shufflevector(__s1_762, __s1_762, 1, 0); \
65258  int32x2_t __rev2_762; __rev2_762 = __builtin_shufflevector(__s2_762, __s2_762, 1, 0); \
65259  int64x2_t __ret_762; \
65260  __ret_762 = __rev0_762 - __noswap_vmull_s32(__rev1_762, __noswap_splat_lane_s32(__rev2_762, __p3_762)); \
65261  __ret_762 = __builtin_shufflevector(__ret_762, __ret_762, 1, 0); \
65262  __ret_762; \
65263 })
65264 #endif
65265 
65266 #ifdef __LITTLE_ENDIAN__
65267 #define vmlsl_lane_s16(__p0_763, __p1_763, __p2_763, __p3_763) __extension__ ({ \
65268  int32x4_t __s0_763 = __p0_763; \
65269  int16x4_t __s1_763 = __p1_763; \
65270  int16x4_t __s2_763 = __p2_763; \
65271  int32x4_t __ret_763; \
65272  __ret_763 = __s0_763 - vmull_s16(__s1_763, splat_lane_s16(__s2_763, __p3_763)); \
65273  __ret_763; \
65274 })
65275 #else
65276 #define vmlsl_lane_s16(__p0_764, __p1_764, __p2_764, __p3_764) __extension__ ({ \
65277  int32x4_t __s0_764 = __p0_764; \
65278  int16x4_t __s1_764 = __p1_764; \
65279  int16x4_t __s2_764 = __p2_764; \
65280  int32x4_t __rev0_764; __rev0_764 = __builtin_shufflevector(__s0_764, __s0_764, 3, 2, 1, 0); \
65281  int16x4_t __rev1_764; __rev1_764 = __builtin_shufflevector(__s1_764, __s1_764, 3, 2, 1, 0); \
65282  int16x4_t __rev2_764; __rev2_764 = __builtin_shufflevector(__s2_764, __s2_764, 3, 2, 1, 0); \
65283  int32x4_t __ret_764; \
65284  __ret_764 = __rev0_764 - __noswap_vmull_s16(__rev1_764, __noswap_splat_lane_s16(__rev2_764, __p3_764)); \
65285  __ret_764 = __builtin_shufflevector(__ret_764, __ret_764, 3, 2, 1, 0); \
65286  __ret_764; \
65287 })
65288 #endif
65289 
65290 #ifdef __LITTLE_ENDIAN__
65291 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
65292  uint64x2_t __ret;
65293  __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
65294  return __ret;
65295 }
65296 #else
65297 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
65298  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65299  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65300  uint64x2_t __ret;
65301  __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
65302  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65303  return __ret;
65304 }
65305 __ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
65306  uint64x2_t __ret;
65307  __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
65308  return __ret;
65309 }
65310 #endif
65311 
65312 #ifdef __LITTLE_ENDIAN__
65313 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
65314  uint32x4_t __ret;
65315  __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
65316  return __ret;
65317 }
65318 #else
65319 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
65320  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65321  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65322  uint32x4_t __ret;
65323  __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
65324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65325  return __ret;
65326 }
65327 __ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
65328  uint32x4_t __ret;
65329  __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
65330  return __ret;
65331 }
65332 #endif
65333 
65334 #ifdef __LITTLE_ENDIAN__
65335 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65336  int64x2_t __ret;
65337  __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
65338  return __ret;
65339 }
65340 #else
65341 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65342  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65343  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65344  int64x2_t __ret;
65345  __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
65346  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65347  return __ret;
65348 }
65349 __ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
65350  int64x2_t __ret;
65351  __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
65352  return __ret;
65353 }
65354 #endif
65355 
65356 #ifdef __LITTLE_ENDIAN__
65357 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65358  int32x4_t __ret;
65359  __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
65360  return __ret;
65361 }
65362 #else
65363 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65364  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65365  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65366  int32x4_t __ret;
65367  __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
65368  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65369  return __ret;
65370 }
65371 __ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
65372  int32x4_t __ret;
65373  __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
65374  return __ret;
65375 }
65376 #endif
65377 
65378 #ifdef __LITTLE_ENDIAN__
65379 #define vset_lane_f16(__p0_765, __p1_765, __p2_765) __extension__ ({ \
65380  float16_t __s0_765 = __p0_765; \
65381  float16x4_t __s1_765 = __p1_765; \
65382  float16x4_t __ret_765; \
65383 float16_t __reint_765 = __s0_765; \
65384 float16x4_t __reint1_765 = __s1_765; \
65385 int16x4_t __reint2_765 = vset_lane_s16(*(int16_t *) &__reint_765, *(int16x4_t *) &__reint1_765, __p2_765); \
65386  __ret_765 = *(float16x4_t *) &__reint2_765; \
65387  __ret_765; \
65388 })
65389 #else
65390 #define vset_lane_f16(__p0_766, __p1_766, __p2_766) __extension__ ({ \
65391  float16_t __s0_766 = __p0_766; \
65392  float16x4_t __s1_766 = __p1_766; \
65393  float16x4_t __rev1_766; __rev1_766 = __builtin_shufflevector(__s1_766, __s1_766, 3, 2, 1, 0); \
65394  float16x4_t __ret_766; \
65395 float16_t __reint_766 = __s0_766; \
65396 float16x4_t __reint1_766 = __rev1_766; \
65397 int16x4_t __reint2_766 = __noswap_vset_lane_s16(*(int16_t *) &__reint_766, *(int16x4_t *) &__reint1_766, __p2_766); \
65398  __ret_766 = *(float16x4_t *) &__reint2_766; \
65399  __ret_766 = __builtin_shufflevector(__ret_766, __ret_766, 3, 2, 1, 0); \
65400  __ret_766; \
65401 })
65402 #endif
65403 
65404 #ifdef __LITTLE_ENDIAN__
65405 #define vsetq_lane_f16(__p0_767, __p1_767, __p2_767) __extension__ ({ \
65406  float16_t __s0_767 = __p0_767; \
65407  float16x8_t __s1_767 = __p1_767; \
65408  float16x8_t __ret_767; \
65409 float16_t __reint_767 = __s0_767; \
65410 float16x8_t __reint1_767 = __s1_767; \
65411 int16x8_t __reint2_767 = vsetq_lane_s16(*(int16_t *) &__reint_767, *(int16x8_t *) &__reint1_767, __p2_767); \
65412  __ret_767 = *(float16x8_t *) &__reint2_767; \
65413  __ret_767; \
65414 })
65415 #else
65416 #define vsetq_lane_f16(__p0_768, __p1_768, __p2_768) __extension__ ({ \
65417  float16_t __s0_768 = __p0_768; \
65418  float16x8_t __s1_768 = __p1_768; \
65419  float16x8_t __rev1_768; __rev1_768 = __builtin_shufflevector(__s1_768, __s1_768, 7, 6, 5, 4, 3, 2, 1, 0); \
65420  float16x8_t __ret_768; \
65421 float16_t __reint_768 = __s0_768; \
65422 float16x8_t __reint1_768 = __rev1_768; \
65423 int16x8_t __reint2_768 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_768, *(int16x8_t *) &__reint1_768, __p2_768); \
65424  __ret_768 = *(float16x8_t *) &__reint2_768; \
65425  __ret_768 = __builtin_shufflevector(__ret_768, __ret_768, 7, 6, 5, 4, 3, 2, 1, 0); \
65426  __ret_768; \
65427 })
65428 #endif
65429 
65430 #if defined(__ARM_FEATURE_BF16_VECTOR_ARITHMETIC)
65431 #ifdef __LITTLE_ENDIAN__
65432 #define vbfmlalbq_lane_f32(__p0_769, __p1_769, __p2_769, __p3_769) __extension__ ({ \
65433  float32x4_t __s0_769 = __p0_769; \
65434  bfloat16x8_t __s1_769 = __p1_769; \
65435  bfloat16x4_t __s2_769 = __p2_769; \
65436  float32x4_t __ret_769; \
65437  __ret_769 = vbfmlalbq_f32(__s0_769, __s1_769, (bfloat16x8_t) {vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769), vget_lane_bf16(__s2_769, __p3_769)}); \
65438  __ret_769; \
65439 })
65440 #else
65441 #define vbfmlalbq_lane_f32(__p0_770, __p1_770, __p2_770, __p3_770) __extension__ ({ \
65442  float32x4_t __s0_770 = __p0_770; \
65443  bfloat16x8_t __s1_770 = __p1_770; \
65444  bfloat16x4_t __s2_770 = __p2_770; \
65445  float32x4_t __rev0_770; __rev0_770 = __builtin_shufflevector(__s0_770, __s0_770, 3, 2, 1, 0); \
65446  bfloat16x8_t __rev1_770; __rev1_770 = __builtin_shufflevector(__s1_770, __s1_770, 7, 6, 5, 4, 3, 2, 1, 0); \
65447  bfloat16x4_t __rev2_770; __rev2_770 = __builtin_shufflevector(__s2_770, __s2_770, 3, 2, 1, 0); \
65448  float32x4_t __ret_770; \
65449  __ret_770 = __noswap_vbfmlalbq_f32(__rev0_770, __rev1_770, (bfloat16x8_t) {__noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770), __noswap_vget_lane_bf16(__rev2_770, __p3_770)}); \
65450  __ret_770 = __builtin_shufflevector(__ret_770, __ret_770, 3, 2, 1, 0); \
65451  __ret_770; \
65452 })
65453 #endif
65454 
65455 #ifdef __LITTLE_ENDIAN__
65456 #define vbfmlalbq_laneq_f32(__p0_771, __p1_771, __p2_771, __p3_771) __extension__ ({ \
65457  float32x4_t __s0_771 = __p0_771; \
65458  bfloat16x8_t __s1_771 = __p1_771; \
65459  bfloat16x8_t __s2_771 = __p2_771; \
65460  float32x4_t __ret_771; \
65461  __ret_771 = vbfmlalbq_f32(__s0_771, __s1_771, (bfloat16x8_t) {vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771), vgetq_lane_bf16(__s2_771, __p3_771)}); \
65462  __ret_771; \
65463 })
65464 #else
65465 #define vbfmlalbq_laneq_f32(__p0_772, __p1_772, __p2_772, __p3_772) __extension__ ({ \
65466  float32x4_t __s0_772 = __p0_772; \
65467  bfloat16x8_t __s1_772 = __p1_772; \
65468  bfloat16x8_t __s2_772 = __p2_772; \
65469  float32x4_t __rev0_772; __rev0_772 = __builtin_shufflevector(__s0_772, __s0_772, 3, 2, 1, 0); \
65470  bfloat16x8_t __rev1_772; __rev1_772 = __builtin_shufflevector(__s1_772, __s1_772, 7, 6, 5, 4, 3, 2, 1, 0); \
65471  bfloat16x8_t __rev2_772; __rev2_772 = __builtin_shufflevector(__s2_772, __s2_772, 7, 6, 5, 4, 3, 2, 1, 0); \
65472  float32x4_t __ret_772; \
65473  __ret_772 = __noswap_vbfmlalbq_f32(__rev0_772, __rev1_772, (bfloat16x8_t) {__noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772), __noswap_vgetq_lane_bf16(__rev2_772, __p3_772)}); \
65474  __ret_772 = __builtin_shufflevector(__ret_772, __ret_772, 3, 2, 1, 0); \
65475  __ret_772; \
65476 })
65477 #endif
65478 
65479 #ifdef __LITTLE_ENDIAN__
65480 #define vbfmlaltq_lane_f32(__p0_773, __p1_773, __p2_773, __p3_773) __extension__ ({ \
65481  float32x4_t __s0_773 = __p0_773; \
65482  bfloat16x8_t __s1_773 = __p1_773; \
65483  bfloat16x4_t __s2_773 = __p2_773; \
65484  float32x4_t __ret_773; \
65485  __ret_773 = vbfmlaltq_f32(__s0_773, __s1_773, (bfloat16x8_t) {vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773), vget_lane_bf16(__s2_773, __p3_773)}); \
65486  __ret_773; \
65487 })
65488 #else
65489 #define vbfmlaltq_lane_f32(__p0_774, __p1_774, __p2_774, __p3_774) __extension__ ({ \
65490  float32x4_t __s0_774 = __p0_774; \
65491  bfloat16x8_t __s1_774 = __p1_774; \
65492  bfloat16x4_t __s2_774 = __p2_774; \
65493  float32x4_t __rev0_774; __rev0_774 = __builtin_shufflevector(__s0_774, __s0_774, 3, 2, 1, 0); \
65494  bfloat16x8_t __rev1_774; __rev1_774 = __builtin_shufflevector(__s1_774, __s1_774, 7, 6, 5, 4, 3, 2, 1, 0); \
65495  bfloat16x4_t __rev2_774; __rev2_774 = __builtin_shufflevector(__s2_774, __s2_774, 3, 2, 1, 0); \
65496  float32x4_t __ret_774; \
65497  __ret_774 = __noswap_vbfmlaltq_f32(__rev0_774, __rev1_774, (bfloat16x8_t) {__noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774), __noswap_vget_lane_bf16(__rev2_774, __p3_774)}); \
65498  __ret_774 = __builtin_shufflevector(__ret_774, __ret_774, 3, 2, 1, 0); \
65499  __ret_774; \
65500 })
65501 #endif
65502 
65503 #ifdef __LITTLE_ENDIAN__
65504 #define vbfmlaltq_laneq_f32(__p0_775, __p1_775, __p2_775, __p3_775) __extension__ ({ \
65505  float32x4_t __s0_775 = __p0_775; \
65506  bfloat16x8_t __s1_775 = __p1_775; \
65507  bfloat16x8_t __s2_775 = __p2_775; \
65508  float32x4_t __ret_775; \
65509  __ret_775 = vbfmlaltq_f32(__s0_775, __s1_775, (bfloat16x8_t) {vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775), vgetq_lane_bf16(__s2_775, __p3_775)}); \
65510  __ret_775; \
65511 })
65512 #else
65513 #define vbfmlaltq_laneq_f32(__p0_776, __p1_776, __p2_776, __p3_776) __extension__ ({ \
65514  float32x4_t __s0_776 = __p0_776; \
65515  bfloat16x8_t __s1_776 = __p1_776; \
65516  bfloat16x8_t __s2_776 = __p2_776; \
65517  float32x4_t __rev0_776; __rev0_776 = __builtin_shufflevector(__s0_776, __s0_776, 3, 2, 1, 0); \
65518  bfloat16x8_t __rev1_776; __rev1_776 = __builtin_shufflevector(__s1_776, __s1_776, 7, 6, 5, 4, 3, 2, 1, 0); \
65519  bfloat16x8_t __rev2_776; __rev2_776 = __builtin_shufflevector(__s2_776, __s2_776, 7, 6, 5, 4, 3, 2, 1, 0); \
65520  float32x4_t __ret_776; \
65521  __ret_776 = __noswap_vbfmlaltq_f32(__rev0_776, __rev1_776, (bfloat16x8_t) {__noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776), __noswap_vgetq_lane_bf16(__rev2_776, __p3_776)}); \
65522  __ret_776 = __builtin_shufflevector(__ret_776, __ret_776, 3, 2, 1, 0); \
65523  __ret_776; \
65524 })
65525 #endif
65526 
65527 #ifdef __LITTLE_ENDIAN__
65528 __ai float32x4_t vcvtq_high_f32_bf16(bfloat16x8_t __p0) {
65529  float32x4_t __ret;
65530  __ret = vcvt_f32_bf16(vget_high_bf16(__p0));
65531  return __ret;
65532 }
65533 #else
65534 __ai float32x4_t vcvtq_high_f32_bf16(bfloat16x8_t __p0) {
65535  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65536  float32x4_t __ret;
65537  __ret = __noswap_vcvt_f32_bf16(__noswap_vget_high_bf16(__rev0));
65538  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65539  return __ret;
65540 }
65541 #endif
65542 
65543 #ifdef __LITTLE_ENDIAN__
65544 __ai float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) {
65545  float32x4_t __ret;
65546  __ret = vcvt_f32_bf16(vget_low_bf16(__p0));
65547  return __ret;
65548 }
65549 #else
65550 __ai float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) {
65551  bfloat16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65552  float32x4_t __ret;
65553  __ret = __noswap_vcvt_f32_bf16(__noswap_vget_low_bf16(__rev0));
65554  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65555  return __ret;
65556 }
65557 #endif
65558 
65559 #endif
65560 #if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__)
65561 #ifdef __LITTLE_ENDIAN__
65562 #define vfmlalq_lane_high_f16(__p0_777, __p1_777, __p2_777, __p3_777) __extension__ ({ \
65563  float32x4_t __s0_777 = __p0_777; \
65564  float16x8_t __s1_777 = __p1_777; \
65565  float16x4_t __s2_777 = __p2_777; \
65566  float32x4_t __ret_777; \
65567  __ret_777 = vfmlalq_high_f16(__s0_777, __s1_777, (float16x8_t) {vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777), vget_lane_f16(__s2_777, __p3_777)}); \
65568  __ret_777; \
65569 })
65570 #else
65571 #define vfmlalq_lane_high_f16(__p0_778, __p1_778, __p2_778, __p3_778) __extension__ ({ \
65572  float32x4_t __s0_778 = __p0_778; \
65573  float16x8_t __s1_778 = __p1_778; \
65574  float16x4_t __s2_778 = __p2_778; \
65575  float32x4_t __rev0_778; __rev0_778 = __builtin_shufflevector(__s0_778, __s0_778, 3, 2, 1, 0); \
65576  float16x8_t __rev1_778; __rev1_778 = __builtin_shufflevector(__s1_778, __s1_778, 7, 6, 5, 4, 3, 2, 1, 0); \
65577  float16x4_t __rev2_778; __rev2_778 = __builtin_shufflevector(__s2_778, __s2_778, 3, 2, 1, 0); \
65578  float32x4_t __ret_778; \
65579  __ret_778 = __noswap_vfmlalq_high_f16(__rev0_778, __rev1_778, (float16x8_t) {__noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778), __noswap_vget_lane_f16(__rev2_778, __p3_778)}); \
65580  __ret_778 = __builtin_shufflevector(__ret_778, __ret_778, 3, 2, 1, 0); \
65581  __ret_778; \
65582 })
65583 #endif
65584 
65585 #ifdef __LITTLE_ENDIAN__
65586 #define vfmlal_lane_high_f16(__p0_779, __p1_779, __p2_779, __p3_779) __extension__ ({ \
65587  float32x2_t __s0_779 = __p0_779; \
65588  float16x4_t __s1_779 = __p1_779; \
65589  float16x4_t __s2_779 = __p2_779; \
65590  float32x2_t __ret_779; \
65591  __ret_779 = vfmlal_high_f16(__s0_779, __s1_779, (float16x4_t) {vget_lane_f16(__s2_779, __p3_779), vget_lane_f16(__s2_779, __p3_779), vget_lane_f16(__s2_779, __p3_779), vget_lane_f16(__s2_779, __p3_779)}); \
65592  __ret_779; \
65593 })
65594 #else
65595 #define vfmlal_lane_high_f16(__p0_780, __p1_780, __p2_780, __p3_780) __extension__ ({ \
65596  float32x2_t __s0_780 = __p0_780; \
65597  float16x4_t __s1_780 = __p1_780; \
65598  float16x4_t __s2_780 = __p2_780; \
65599  float32x2_t __rev0_780; __rev0_780 = __builtin_shufflevector(__s0_780, __s0_780, 1, 0); \
65600  float16x4_t __rev1_780; __rev1_780 = __builtin_shufflevector(__s1_780, __s1_780, 3, 2, 1, 0); \
65601  float16x4_t __rev2_780; __rev2_780 = __builtin_shufflevector(__s2_780, __s2_780, 3, 2, 1, 0); \
65602  float32x2_t __ret_780; \
65603  __ret_780 = __noswap_vfmlal_high_f16(__rev0_780, __rev1_780, (float16x4_t) {__noswap_vget_lane_f16(__rev2_780, __p3_780), __noswap_vget_lane_f16(__rev2_780, __p3_780), __noswap_vget_lane_f16(__rev2_780, __p3_780), __noswap_vget_lane_f16(__rev2_780, __p3_780)}); \
65604  __ret_780 = __builtin_shufflevector(__ret_780, __ret_780, 1, 0); \
65605  __ret_780; \
65606 })
65607 #endif
65608 
65609 #ifdef __LITTLE_ENDIAN__
65610 #define vfmlalq_lane_low_f16(__p0_781, __p1_781, __p2_781, __p3_781) __extension__ ({ \
65611  float32x4_t __s0_781 = __p0_781; \
65612  float16x8_t __s1_781 = __p1_781; \
65613  float16x4_t __s2_781 = __p2_781; \
65614  float32x4_t __ret_781; \
65615  __ret_781 = vfmlalq_low_f16(__s0_781, __s1_781, (float16x8_t) {vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781), vget_lane_f16(__s2_781, __p3_781)}); \
65616  __ret_781; \
65617 })
65618 #else
65619 #define vfmlalq_lane_low_f16(__p0_782, __p1_782, __p2_782, __p3_782) __extension__ ({ \
65620  float32x4_t __s0_782 = __p0_782; \
65621  float16x8_t __s1_782 = __p1_782; \
65622  float16x4_t __s2_782 = __p2_782; \
65623  float32x4_t __rev0_782; __rev0_782 = __builtin_shufflevector(__s0_782, __s0_782, 3, 2, 1, 0); \
65624  float16x8_t __rev1_782; __rev1_782 = __builtin_shufflevector(__s1_782, __s1_782, 7, 6, 5, 4, 3, 2, 1, 0); \
65625  float16x4_t __rev2_782; __rev2_782 = __builtin_shufflevector(__s2_782, __s2_782, 3, 2, 1, 0); \
65626  float32x4_t __ret_782; \
65627  __ret_782 = __noswap_vfmlalq_low_f16(__rev0_782, __rev1_782, (float16x8_t) {__noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782), __noswap_vget_lane_f16(__rev2_782, __p3_782)}); \
65628  __ret_782 = __builtin_shufflevector(__ret_782, __ret_782, 3, 2, 1, 0); \
65629  __ret_782; \
65630 })
65631 #endif
65632 
65633 #ifdef __LITTLE_ENDIAN__
65634 #define vfmlal_lane_low_f16(__p0_783, __p1_783, __p2_783, __p3_783) __extension__ ({ \
65635  float32x2_t __s0_783 = __p0_783; \
65636  float16x4_t __s1_783 = __p1_783; \
65637  float16x4_t __s2_783 = __p2_783; \
65638  float32x2_t __ret_783; \
65639  __ret_783 = vfmlal_low_f16(__s0_783, __s1_783, (float16x4_t) {vget_lane_f16(__s2_783, __p3_783), vget_lane_f16(__s2_783, __p3_783), vget_lane_f16(__s2_783, __p3_783), vget_lane_f16(__s2_783, __p3_783)}); \
65640  __ret_783; \
65641 })
65642 #else
65643 #define vfmlal_lane_low_f16(__p0_784, __p1_784, __p2_784, __p3_784) __extension__ ({ \
65644  float32x2_t __s0_784 = __p0_784; \
65645  float16x4_t __s1_784 = __p1_784; \
65646  float16x4_t __s2_784 = __p2_784; \
65647  float32x2_t __rev0_784; __rev0_784 = __builtin_shufflevector(__s0_784, __s0_784, 1, 0); \
65648  float16x4_t __rev1_784; __rev1_784 = __builtin_shufflevector(__s1_784, __s1_784, 3, 2, 1, 0); \
65649  float16x4_t __rev2_784; __rev2_784 = __builtin_shufflevector(__s2_784, __s2_784, 3, 2, 1, 0); \
65650  float32x2_t __ret_784; \
65651  __ret_784 = __noswap_vfmlal_low_f16(__rev0_784, __rev1_784, (float16x4_t) {__noswap_vget_lane_f16(__rev2_784, __p3_784), __noswap_vget_lane_f16(__rev2_784, __p3_784), __noswap_vget_lane_f16(__rev2_784, __p3_784), __noswap_vget_lane_f16(__rev2_784, __p3_784)}); \
65652  __ret_784 = __builtin_shufflevector(__ret_784, __ret_784, 1, 0); \
65653  __ret_784; \
65654 })
65655 #endif
65656 
65657 #ifdef __LITTLE_ENDIAN__
65658 #define vfmlalq_laneq_high_f16(__p0_785, __p1_785, __p2_785, __p3_785) __extension__ ({ \
65659  float32x4_t __s0_785 = __p0_785; \
65660  float16x8_t __s1_785 = __p1_785; \
65661  float16x8_t __s2_785 = __p2_785; \
65662  float32x4_t __ret_785; \
65663  __ret_785 = vfmlalq_high_f16(__s0_785, __s1_785, (float16x8_t) {vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785), vgetq_lane_f16(__s2_785, __p3_785)}); \
65664  __ret_785; \
65665 })
65666 #else
65667 #define vfmlalq_laneq_high_f16(__p0_786, __p1_786, __p2_786, __p3_786) __extension__ ({ \
65668  float32x4_t __s0_786 = __p0_786; \
65669  float16x8_t __s1_786 = __p1_786; \
65670  float16x8_t __s2_786 = __p2_786; \
65671  float32x4_t __rev0_786; __rev0_786 = __builtin_shufflevector(__s0_786, __s0_786, 3, 2, 1, 0); \
65672  float16x8_t __rev1_786; __rev1_786 = __builtin_shufflevector(__s1_786, __s1_786, 7, 6, 5, 4, 3, 2, 1, 0); \
65673  float16x8_t __rev2_786; __rev2_786 = __builtin_shufflevector(__s2_786, __s2_786, 7, 6, 5, 4, 3, 2, 1, 0); \
65674  float32x4_t __ret_786; \
65675  __ret_786 = __noswap_vfmlalq_high_f16(__rev0_786, __rev1_786, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786), __noswap_vgetq_lane_f16(__rev2_786, __p3_786)}); \
65676  __ret_786 = __builtin_shufflevector(__ret_786, __ret_786, 3, 2, 1, 0); \
65677  __ret_786; \
65678 })
65679 #endif
65680 
65681 #ifdef __LITTLE_ENDIAN__
65682 #define vfmlal_laneq_high_f16(__p0_787, __p1_787, __p2_787, __p3_787) __extension__ ({ \
65683  float32x2_t __s0_787 = __p0_787; \
65684  float16x4_t __s1_787 = __p1_787; \
65685  float16x8_t __s2_787 = __p2_787; \
65686  float32x2_t __ret_787; \
65687  __ret_787 = vfmlal_high_f16(__s0_787, __s1_787, (float16x4_t) {vgetq_lane_f16(__s2_787, __p3_787), vgetq_lane_f16(__s2_787, __p3_787), vgetq_lane_f16(__s2_787, __p3_787), vgetq_lane_f16(__s2_787, __p3_787)}); \
65688  __ret_787; \
65689 })
65690 #else
65691 #define vfmlal_laneq_high_f16(__p0_788, __p1_788, __p2_788, __p3_788) __extension__ ({ \
65692  float32x2_t __s0_788 = __p0_788; \
65693  float16x4_t __s1_788 = __p1_788; \
65694  float16x8_t __s2_788 = __p2_788; \
65695  float32x2_t __rev0_788; __rev0_788 = __builtin_shufflevector(__s0_788, __s0_788, 1, 0); \
65696  float16x4_t __rev1_788; __rev1_788 = __builtin_shufflevector(__s1_788, __s1_788, 3, 2, 1, 0); \
65697  float16x8_t __rev2_788; __rev2_788 = __builtin_shufflevector(__s2_788, __s2_788, 7, 6, 5, 4, 3, 2, 1, 0); \
65698  float32x2_t __ret_788; \
65699  __ret_788 = __noswap_vfmlal_high_f16(__rev0_788, __rev1_788, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_788, __p3_788), __noswap_vgetq_lane_f16(__rev2_788, __p3_788), __noswap_vgetq_lane_f16(__rev2_788, __p3_788), __noswap_vgetq_lane_f16(__rev2_788, __p3_788)}); \
65700  __ret_788 = __builtin_shufflevector(__ret_788, __ret_788, 1, 0); \
65701  __ret_788; \
65702 })
65703 #endif
65704 
65705 #ifdef __LITTLE_ENDIAN__
65706 #define vfmlalq_laneq_low_f16(__p0_789, __p1_789, __p2_789, __p3_789) __extension__ ({ \
65707  float32x4_t __s0_789 = __p0_789; \
65708  float16x8_t __s1_789 = __p1_789; \
65709  float16x8_t __s2_789 = __p2_789; \
65710  float32x4_t __ret_789; \
65711  __ret_789 = vfmlalq_low_f16(__s0_789, __s1_789, (float16x8_t) {vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789), vgetq_lane_f16(__s2_789, __p3_789)}); \
65712  __ret_789; \
65713 })
65714 #else
65715 #define vfmlalq_laneq_low_f16(__p0_790, __p1_790, __p2_790, __p3_790) __extension__ ({ \
65716  float32x4_t __s0_790 = __p0_790; \
65717  float16x8_t __s1_790 = __p1_790; \
65718  float16x8_t __s2_790 = __p2_790; \
65719  float32x4_t __rev0_790; __rev0_790 = __builtin_shufflevector(__s0_790, __s0_790, 3, 2, 1, 0); \
65720  float16x8_t __rev1_790; __rev1_790 = __builtin_shufflevector(__s1_790, __s1_790, 7, 6, 5, 4, 3, 2, 1, 0); \
65721  float16x8_t __rev2_790; __rev2_790 = __builtin_shufflevector(__s2_790, __s2_790, 7, 6, 5, 4, 3, 2, 1, 0); \
65722  float32x4_t __ret_790; \
65723  __ret_790 = __noswap_vfmlalq_low_f16(__rev0_790, __rev1_790, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790), __noswap_vgetq_lane_f16(__rev2_790, __p3_790)}); \
65724  __ret_790 = __builtin_shufflevector(__ret_790, __ret_790, 3, 2, 1, 0); \
65725  __ret_790; \
65726 })
65727 #endif
65728 
65729 #ifdef __LITTLE_ENDIAN__
65730 #define vfmlal_laneq_low_f16(__p0_791, __p1_791, __p2_791, __p3_791) __extension__ ({ \
65731  float32x2_t __s0_791 = __p0_791; \
65732  float16x4_t __s1_791 = __p1_791; \
65733  float16x8_t __s2_791 = __p2_791; \
65734  float32x2_t __ret_791; \
65735  __ret_791 = vfmlal_low_f16(__s0_791, __s1_791, (float16x4_t) {vgetq_lane_f16(__s2_791, __p3_791), vgetq_lane_f16(__s2_791, __p3_791), vgetq_lane_f16(__s2_791, __p3_791), vgetq_lane_f16(__s2_791, __p3_791)}); \
65736  __ret_791; \
65737 })
65738 #else
65739 #define vfmlal_laneq_low_f16(__p0_792, __p1_792, __p2_792, __p3_792) __extension__ ({ \
65740  float32x2_t __s0_792 = __p0_792; \
65741  float16x4_t __s1_792 = __p1_792; \
65742  float16x8_t __s2_792 = __p2_792; \
65743  float32x2_t __rev0_792; __rev0_792 = __builtin_shufflevector(__s0_792, __s0_792, 1, 0); \
65744  float16x4_t __rev1_792; __rev1_792 = __builtin_shufflevector(__s1_792, __s1_792, 3, 2, 1, 0); \
65745  float16x8_t __rev2_792; __rev2_792 = __builtin_shufflevector(__s2_792, __s2_792, 7, 6, 5, 4, 3, 2, 1, 0); \
65746  float32x2_t __ret_792; \
65747  __ret_792 = __noswap_vfmlal_low_f16(__rev0_792, __rev1_792, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_792, __p3_792), __noswap_vgetq_lane_f16(__rev2_792, __p3_792), __noswap_vgetq_lane_f16(__rev2_792, __p3_792), __noswap_vgetq_lane_f16(__rev2_792, __p3_792)}); \
65748  __ret_792 = __builtin_shufflevector(__ret_792, __ret_792, 1, 0); \
65749  __ret_792; \
65750 })
65751 #endif
65752 
65753 #ifdef __LITTLE_ENDIAN__
65754 #define vfmlslq_lane_high_f16(__p0_793, __p1_793, __p2_793, __p3_793) __extension__ ({ \
65755  float32x4_t __s0_793 = __p0_793; \
65756  float16x8_t __s1_793 = __p1_793; \
65757  float16x4_t __s2_793 = __p2_793; \
65758  float32x4_t __ret_793; \
65759  __ret_793 = vfmlslq_high_f16(__s0_793, __s1_793, (float16x8_t) {vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793), vget_lane_f16(__s2_793, __p3_793)}); \
65760  __ret_793; \
65761 })
65762 #else
65763 #define vfmlslq_lane_high_f16(__p0_794, __p1_794, __p2_794, __p3_794) __extension__ ({ \
65764  float32x4_t __s0_794 = __p0_794; \
65765  float16x8_t __s1_794 = __p1_794; \
65766  float16x4_t __s2_794 = __p2_794; \
65767  float32x4_t __rev0_794; __rev0_794 = __builtin_shufflevector(__s0_794, __s0_794, 3, 2, 1, 0); \
65768  float16x8_t __rev1_794; __rev1_794 = __builtin_shufflevector(__s1_794, __s1_794, 7, 6, 5, 4, 3, 2, 1, 0); \
65769  float16x4_t __rev2_794; __rev2_794 = __builtin_shufflevector(__s2_794, __s2_794, 3, 2, 1, 0); \
65770  float32x4_t __ret_794; \
65771  __ret_794 = __noswap_vfmlslq_high_f16(__rev0_794, __rev1_794, (float16x8_t) {__noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794), __noswap_vget_lane_f16(__rev2_794, __p3_794)}); \
65772  __ret_794 = __builtin_shufflevector(__ret_794, __ret_794, 3, 2, 1, 0); \
65773  __ret_794; \
65774 })
65775 #endif
65776 
65777 #ifdef __LITTLE_ENDIAN__
65778 #define vfmlsl_lane_high_f16(__p0_795, __p1_795, __p2_795, __p3_795) __extension__ ({ \
65779  float32x2_t __s0_795 = __p0_795; \
65780  float16x4_t __s1_795 = __p1_795; \
65781  float16x4_t __s2_795 = __p2_795; \
65782  float32x2_t __ret_795; \
65783  __ret_795 = vfmlsl_high_f16(__s0_795, __s1_795, (float16x4_t) {vget_lane_f16(__s2_795, __p3_795), vget_lane_f16(__s2_795, __p3_795), vget_lane_f16(__s2_795, __p3_795), vget_lane_f16(__s2_795, __p3_795)}); \
65784  __ret_795; \
65785 })
65786 #else
65787 #define vfmlsl_lane_high_f16(__p0_796, __p1_796, __p2_796, __p3_796) __extension__ ({ \
65788  float32x2_t __s0_796 = __p0_796; \
65789  float16x4_t __s1_796 = __p1_796; \
65790  float16x4_t __s2_796 = __p2_796; \
65791  float32x2_t __rev0_796; __rev0_796 = __builtin_shufflevector(__s0_796, __s0_796, 1, 0); \
65792  float16x4_t __rev1_796; __rev1_796 = __builtin_shufflevector(__s1_796, __s1_796, 3, 2, 1, 0); \
65793  float16x4_t __rev2_796; __rev2_796 = __builtin_shufflevector(__s2_796, __s2_796, 3, 2, 1, 0); \
65794  float32x2_t __ret_796; \
65795  __ret_796 = __noswap_vfmlsl_high_f16(__rev0_796, __rev1_796, (float16x4_t) {__noswap_vget_lane_f16(__rev2_796, __p3_796), __noswap_vget_lane_f16(__rev2_796, __p3_796), __noswap_vget_lane_f16(__rev2_796, __p3_796), __noswap_vget_lane_f16(__rev2_796, __p3_796)}); \
65796  __ret_796 = __builtin_shufflevector(__ret_796, __ret_796, 1, 0); \
65797  __ret_796; \
65798 })
65799 #endif
65800 
65801 #ifdef __LITTLE_ENDIAN__
65802 #define vfmlslq_lane_low_f16(__p0_797, __p1_797, __p2_797, __p3_797) __extension__ ({ \
65803  float32x4_t __s0_797 = __p0_797; \
65804  float16x8_t __s1_797 = __p1_797; \
65805  float16x4_t __s2_797 = __p2_797; \
65806  float32x4_t __ret_797; \
65807  __ret_797 = vfmlslq_low_f16(__s0_797, __s1_797, (float16x8_t) {vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797), vget_lane_f16(__s2_797, __p3_797)}); \
65808  __ret_797; \
65809 })
65810 #else
65811 #define vfmlslq_lane_low_f16(__p0_798, __p1_798, __p2_798, __p3_798) __extension__ ({ \
65812  float32x4_t __s0_798 = __p0_798; \
65813  float16x8_t __s1_798 = __p1_798; \
65814  float16x4_t __s2_798 = __p2_798; \
65815  float32x4_t __rev0_798; __rev0_798 = __builtin_shufflevector(__s0_798, __s0_798, 3, 2, 1, 0); \
65816  float16x8_t __rev1_798; __rev1_798 = __builtin_shufflevector(__s1_798, __s1_798, 7, 6, 5, 4, 3, 2, 1, 0); \
65817  float16x4_t __rev2_798; __rev2_798 = __builtin_shufflevector(__s2_798, __s2_798, 3, 2, 1, 0); \
65818  float32x4_t __ret_798; \
65819  __ret_798 = __noswap_vfmlslq_low_f16(__rev0_798, __rev1_798, (float16x8_t) {__noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798), __noswap_vget_lane_f16(__rev2_798, __p3_798)}); \
65820  __ret_798 = __builtin_shufflevector(__ret_798, __ret_798, 3, 2, 1, 0); \
65821  __ret_798; \
65822 })
65823 #endif
65824 
65825 #ifdef __LITTLE_ENDIAN__
65826 #define vfmlsl_lane_low_f16(__p0_799, __p1_799, __p2_799, __p3_799) __extension__ ({ \
65827  float32x2_t __s0_799 = __p0_799; \
65828  float16x4_t __s1_799 = __p1_799; \
65829  float16x4_t __s2_799 = __p2_799; \
65830  float32x2_t __ret_799; \
65831  __ret_799 = vfmlsl_low_f16(__s0_799, __s1_799, (float16x4_t) {vget_lane_f16(__s2_799, __p3_799), vget_lane_f16(__s2_799, __p3_799), vget_lane_f16(__s2_799, __p3_799), vget_lane_f16(__s2_799, __p3_799)}); \
65832  __ret_799; \
65833 })
65834 #else
65835 #define vfmlsl_lane_low_f16(__p0_800, __p1_800, __p2_800, __p3_800) __extension__ ({ \
65836  float32x2_t __s0_800 = __p0_800; \
65837  float16x4_t __s1_800 = __p1_800; \
65838  float16x4_t __s2_800 = __p2_800; \
65839  float32x2_t __rev0_800; __rev0_800 = __builtin_shufflevector(__s0_800, __s0_800, 1, 0); \
65840  float16x4_t __rev1_800; __rev1_800 = __builtin_shufflevector(__s1_800, __s1_800, 3, 2, 1, 0); \
65841  float16x4_t __rev2_800; __rev2_800 = __builtin_shufflevector(__s2_800, __s2_800, 3, 2, 1, 0); \
65842  float32x2_t __ret_800; \
65843  __ret_800 = __noswap_vfmlsl_low_f16(__rev0_800, __rev1_800, (float16x4_t) {__noswap_vget_lane_f16(__rev2_800, __p3_800), __noswap_vget_lane_f16(__rev2_800, __p3_800), __noswap_vget_lane_f16(__rev2_800, __p3_800), __noswap_vget_lane_f16(__rev2_800, __p3_800)}); \
65844  __ret_800 = __builtin_shufflevector(__ret_800, __ret_800, 1, 0); \
65845  __ret_800; \
65846 })
65847 #endif
65848 
65849 #ifdef __LITTLE_ENDIAN__
65850 #define vfmlslq_laneq_high_f16(__p0_801, __p1_801, __p2_801, __p3_801) __extension__ ({ \
65851  float32x4_t __s0_801 = __p0_801; \
65852  float16x8_t __s1_801 = __p1_801; \
65853  float16x8_t __s2_801 = __p2_801; \
65854  float32x4_t __ret_801; \
65855  __ret_801 = vfmlslq_high_f16(__s0_801, __s1_801, (float16x8_t) {vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801), vgetq_lane_f16(__s2_801, __p3_801)}); \
65856  __ret_801; \
65857 })
65858 #else
65859 #define vfmlslq_laneq_high_f16(__p0_802, __p1_802, __p2_802, __p3_802) __extension__ ({ \
65860  float32x4_t __s0_802 = __p0_802; \
65861  float16x8_t __s1_802 = __p1_802; \
65862  float16x8_t __s2_802 = __p2_802; \
65863  float32x4_t __rev0_802; __rev0_802 = __builtin_shufflevector(__s0_802, __s0_802, 3, 2, 1, 0); \
65864  float16x8_t __rev1_802; __rev1_802 = __builtin_shufflevector(__s1_802, __s1_802, 7, 6, 5, 4, 3, 2, 1, 0); \
65865  float16x8_t __rev2_802; __rev2_802 = __builtin_shufflevector(__s2_802, __s2_802, 7, 6, 5, 4, 3, 2, 1, 0); \
65866  float32x4_t __ret_802; \
65867  __ret_802 = __noswap_vfmlslq_high_f16(__rev0_802, __rev1_802, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802), __noswap_vgetq_lane_f16(__rev2_802, __p3_802)}); \
65868  __ret_802 = __builtin_shufflevector(__ret_802, __ret_802, 3, 2, 1, 0); \
65869  __ret_802; \
65870 })
65871 #endif
65872 
65873 #ifdef __LITTLE_ENDIAN__
65874 #define vfmlsl_laneq_high_f16(__p0_803, __p1_803, __p2_803, __p3_803) __extension__ ({ \
65875  float32x2_t __s0_803 = __p0_803; \
65876  float16x4_t __s1_803 = __p1_803; \
65877  float16x8_t __s2_803 = __p2_803; \
65878  float32x2_t __ret_803; \
65879  __ret_803 = vfmlsl_high_f16(__s0_803, __s1_803, (float16x4_t) {vgetq_lane_f16(__s2_803, __p3_803), vgetq_lane_f16(__s2_803, __p3_803), vgetq_lane_f16(__s2_803, __p3_803), vgetq_lane_f16(__s2_803, __p3_803)}); \
65880  __ret_803; \
65881 })
65882 #else
65883 #define vfmlsl_laneq_high_f16(__p0_804, __p1_804, __p2_804, __p3_804) __extension__ ({ \
65884  float32x2_t __s0_804 = __p0_804; \
65885  float16x4_t __s1_804 = __p1_804; \
65886  float16x8_t __s2_804 = __p2_804; \
65887  float32x2_t __rev0_804; __rev0_804 = __builtin_shufflevector(__s0_804, __s0_804, 1, 0); \
65888  float16x4_t __rev1_804; __rev1_804 = __builtin_shufflevector(__s1_804, __s1_804, 3, 2, 1, 0); \
65889  float16x8_t __rev2_804; __rev2_804 = __builtin_shufflevector(__s2_804, __s2_804, 7, 6, 5, 4, 3, 2, 1, 0); \
65890  float32x2_t __ret_804; \
65891  __ret_804 = __noswap_vfmlsl_high_f16(__rev0_804, __rev1_804, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_804, __p3_804), __noswap_vgetq_lane_f16(__rev2_804, __p3_804), __noswap_vgetq_lane_f16(__rev2_804, __p3_804), __noswap_vgetq_lane_f16(__rev2_804, __p3_804)}); \
65892  __ret_804 = __builtin_shufflevector(__ret_804, __ret_804, 1, 0); \
65893  __ret_804; \
65894 })
65895 #endif
65896 
65897 #ifdef __LITTLE_ENDIAN__
65898 #define vfmlslq_laneq_low_f16(__p0_805, __p1_805, __p2_805, __p3_805) __extension__ ({ \
65899  float32x4_t __s0_805 = __p0_805; \
65900  float16x8_t __s1_805 = __p1_805; \
65901  float16x8_t __s2_805 = __p2_805; \
65902  float32x4_t __ret_805; \
65903  __ret_805 = vfmlslq_low_f16(__s0_805, __s1_805, (float16x8_t) {vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805), vgetq_lane_f16(__s2_805, __p3_805)}); \
65904  __ret_805; \
65905 })
65906 #else
65907 #define vfmlslq_laneq_low_f16(__p0_806, __p1_806, __p2_806, __p3_806) __extension__ ({ \
65908  float32x4_t __s0_806 = __p0_806; \
65909  float16x8_t __s1_806 = __p1_806; \
65910  float16x8_t __s2_806 = __p2_806; \
65911  float32x4_t __rev0_806; __rev0_806 = __builtin_shufflevector(__s0_806, __s0_806, 3, 2, 1, 0); \
65912  float16x8_t __rev1_806; __rev1_806 = __builtin_shufflevector(__s1_806, __s1_806, 7, 6, 5, 4, 3, 2, 1, 0); \
65913  float16x8_t __rev2_806; __rev2_806 = __builtin_shufflevector(__s2_806, __s2_806, 7, 6, 5, 4, 3, 2, 1, 0); \
65914  float32x4_t __ret_806; \
65915  __ret_806 = __noswap_vfmlslq_low_f16(__rev0_806, __rev1_806, (float16x8_t) {__noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806), __noswap_vgetq_lane_f16(__rev2_806, __p3_806)}); \
65916  __ret_806 = __builtin_shufflevector(__ret_806, __ret_806, 3, 2, 1, 0); \
65917  __ret_806; \
65918 })
65919 #endif
65920 
65921 #ifdef __LITTLE_ENDIAN__
65922 #define vfmlsl_laneq_low_f16(__p0_807, __p1_807, __p2_807, __p3_807) __extension__ ({ \
65923  float32x2_t __s0_807 = __p0_807; \
65924  float16x4_t __s1_807 = __p1_807; \
65925  float16x8_t __s2_807 = __p2_807; \
65926  float32x2_t __ret_807; \
65927  __ret_807 = vfmlsl_low_f16(__s0_807, __s1_807, (float16x4_t) {vgetq_lane_f16(__s2_807, __p3_807), vgetq_lane_f16(__s2_807, __p3_807), vgetq_lane_f16(__s2_807, __p3_807), vgetq_lane_f16(__s2_807, __p3_807)}); \
65928  __ret_807; \
65929 })
65930 #else
65931 #define vfmlsl_laneq_low_f16(__p0_808, __p1_808, __p2_808, __p3_808) __extension__ ({ \
65932  float32x2_t __s0_808 = __p0_808; \
65933  float16x4_t __s1_808 = __p1_808; \
65934  float16x8_t __s2_808 = __p2_808; \
65935  float32x2_t __rev0_808; __rev0_808 = __builtin_shufflevector(__s0_808, __s0_808, 1, 0); \
65936  float16x4_t __rev1_808; __rev1_808 = __builtin_shufflevector(__s1_808, __s1_808, 3, 2, 1, 0); \
65937  float16x8_t __rev2_808; __rev2_808 = __builtin_shufflevector(__s2_808, __s2_808, 7, 6, 5, 4, 3, 2, 1, 0); \
65938  float32x2_t __ret_808; \
65939  __ret_808 = __noswap_vfmlsl_low_f16(__rev0_808, __rev1_808, (float16x4_t) {__noswap_vgetq_lane_f16(__rev2_808, __p3_808), __noswap_vgetq_lane_f16(__rev2_808, __p3_808), __noswap_vgetq_lane_f16(__rev2_808, __p3_808), __noswap_vgetq_lane_f16(__rev2_808, __p3_808)}); \
65940  __ret_808 = __builtin_shufflevector(__ret_808, __ret_808, 1, 0); \
65941  __ret_808; \
65942 })
65943 #endif
65944 
65945 #endif
65946 #if defined(__ARM_FEATURE_FP16_VECTOR_ARITHMETIC) && defined(__aarch64__)
65947 #ifdef __LITTLE_ENDIAN__
65948 #define vmulh_lane_f16(__p0_809, __p1_809, __p2_809) __extension__ ({ \
65949  float16_t __s0_809 = __p0_809; \
65950  float16x4_t __s1_809 = __p1_809; \
65951  float16_t __ret_809; \
65952  __ret_809 = __s0_809 * vget_lane_f16(__s1_809, __p2_809); \
65953  __ret_809; \
65954 })
65955 #else
65956 #define vmulh_lane_f16(__p0_810, __p1_810, __p2_810) __extension__ ({ \
65957  float16_t __s0_810 = __p0_810; \
65958  float16x4_t __s1_810 = __p1_810; \
65959  float16x4_t __rev1_810; __rev1_810 = __builtin_shufflevector(__s1_810, __s1_810, 3, 2, 1, 0); \
65960  float16_t __ret_810; \
65961  __ret_810 = __s0_810 * __noswap_vget_lane_f16(__rev1_810, __p2_810); \
65962  __ret_810; \
65963 })
65964 #endif
65965 
65966 #ifdef __LITTLE_ENDIAN__
65967 #define vmulh_laneq_f16(__p0_811, __p1_811, __p2_811) __extension__ ({ \
65968  float16_t __s0_811 = __p0_811; \
65969  float16x8_t __s1_811 = __p1_811; \
65970  float16_t __ret_811; \
65971  __ret_811 = __s0_811 * vgetq_lane_f16(__s1_811, __p2_811); \
65972  __ret_811; \
65973 })
65974 #else
65975 #define vmulh_laneq_f16(__p0_812, __p1_812, __p2_812) __extension__ ({ \
65976  float16_t __s0_812 = __p0_812; \
65977  float16x8_t __s1_812 = __p1_812; \
65978  float16x8_t __rev1_812; __rev1_812 = __builtin_shufflevector(__s1_812, __s1_812, 7, 6, 5, 4, 3, 2, 1, 0); \
65979  float16_t __ret_812; \
65980  __ret_812 = __s0_812 * __noswap_vgetq_lane_f16(__rev1_812, __p2_812); \
65981  __ret_812; \
65982 })
65983 #endif
65984 
65985 #endif
65986 #if defined(__ARM_FEATURE_MATMUL_INT8)
65987 #ifdef __LITTLE_ENDIAN__
65988 #define vsudotq_lane_s32(__p0_813, __p1_813, __p2_813, __p3_813) __extension__ ({ \
65989  int32x4_t __s0_813 = __p0_813; \
65990  int8x16_t __s1_813 = __p1_813; \
65991  uint8x8_t __s2_813 = __p2_813; \
65992  int32x4_t __ret_813; \
65993 uint8x8_t __reint_813 = __s2_813; \
65994  __ret_813 = vusdotq_s32(__s0_813, (uint8x16_t)(splatq_lane_s32(*(int32x2_t *) &__reint_813, __p3_813)), __s1_813); \
65995  __ret_813; \
65996 })
65997 #else
65998 #define vsudotq_lane_s32(__p0_814, __p1_814, __p2_814, __p3_814) __extension__ ({ \
65999  int32x4_t __s0_814 = __p0_814; \
66000  int8x16_t __s1_814 = __p1_814; \
66001  uint8x8_t __s2_814 = __p2_814; \
66002  int32x4_t __rev0_814; __rev0_814 = __builtin_shufflevector(__s0_814, __s0_814, 3, 2, 1, 0); \
66003  int8x16_t __rev1_814; __rev1_814 = __builtin_shufflevector(__s1_814, __s1_814, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
66004  uint8x8_t __rev2_814; __rev2_814 = __builtin_shufflevector(__s2_814, __s2_814, 7, 6, 5, 4, 3, 2, 1, 0); \
66005  int32x4_t __ret_814; \
66006 uint8x8_t __reint_814 = __rev2_814; \
66007  __ret_814 = __noswap_vusdotq_s32(__rev0_814, (uint8x16_t)(__noswap_splatq_lane_s32(*(int32x2_t *) &__reint_814, __p3_814)), __rev1_814); \
66008  __ret_814 = __builtin_shufflevector(__ret_814, __ret_814, 3, 2, 1, 0); \
66009  __ret_814; \
66010 })
66011 #endif
66012 
66013 #ifdef __LITTLE_ENDIAN__
66014 #define vsudot_lane_s32(__p0_815, __p1_815, __p2_815, __p3_815) __extension__ ({ \
66015  int32x2_t __s0_815 = __p0_815; \
66016  int8x8_t __s1_815 = __p1_815; \
66017  uint8x8_t __s2_815 = __p2_815; \
66018  int32x2_t __ret_815; \
66019 uint8x8_t __reint_815 = __s2_815; \
66020  __ret_815 = vusdot_s32(__s0_815, (uint8x8_t)(splat_lane_s32(*(int32x2_t *) &__reint_815, __p3_815)), __s1_815); \
66021  __ret_815; \
66022 })
66023 #else
66024 #define vsudot_lane_s32(__p0_816, __p1_816, __p2_816, __p3_816) __extension__ ({ \
66025  int32x2_t __s0_816 = __p0_816; \
66026  int8x8_t __s1_816 = __p1_816; \
66027  uint8x8_t __s2_816 = __p2_816; \
66028  int32x2_t __rev0_816; __rev0_816 = __builtin_shufflevector(__s0_816, __s0_816, 1, 0); \
66029  int8x8_t __rev1_816; __rev1_816 = __builtin_shufflevector(__s1_816, __s1_816, 7, 6, 5, 4, 3, 2, 1, 0); \
66030  uint8x8_t __rev2_816; __rev2_816 = __builtin_shufflevector(__s2_816, __s2_816, 7, 6, 5, 4, 3, 2, 1, 0); \
66031  int32x2_t __ret_816; \
66032 uint8x8_t __reint_816 = __rev2_816; \
66033  __ret_816 = __noswap_vusdot_s32(__rev0_816, (uint8x8_t)(__noswap_splat_lane_s32(*(int32x2_t *) &__reint_816, __p3_816)), __rev1_816); \
66034  __ret_816 = __builtin_shufflevector(__ret_816, __ret_816, 1, 0); \
66035  __ret_816; \
66036 })
66037 #endif
66038 
66039 #endif
66040 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
66041 __ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
66042  int32_t __ret;
66043  __ret = vqadds_s32(__p0, vqrdmulhs_s32(__p1, __p2));
66044  return __ret;
66045 }
66046 __ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
66047  int16_t __ret;
66048  __ret = vqaddh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
66049  return __ret;
66050 }
66051 #ifdef __LITTLE_ENDIAN__
66052 #define vqrdmlahs_lane_s32(__p0_817, __p1_817, __p2_817, __p3_817) __extension__ ({ \
66053  int32_t __s0_817 = __p0_817; \
66054  int32_t __s1_817 = __p1_817; \
66055  int32x2_t __s2_817 = __p2_817; \
66056  int32_t __ret_817; \
66057  __ret_817 = vqadds_s32(__s0_817, vqrdmulhs_s32(__s1_817, vget_lane_s32(__s2_817, __p3_817))); \
66058  __ret_817; \
66059 })
66060 #else
66061 #define vqrdmlahs_lane_s32(__p0_818, __p1_818, __p2_818, __p3_818) __extension__ ({ \
66062  int32_t __s0_818 = __p0_818; \
66063  int32_t __s1_818 = __p1_818; \
66064  int32x2_t __s2_818 = __p2_818; \
66065  int32x2_t __rev2_818; __rev2_818 = __builtin_shufflevector(__s2_818, __s2_818, 1, 0); \
66066  int32_t __ret_818; \
66067  __ret_818 = vqadds_s32(__s0_818, vqrdmulhs_s32(__s1_818, __noswap_vget_lane_s32(__rev2_818, __p3_818))); \
66068  __ret_818; \
66069 })
66070 #endif
66071 
66072 #ifdef __LITTLE_ENDIAN__
66073 #define vqrdmlahh_lane_s16(__p0_819, __p1_819, __p2_819, __p3_819) __extension__ ({ \
66074  int16_t __s0_819 = __p0_819; \
66075  int16_t __s1_819 = __p1_819; \
66076  int16x4_t __s2_819 = __p2_819; \
66077  int16_t __ret_819; \
66078  __ret_819 = vqaddh_s16(__s0_819, vqrdmulhh_s16(__s1_819, vget_lane_s16(__s2_819, __p3_819))); \
66079  __ret_819; \
66080 })
66081 #else
66082 #define vqrdmlahh_lane_s16(__p0_820, __p1_820, __p2_820, __p3_820) __extension__ ({ \
66083  int16_t __s0_820 = __p0_820; \
66084  int16_t __s1_820 = __p1_820; \
66085  int16x4_t __s2_820 = __p2_820; \
66086  int16x4_t __rev2_820; __rev2_820 = __builtin_shufflevector(__s2_820, __s2_820, 3, 2, 1, 0); \
66087  int16_t __ret_820; \
66088  __ret_820 = vqaddh_s16(__s0_820, vqrdmulhh_s16(__s1_820, __noswap_vget_lane_s16(__rev2_820, __p3_820))); \
66089  __ret_820; \
66090 })
66091 #endif
66092 
66093 #ifdef __LITTLE_ENDIAN__
66094 #define vqrdmlahs_laneq_s32(__p0_821, __p1_821, __p2_821, __p3_821) __extension__ ({ \
66095  int32_t __s0_821 = __p0_821; \
66096  int32_t __s1_821 = __p1_821; \
66097  int32x4_t __s2_821 = __p2_821; \
66098  int32_t __ret_821; \
66099  __ret_821 = vqadds_s32(__s0_821, vqrdmulhs_s32(__s1_821, vgetq_lane_s32(__s2_821, __p3_821))); \
66100  __ret_821; \
66101 })
66102 #else
66103 #define vqrdmlahs_laneq_s32(__p0_822, __p1_822, __p2_822, __p3_822) __extension__ ({ \
66104  int32_t __s0_822 = __p0_822; \
66105  int32_t __s1_822 = __p1_822; \
66106  int32x4_t __s2_822 = __p2_822; \
66107  int32x4_t __rev2_822; __rev2_822 = __builtin_shufflevector(__s2_822, __s2_822, 3, 2, 1, 0); \
66108  int32_t __ret_822; \
66109  __ret_822 = vqadds_s32(__s0_822, vqrdmulhs_s32(__s1_822, __noswap_vgetq_lane_s32(__rev2_822, __p3_822))); \
66110  __ret_822; \
66111 })
66112 #endif
66113 
66114 #ifdef __LITTLE_ENDIAN__
66115 #define vqrdmlahh_laneq_s16(__p0_823, __p1_823, __p2_823, __p3_823) __extension__ ({ \
66116  int16_t __s0_823 = __p0_823; \
66117  int16_t __s1_823 = __p1_823; \
66118  int16x8_t __s2_823 = __p2_823; \
66119  int16_t __ret_823; \
66120  __ret_823 = vqaddh_s16(__s0_823, vqrdmulhh_s16(__s1_823, vgetq_lane_s16(__s2_823, __p3_823))); \
66121  __ret_823; \
66122 })
66123 #else
66124 #define vqrdmlahh_laneq_s16(__p0_824, __p1_824, __p2_824, __p3_824) __extension__ ({ \
66125  int16_t __s0_824 = __p0_824; \
66126  int16_t __s1_824 = __p1_824; \
66127  int16x8_t __s2_824 = __p2_824; \
66128  int16x8_t __rev2_824; __rev2_824 = __builtin_shufflevector(__s2_824, __s2_824, 7, 6, 5, 4, 3, 2, 1, 0); \
66129  int16_t __ret_824; \
66130  __ret_824 = vqaddh_s16(__s0_824, vqrdmulhh_s16(__s1_824, __noswap_vgetq_lane_s16(__rev2_824, __p3_824))); \
66131  __ret_824; \
66132 })
66133 #endif
66134 
66135 __ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
66136  int32_t __ret;
66137  __ret = vqsubs_s32(__p0, vqrdmulhs_s32(__p1, __p2));
66138  return __ret;
66139 }
66140 __ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
66141  int16_t __ret;
66142  __ret = vqsubh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
66143  return __ret;
66144 }
66145 #ifdef __LITTLE_ENDIAN__
66146 #define vqrdmlshs_lane_s32(__p0_825, __p1_825, __p2_825, __p3_825) __extension__ ({ \
66147  int32_t __s0_825 = __p0_825; \
66148  int32_t __s1_825 = __p1_825; \
66149  int32x2_t __s2_825 = __p2_825; \
66150  int32_t __ret_825; \
66151  __ret_825 = vqsubs_s32(__s0_825, vqrdmulhs_s32(__s1_825, vget_lane_s32(__s2_825, __p3_825))); \
66152  __ret_825; \
66153 })
66154 #else
66155 #define vqrdmlshs_lane_s32(__p0_826, __p1_826, __p2_826, __p3_826) __extension__ ({ \
66156  int32_t __s0_826 = __p0_826; \
66157  int32_t __s1_826 = __p1_826; \
66158  int32x2_t __s2_826 = __p2_826; \
66159  int32x2_t __rev2_826; __rev2_826 = __builtin_shufflevector(__s2_826, __s2_826, 1, 0); \
66160  int32_t __ret_826; \
66161  __ret_826 = vqsubs_s32(__s0_826, vqrdmulhs_s32(__s1_826, __noswap_vget_lane_s32(__rev2_826, __p3_826))); \
66162  __ret_826; \
66163 })
66164 #endif
66165 
66166 #ifdef __LITTLE_ENDIAN__
66167 #define vqrdmlshh_lane_s16(__p0_827, __p1_827, __p2_827, __p3_827) __extension__ ({ \
66168  int16_t __s0_827 = __p0_827; \
66169  int16_t __s1_827 = __p1_827; \
66170  int16x4_t __s2_827 = __p2_827; \
66171  int16_t __ret_827; \
66172  __ret_827 = vqsubh_s16(__s0_827, vqrdmulhh_s16(__s1_827, vget_lane_s16(__s2_827, __p3_827))); \
66173  __ret_827; \
66174 })
66175 #else
66176 #define vqrdmlshh_lane_s16(__p0_828, __p1_828, __p2_828, __p3_828) __extension__ ({ \
66177  int16_t __s0_828 = __p0_828; \
66178  int16_t __s1_828 = __p1_828; \
66179  int16x4_t __s2_828 = __p2_828; \
66180  int16x4_t __rev2_828; __rev2_828 = __builtin_shufflevector(__s2_828, __s2_828, 3, 2, 1, 0); \
66181  int16_t __ret_828; \
66182  __ret_828 = vqsubh_s16(__s0_828, vqrdmulhh_s16(__s1_828, __noswap_vget_lane_s16(__rev2_828, __p3_828))); \
66183  __ret_828; \
66184 })
66185 #endif
66186 
66187 #ifdef __LITTLE_ENDIAN__
66188 #define vqrdmlshs_laneq_s32(__p0_829, __p1_829, __p2_829, __p3_829) __extension__ ({ \
66189  int32_t __s0_829 = __p0_829; \
66190  int32_t __s1_829 = __p1_829; \
66191  int32x4_t __s2_829 = __p2_829; \
66192  int32_t __ret_829; \
66193  __ret_829 = vqsubs_s32(__s0_829, vqrdmulhs_s32(__s1_829, vgetq_lane_s32(__s2_829, __p3_829))); \
66194  __ret_829; \
66195 })
66196 #else
66197 #define vqrdmlshs_laneq_s32(__p0_830, __p1_830, __p2_830, __p3_830) __extension__ ({ \
66198  int32_t __s0_830 = __p0_830; \
66199  int32_t __s1_830 = __p1_830; \
66200  int32x4_t __s2_830 = __p2_830; \
66201  int32x4_t __rev2_830; __rev2_830 = __builtin_shufflevector(__s2_830, __s2_830, 3, 2, 1, 0); \
66202  int32_t __ret_830; \
66203  __ret_830 = vqsubs_s32(__s0_830, vqrdmulhs_s32(__s1_830, __noswap_vgetq_lane_s32(__rev2_830, __p3_830))); \
66204  __ret_830; \
66205 })
66206 #endif
66207 
66208 #ifdef __LITTLE_ENDIAN__
66209 #define vqrdmlshh_laneq_s16(__p0_831, __p1_831, __p2_831, __p3_831) __extension__ ({ \
66210  int16_t __s0_831 = __p0_831; \
66211  int16_t __s1_831 = __p1_831; \
66212  int16x8_t __s2_831 = __p2_831; \
66213  int16_t __ret_831; \
66214  __ret_831 = vqsubh_s16(__s0_831, vqrdmulhh_s16(__s1_831, vgetq_lane_s16(__s2_831, __p3_831))); \
66215  __ret_831; \
66216 })
66217 #else
66218 #define vqrdmlshh_laneq_s16(__p0_832, __p1_832, __p2_832, __p3_832) __extension__ ({ \
66219  int16_t __s0_832 = __p0_832; \
66220  int16_t __s1_832 = __p1_832; \
66221  int16x8_t __s2_832 = __p2_832; \
66222  int16x8_t __rev2_832; __rev2_832 = __builtin_shufflevector(__s2_832, __s2_832, 7, 6, 5, 4, 3, 2, 1, 0); \
66223  int16_t __ret_832; \
66224  __ret_832 = vqsubh_s16(__s0_832, vqrdmulhh_s16(__s1_832, __noswap_vgetq_lane_s16(__rev2_832, __p3_832))); \
66225  __ret_832; \
66226 })
66227 #endif
66228 
66229 #endif
66230 #if defined(__aarch64__)
66231 #ifdef __LITTLE_ENDIAN__
66232 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
66233  uint16x8_t __ret;
66234  __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
66235  return __ret;
66236 }
66237 #else
66238 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
66239  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66240  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66241  uint16x8_t __ret;
66242  __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
66243  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66244  return __ret;
66245 }
66246 #endif
66247 
66248 #ifdef __LITTLE_ENDIAN__
66249 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
66250  uint64x2_t __ret;
66251  __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
66252  return __ret;
66253 }
66254 #else
66255 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
66256  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66257  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66258  uint64x2_t __ret;
66259  __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
66260  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66261  return __ret;
66262 }
66263 #endif
66264 
66265 #ifdef __LITTLE_ENDIAN__
66266 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
66267  uint32x4_t __ret;
66268  __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
66269  return __ret;
66270 }
66271 #else
66272 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
66273  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66274  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66275  uint32x4_t __ret;
66276  __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
66277  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66278  return __ret;
66279 }
66280 #endif
66281 
66282 #ifdef __LITTLE_ENDIAN__
66283 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
66284  int16x8_t __ret;
66285  __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
66286  return __ret;
66287 }
66288 #else
66289 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
66290  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66291  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66292  int16x8_t __ret;
66293  __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
66294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66295  return __ret;
66296 }
66297 #endif
66298 
66299 #ifdef __LITTLE_ENDIAN__
66300 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
66301  int64x2_t __ret;
66302  __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
66303  return __ret;
66304 }
66305 #else
66306 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
66307  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66308  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66309  int64x2_t __ret;
66310  __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
66311  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66312  return __ret;
66313 }
66314 #endif
66315 
66316 #ifdef __LITTLE_ENDIAN__
66317 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
66318  int32x4_t __ret;
66319  __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
66320  return __ret;
66321 }
66322 #else
66323 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
66324  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66325  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66326  int32x4_t __ret;
66327  __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
66328  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66329  return __ret;
66330 }
66331 #endif
66332 
66333 #ifdef __LITTLE_ENDIAN__
66334 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
66335  uint16x8_t __ret;
66336  __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
66337  return __ret;
66338 }
66339 #else
66340 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
66341  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66342  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66343  uint16x8_t __ret;
66344  __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
66345  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66346  return __ret;
66347 }
66348 #endif
66349 
66350 #ifdef __LITTLE_ENDIAN__
66351 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
66352  uint64x2_t __ret;
66353  __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
66354  return __ret;
66355 }
66356 #else
66357 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
66358  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66359  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66360  uint64x2_t __ret;
66361  __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
66362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66363  return __ret;
66364 }
66365 #endif
66366 
66367 #ifdef __LITTLE_ENDIAN__
66368 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
66369  uint32x4_t __ret;
66370  __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
66371  return __ret;
66372 }
66373 #else
66374 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
66375  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66376  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66377  uint32x4_t __ret;
66378  __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
66379  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66380  return __ret;
66381 }
66382 #endif
66383 
66384 #ifdef __LITTLE_ENDIAN__
66385 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
66386  int16x8_t __ret;
66387  __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
66388  return __ret;
66389 }
66390 #else
66391 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
66392  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66393  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66394  int16x8_t __ret;
66395  __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
66396  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66397  return __ret;
66398 }
66399 #endif
66400 
66401 #ifdef __LITTLE_ENDIAN__
66402 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
66403  int64x2_t __ret;
66404  __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
66405  return __ret;
66406 }
66407 #else
66408 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
66409  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66410  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66411  int64x2_t __ret;
66412  __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
66413  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66414  return __ret;
66415 }
66416 #endif
66417 
66418 #ifdef __LITTLE_ENDIAN__
66419 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
66420  int32x4_t __ret;
66421  __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
66422  return __ret;
66423 }
66424 #else
66425 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
66426  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66427  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66428  int32x4_t __ret;
66429  __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
66430  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66431  return __ret;
66432 }
66433 #endif
66434 
66435 #ifdef __LITTLE_ENDIAN__
66436 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
66437  uint16x8_t __ret;
66438  __ret = __p0 + vmovl_high_u8(__p1);
66439  return __ret;
66440 }
66441 #else
66442 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
66443  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66444  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66445  uint16x8_t __ret;
66446  __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
66447  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66448  return __ret;
66449 }
66450 #endif
66451 
66452 #ifdef __LITTLE_ENDIAN__
66453 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
66454  uint64x2_t __ret;
66455  __ret = __p0 + vmovl_high_u32(__p1);
66456  return __ret;
66457 }
66458 #else
66459 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
66460  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66461  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66462  uint64x2_t __ret;
66463  __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
66464  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66465  return __ret;
66466 }
66467 #endif
66468 
66469 #ifdef __LITTLE_ENDIAN__
66470 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
66471  uint32x4_t __ret;
66472  __ret = __p0 + vmovl_high_u16(__p1);
66473  return __ret;
66474 }
66475 #else
66476 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
66477  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66478  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66479  uint32x4_t __ret;
66480  __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
66481  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66482  return __ret;
66483 }
66484 #endif
66485 
66486 #ifdef __LITTLE_ENDIAN__
66487 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
66488  int16x8_t __ret;
66489  __ret = __p0 + vmovl_high_s8(__p1);
66490  return __ret;
66491 }
66492 #else
66493 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
66494  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66495  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66496  int16x8_t __ret;
66497  __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
66498  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66499  return __ret;
66500 }
66501 #endif
66502 
66503 #ifdef __LITTLE_ENDIAN__
66504 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
66505  int64x2_t __ret;
66506  __ret = __p0 + vmovl_high_s32(__p1);
66507  return __ret;
66508 }
66509 #else
66510 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
66511  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66512  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66513  int64x2_t __ret;
66514  __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
66515  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66516  return __ret;
66517 }
66518 #endif
66519 
66520 #ifdef __LITTLE_ENDIAN__
66521 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
66522  int32x4_t __ret;
66523  __ret = __p0 + vmovl_high_s16(__p1);
66524  return __ret;
66525 }
66526 #else
66527 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
66528  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66529  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66530  int32x4_t __ret;
66531  __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
66532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66533  return __ret;
66534 }
66535 #endif
66536 
66537 #ifdef __LITTLE_ENDIAN__
66538 #define vcopyq_lane_p64(__p0_833, __p1_833, __p2_833, __p3_833) __extension__ ({ \
66539  poly64x2_t __s0_833 = __p0_833; \
66540  poly64x1_t __s2_833 = __p2_833; \
66541  poly64x2_t __ret_833; \
66542  __ret_833 = vsetq_lane_p64(vget_lane_p64(__s2_833, __p3_833), __s0_833, __p1_833); \
66543  __ret_833; \
66544 })
66545 #else
66546 #define vcopyq_lane_p64(__p0_834, __p1_834, __p2_834, __p3_834) __extension__ ({ \
66547  poly64x2_t __s0_834 = __p0_834; \
66548  poly64x1_t __s2_834 = __p2_834; \
66549  poly64x2_t __rev0_834; __rev0_834 = __builtin_shufflevector(__s0_834, __s0_834, 1, 0); \
66550  poly64x2_t __ret_834; \
66551  __ret_834 = __noswap_vsetq_lane_p64(vget_lane_p64(__s2_834, __p3_834), __rev0_834, __p1_834); \
66552  __ret_834 = __builtin_shufflevector(__ret_834, __ret_834, 1, 0); \
66553  __ret_834; \
66554 })
66555 #endif
66556 
66557 #ifdef __LITTLE_ENDIAN__
66558 #define vcopyq_lane_f64(__p0_835, __p1_835, __p2_835, __p3_835) __extension__ ({ \
66559  float64x2_t __s0_835 = __p0_835; \
66560  float64x1_t __s2_835 = __p2_835; \
66561  float64x2_t __ret_835; \
66562  __ret_835 = vsetq_lane_f64(vget_lane_f64(__s2_835, __p3_835), __s0_835, __p1_835); \
66563  __ret_835; \
66564 })
66565 #else
66566 #define vcopyq_lane_f64(__p0_836, __p1_836, __p2_836, __p3_836) __extension__ ({ \
66567  float64x2_t __s0_836 = __p0_836; \
66568  float64x1_t __s2_836 = __p2_836; \
66569  float64x2_t __rev0_836; __rev0_836 = __builtin_shufflevector(__s0_836, __s0_836, 1, 0); \
66570  float64x2_t __ret_836; \
66571  __ret_836 = __noswap_vsetq_lane_f64(vget_lane_f64(__s2_836, __p3_836), __rev0_836, __p1_836); \
66572  __ret_836 = __builtin_shufflevector(__ret_836, __ret_836, 1, 0); \
66573  __ret_836; \
66574 })
66575 #endif
66576 
66577 #define vcopy_lane_p64(__p0_837, __p1_837, __p2_837, __p3_837) __extension__ ({ \
66578  poly64x1_t __s0_837 = __p0_837; \
66579  poly64x1_t __s2_837 = __p2_837; \
66580  poly64x1_t __ret_837; \
66581  __ret_837 = vset_lane_p64(vget_lane_p64(__s2_837, __p3_837), __s0_837, __p1_837); \
66582  __ret_837; \
66583 })
66584 #define vcopy_lane_f64(__p0_838, __p1_838, __p2_838, __p3_838) __extension__ ({ \
66585  float64x1_t __s0_838 = __p0_838; \
66586  float64x1_t __s2_838 = __p2_838; \
66587  float64x1_t __ret_838; \
66588  __ret_838 = vset_lane_f64(vget_lane_f64(__s2_838, __p3_838), __s0_838, __p1_838); \
66589  __ret_838; \
66590 })
66591 #ifdef __LITTLE_ENDIAN__
66592 #define vcopyq_laneq_p64(__p0_839, __p1_839, __p2_839, __p3_839) __extension__ ({ \
66593  poly64x2_t __s0_839 = __p0_839; \
66594  poly64x2_t __s2_839 = __p2_839; \
66595  poly64x2_t __ret_839; \
66596  __ret_839 = vsetq_lane_p64(vgetq_lane_p64(__s2_839, __p3_839), __s0_839, __p1_839); \
66597  __ret_839; \
66598 })
66599 #else
66600 #define vcopyq_laneq_p64(__p0_840, __p1_840, __p2_840, __p3_840) __extension__ ({ \
66601  poly64x2_t __s0_840 = __p0_840; \
66602  poly64x2_t __s2_840 = __p2_840; \
66603  poly64x2_t __rev0_840; __rev0_840 = __builtin_shufflevector(__s0_840, __s0_840, 1, 0); \
66604  poly64x2_t __rev2_840; __rev2_840 = __builtin_shufflevector(__s2_840, __s2_840, 1, 0); \
66605  poly64x2_t __ret_840; \
66606  __ret_840 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_840, __p3_840), __rev0_840, __p1_840); \
66607  __ret_840 = __builtin_shufflevector(__ret_840, __ret_840, 1, 0); \
66608  __ret_840; \
66609 })
66610 #endif
66611 
66612 #ifdef __LITTLE_ENDIAN__
66613 #define vcopyq_laneq_f64(__p0_841, __p1_841, __p2_841, __p3_841) __extension__ ({ \
66614  float64x2_t __s0_841 = __p0_841; \
66615  float64x2_t __s2_841 = __p2_841; \
66616  float64x2_t __ret_841; \
66617  __ret_841 = vsetq_lane_f64(vgetq_lane_f64(__s2_841, __p3_841), __s0_841, __p1_841); \
66618  __ret_841; \
66619 })
66620 #else
66621 #define vcopyq_laneq_f64(__p0_842, __p1_842, __p2_842, __p3_842) __extension__ ({ \
66622  float64x2_t __s0_842 = __p0_842; \
66623  float64x2_t __s2_842 = __p2_842; \
66624  float64x2_t __rev0_842; __rev0_842 = __builtin_shufflevector(__s0_842, __s0_842, 1, 0); \
66625  float64x2_t __rev2_842; __rev2_842 = __builtin_shufflevector(__s2_842, __s2_842, 1, 0); \
66626  float64x2_t __ret_842; \
66627  __ret_842 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_842, __p3_842), __rev0_842, __p1_842); \
66628  __ret_842 = __builtin_shufflevector(__ret_842, __ret_842, 1, 0); \
66629  __ret_842; \
66630 })
66631 #endif
66632 
66633 #ifdef __LITTLE_ENDIAN__
66634 #define vcopy_laneq_p64(__p0_843, __p1_843, __p2_843, __p3_843) __extension__ ({ \
66635  poly64x1_t __s0_843 = __p0_843; \
66636  poly64x2_t __s2_843 = __p2_843; \
66637  poly64x1_t __ret_843; \
66638  __ret_843 = vset_lane_p64(vgetq_lane_p64(__s2_843, __p3_843), __s0_843, __p1_843); \
66639  __ret_843; \
66640 })
66641 #else
66642 #define vcopy_laneq_p64(__p0_844, __p1_844, __p2_844, __p3_844) __extension__ ({ \
66643  poly64x1_t __s0_844 = __p0_844; \
66644  poly64x2_t __s2_844 = __p2_844; \
66645  poly64x2_t __rev2_844; __rev2_844 = __builtin_shufflevector(__s2_844, __s2_844, 1, 0); \
66646  poly64x1_t __ret_844; \
66647  __ret_844 = vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_844, __p3_844), __s0_844, __p1_844); \
66648  __ret_844; \
66649 })
66650 #endif
66651 
66652 #ifdef __LITTLE_ENDIAN__
66653 #define vcopy_laneq_f64(__p0_845, __p1_845, __p2_845, __p3_845) __extension__ ({ \
66654  float64x1_t __s0_845 = __p0_845; \
66655  float64x2_t __s2_845 = __p2_845; \
66656  float64x1_t __ret_845; \
66657  __ret_845 = vset_lane_f64(vgetq_lane_f64(__s2_845, __p3_845), __s0_845, __p1_845); \
66658  __ret_845; \
66659 })
66660 #else
66661 #define vcopy_laneq_f64(__p0_846, __p1_846, __p2_846, __p3_846) __extension__ ({ \
66662  float64x1_t __s0_846 = __p0_846; \
66663  float64x2_t __s2_846 = __p2_846; \
66664  float64x2_t __rev2_846; __rev2_846 = __builtin_shufflevector(__s2_846, __s2_846, 1, 0); \
66665  float64x1_t __ret_846; \
66666  __ret_846 = vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_846, __p3_846), __s0_846, __p1_846); \
66667  __ret_846; \
66668 })
66669 #endif
66670 
66671 #ifdef __LITTLE_ENDIAN__
66672 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66673  uint16x8_t __ret;
66674  __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
66675  return __ret;
66676 }
66677 #else
66678 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66679  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66680  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66681  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66682  uint16x8_t __ret;
66683  __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
66684  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66685  return __ret;
66686 }
66687 #endif
66688 
66689 #ifdef __LITTLE_ENDIAN__
66690 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66691  uint64x2_t __ret;
66692  __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
66693  return __ret;
66694 }
66695 #else
66696 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66697  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66698  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66699  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66700  uint64x2_t __ret;
66701  __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
66702  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66703  return __ret;
66704 }
66705 #endif
66706 
66707 #ifdef __LITTLE_ENDIAN__
66708 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66709  uint32x4_t __ret;
66710  __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
66711  return __ret;
66712 }
66713 #else
66714 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66715  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66716  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66717  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66718  uint32x4_t __ret;
66719  __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
66720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66721  return __ret;
66722 }
66723 #endif
66724 
66725 #ifdef __LITTLE_ENDIAN__
66726 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
66727  int16x8_t __ret;
66728  __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
66729  return __ret;
66730 }
66731 #else
66732 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
66733  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66734  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66735  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66736  int16x8_t __ret;
66737  __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
66738  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66739  return __ret;
66740 }
66741 #endif
66742 
66743 #ifdef __LITTLE_ENDIAN__
66744 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
66745  int64x2_t __ret;
66746  __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
66747  return __ret;
66748 }
66749 #else
66750 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
66751  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66752  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66753  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66754  int64x2_t __ret;
66755  __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
66756  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66757  return __ret;
66758 }
66759 #endif
66760 
66761 #ifdef __LITTLE_ENDIAN__
66762 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
66763  int32x4_t __ret;
66764  __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
66765  return __ret;
66766 }
66767 #else
66768 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
66769  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66770  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66771  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66772  int32x4_t __ret;
66773  __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
66774  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66775  return __ret;
66776 }
66777 #endif
66778 
66779 #ifdef __LITTLE_ENDIAN__
66780 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
66781  uint64x2_t __ret;
66782  __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
66783  return __ret;
66784 }
66785 #else
66786 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
66787  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66788  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66789  uint64x2_t __ret;
66790  __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
66791  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66792  return __ret;
66793 }
66794 #endif
66795 
66796 #ifdef __LITTLE_ENDIAN__
66797 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
66798  uint32x4_t __ret;
66799  __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
66800  return __ret;
66801 }
66802 #else
66803 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
66804  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66805  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66806  uint32x4_t __ret;
66807  __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
66808  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66809  return __ret;
66810 }
66811 #endif
66812 
66813 #ifdef __LITTLE_ENDIAN__
66814 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
66815  int64x2_t __ret;
66816  __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
66817  return __ret;
66818 }
66819 #else
66820 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
66821  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66822  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66823  int64x2_t __ret;
66824  __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
66825  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66826  return __ret;
66827 }
66828 #endif
66829 
66830 #ifdef __LITTLE_ENDIAN__
66831 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
66832  int32x4_t __ret;
66833  __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
66834  return __ret;
66835 }
66836 #else
66837 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
66838  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66839  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66840  int32x4_t __ret;
66841  __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
66842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66843  return __ret;
66844 }
66845 #endif
66846 
66847 #ifdef __LITTLE_ENDIAN__
66848 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66849  uint16x8_t __ret;
66850  __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
66851  return __ret;
66852 }
66853 #else
66854 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66855  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66856  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66857  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66858  uint16x8_t __ret;
66859  __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
66860  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66861  return __ret;
66862 }
66863 #endif
66864 
66865 #ifdef __LITTLE_ENDIAN__
66866 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66867  uint64x2_t __ret;
66868  __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
66869  return __ret;
66870 }
66871 #else
66872 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66873  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66874  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66875  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66876  uint64x2_t __ret;
66877  __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
66878  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66879  return __ret;
66880 }
66881 #endif
66882 
66883 #ifdef __LITTLE_ENDIAN__
66884 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66885  uint32x4_t __ret;
66886  __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
66887  return __ret;
66888 }
66889 #else
66890 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66891  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66892  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66893  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66894  uint32x4_t __ret;
66895  __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
66896  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66897  return __ret;
66898 }
66899 #endif
66900 
66901 #ifdef __LITTLE_ENDIAN__
66902 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
66903  int16x8_t __ret;
66904  __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
66905  return __ret;
66906 }
66907 #else
66908 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
66909  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66910  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66911  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66912  int16x8_t __ret;
66913  __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
66914  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66915  return __ret;
66916 }
66917 #endif
66918 
66919 #ifdef __LITTLE_ENDIAN__
66920 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
66921  int64x2_t __ret;
66922  __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
66923  return __ret;
66924 }
66925 #else
66926 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
66927  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66928  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66929  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66930  int64x2_t __ret;
66931  __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
66932  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66933  return __ret;
66934 }
66935 #endif
66936 
66937 #ifdef __LITTLE_ENDIAN__
66938 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
66939  int32x4_t __ret;
66940  __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
66941  return __ret;
66942 }
66943 #else
66944 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
66945  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66946  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66947  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66948  int32x4_t __ret;
66949  __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
66950  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66951  return __ret;
66952 }
66953 #endif
66954 
66955 #ifdef __LITTLE_ENDIAN__
66956 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
66957  uint64x2_t __ret;
66958  __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
66959  return __ret;
66960 }
66961 #else
66962 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
66963  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66964  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66965  uint64x2_t __ret;
66966  __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
66967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66968  return __ret;
66969 }
66970 #endif
66971 
66972 #ifdef __LITTLE_ENDIAN__
66973 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
66974  uint32x4_t __ret;
66975  __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
66976  return __ret;
66977 }
66978 #else
66979 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
66980  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66981  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66982  uint32x4_t __ret;
66983  __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
66984  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66985  return __ret;
66986 }
66987 #endif
66988 
66989 #ifdef __LITTLE_ENDIAN__
66990 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
66991  int64x2_t __ret;
66992  __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
66993  return __ret;
66994 }
66995 #else
66996 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
66997  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66998  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66999  int64x2_t __ret;
67000  __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
67001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67002  return __ret;
67003 }
67004 #endif
67005 
67006 #ifdef __LITTLE_ENDIAN__
67007 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
67008  int32x4_t __ret;
67009  __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
67010  return __ret;
67011 }
67012 #else
67013 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
67014  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67015  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67016  int32x4_t __ret;
67017  __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
67018  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67019  return __ret;
67020 }
67021 #endif
67022 
67023 #define vmulx_lane_f64(__p0_847, __p1_847, __p2_847) __extension__ ({ \
67024  float64x1_t __s0_847 = __p0_847; \
67025  float64x1_t __s1_847 = __p1_847; \
67026  float64x1_t __ret_847; \
67027  float64_t __x_847 = vget_lane_f64(__s0_847, 0); \
67028  float64_t __y_847 = vget_lane_f64(__s1_847, __p2_847); \
67029  float64_t __z_847 = vmulxd_f64(__x_847, __y_847); \
67030  __ret_847 = vset_lane_f64(__z_847, __s0_847, __p2_847); \
67031  __ret_847; \
67032 })
67033 #ifdef __LITTLE_ENDIAN__
67034 #define vmulx_laneq_f64(__p0_848, __p1_848, __p2_848) __extension__ ({ \
67035  float64x1_t __s0_848 = __p0_848; \
67036  float64x2_t __s1_848 = __p1_848; \
67037  float64x1_t __ret_848; \
67038  float64_t __x_848 = vget_lane_f64(__s0_848, 0); \
67039  float64_t __y_848 = vgetq_lane_f64(__s1_848, __p2_848); \
67040  float64_t __z_848 = vmulxd_f64(__x_848, __y_848); \
67041  __ret_848 = vset_lane_f64(__z_848, __s0_848, 0); \
67042  __ret_848; \
67043 })
67044 #else
67045 #define vmulx_laneq_f64(__p0_849, __p1_849, __p2_849) __extension__ ({ \
67046  float64x1_t __s0_849 = __p0_849; \
67047  float64x2_t __s1_849 = __p1_849; \
67048  float64x2_t __rev1_849; __rev1_849 = __builtin_shufflevector(__s1_849, __s1_849, 1, 0); \
67049  float64x1_t __ret_849; \
67050  float64_t __x_849 = vget_lane_f64(__s0_849, 0); \
67051  float64_t __y_849 = __noswap_vgetq_lane_f64(__rev1_849, __p2_849); \
67052  float64_t __z_849 = vmulxd_f64(__x_849, __y_849); \
67053  __ret_849 = vset_lane_f64(__z_849, __s0_849, 0); \
67054  __ret_849; \
67055 })
67056 #endif
67057 
67058 #endif
67059 #ifdef __LITTLE_ENDIAN__
67060 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67061  uint16x8_t __ret;
67062  __ret = __p0 + vabdl_u8(__p1, __p2);
67063  return __ret;
67064 }
67065 #else
67066 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67067  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67068  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67069  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67070  uint16x8_t __ret;
67071  __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
67072  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67073  return __ret;
67074 }
67075 __ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67076  uint16x8_t __ret;
67077  __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
67078  return __ret;
67079 }
67080 #endif
67081 
67082 #ifdef __LITTLE_ENDIAN__
67083 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67084  uint64x2_t __ret;
67085  __ret = __p0 + vabdl_u32(__p1, __p2);
67086  return __ret;
67087 }
67088 #else
67089 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67090  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67091  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67092  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67093  uint64x2_t __ret;
67094  __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
67095  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67096  return __ret;
67097 }
67098 __ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67099  uint64x2_t __ret;
67100  __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
67101  return __ret;
67102 }
67103 #endif
67104 
67105 #ifdef __LITTLE_ENDIAN__
67106 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67107  uint32x4_t __ret;
67108  __ret = __p0 + vabdl_u16(__p1, __p2);
67109  return __ret;
67110 }
67111 #else
67112 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67113  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67114  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67115  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67116  uint32x4_t __ret;
67117  __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
67118  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67119  return __ret;
67120 }
67121 __ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67122  uint32x4_t __ret;
67123  __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
67124  return __ret;
67125 }
67126 #endif
67127 
67128 #ifdef __LITTLE_ENDIAN__
67129 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67130  int16x8_t __ret;
67131  __ret = __p0 + vabdl_s8(__p1, __p2);
67132  return __ret;
67133 }
67134 #else
67135 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67136  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67137  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67138  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67139  int16x8_t __ret;
67140  __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
67141  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67142  return __ret;
67143 }
67144 __ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67145  int16x8_t __ret;
67146  __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
67147  return __ret;
67148 }
67149 #endif
67150 
67151 #ifdef __LITTLE_ENDIAN__
67152 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67153  int64x2_t __ret;
67154  __ret = __p0 + vabdl_s32(__p1, __p2);
67155  return __ret;
67156 }
67157 #else
67158 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67159  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67160  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67161  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67162  int64x2_t __ret;
67163  __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
67164  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67165  return __ret;
67166 }
67167 __ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67168  int64x2_t __ret;
67169  __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
67170  return __ret;
67171 }
67172 #endif
67173 
67174 #ifdef __LITTLE_ENDIAN__
67175 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67176  int32x4_t __ret;
67177  __ret = __p0 + vabdl_s16(__p1, __p2);
67178  return __ret;
67179 }
67180 #else
67181 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67182  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67183  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67184  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67185  int32x4_t __ret;
67186  __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
67187  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67188  return __ret;
67189 }
67190 __ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67191  int32x4_t __ret;
67192  __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
67193  return __ret;
67194 }
67195 #endif
67196 
67197 #if defined(__aarch64__)
67198 #ifdef __LITTLE_ENDIAN__
67199 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67200  uint16x8_t __ret;
67201  __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
67202  return __ret;
67203 }
67204 #else
67205 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67206  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67207  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67208  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67209  uint16x8_t __ret;
67210  __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
67211  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67212  return __ret;
67213 }
67214 #endif
67215 
67216 #ifdef __LITTLE_ENDIAN__
67217 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67218  uint64x2_t __ret;
67219  __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
67220  return __ret;
67221 }
67222 #else
67223 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67224  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67225  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67226  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67227  uint64x2_t __ret;
67228  __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
67229  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67230  return __ret;
67231 }
67232 #endif
67233 
67234 #ifdef __LITTLE_ENDIAN__
67235 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67236  uint32x4_t __ret;
67237  __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
67238  return __ret;
67239 }
67240 #else
67241 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67242  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67243  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67244  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67245  uint32x4_t __ret;
67246  __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
67247  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67248  return __ret;
67249 }
67250 #endif
67251 
67252 #ifdef __LITTLE_ENDIAN__
67253 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67254  int16x8_t __ret;
67255  __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
67256  return __ret;
67257 }
67258 #else
67259 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67260  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67261  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67262  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67263  int16x8_t __ret;
67264  __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
67265  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67266  return __ret;
67267 }
67268 #endif
67269 
67270 #ifdef __LITTLE_ENDIAN__
67271 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67272  int64x2_t __ret;
67273  __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
67274  return __ret;
67275 }
67276 #else
67277 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67278  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67279  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67280  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67281  int64x2_t __ret;
67282  __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
67283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67284  return __ret;
67285 }
67286 #endif
67287 
67288 #ifdef __LITTLE_ENDIAN__
67289 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
67290  int32x4_t __ret;
67291  __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
67292  return __ret;
67293 }
67294 #else
67295 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
67296  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67297  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67298  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67299  int32x4_t __ret;
67300  __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
67301  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67302  return __ret;
67303 }
67304 #endif
67305 
67306 #endif
67307 
67308 #undef __ai
67309 
67310 #endif /* if !defined(__ARM_NEON) */
67311 #endif /* ifndef __ARM_FP */
67312 #endif /* __ARM_NEON_H */
_Float16 __2f16 __attribute__((ext_vector_type(2)))
Zeroes the upper 128 bits (bits 255:128) of all YMM registers.
__bf16 bfloat16_t
Definition: arm_bf16.h:14
#define __ai
Definition: arm_bf16.h:15
static __inline__ uint32_t
Definition: arm_cde.h:25
static __inline__ uint64_t
Definition: arm_cde.h:31
__fp16 float16_t
Definition: arm_fp16.h:29
static __inline__ int32_t
Definition: arm_mve.h:51
static __inline__ int8x16x2_t
Definition: arm_mve.h:7233
static __inline__ int32x4x4_t
Definition: arm_mve.h:7253
static __inline__ int16x8x4_t
Definition: arm_mve.h:7249
static __inline__ uint8x16_t
Definition: arm_mve.h:103
static __inline__ int8x16x4_t
Definition: arm_mve.h:7257
static __inline__ int8x16_t
Definition: arm_mve.h:91
static __inline__ uint16x8x4_t
Definition: arm_mve.h:7261
static __inline__ uint64x2_t
Definition: arm_mve.h:2539
static __inline__ int8_t
Definition: arm_mve.h:311
static __inline__ int32x4x2_t
Definition: arm_mve.h:7229
static __inline__ uint16x8x2_t
Definition: arm_mve.h:7237
static __inline__ uint16x8_t
Definition: arm_mve.h:95
static __inline__ uint32x4_t
Definition: arm_mve.h:99
static __inline__ int16x8_t
Definition: arm_mve.h:83
static __inline__ int32x4_t
Definition: arm_mve.h:87
static __inline__ uint32x4x2_t
Definition: arm_mve.h:7241
static __inline__ uint8x16x4_t
Definition: arm_mve.h:7269
static __inline__ uint8_t
Definition: arm_mve.h:323
static __inline__ int64x2_t
Definition: arm_mve.h:4433
static __inline__ int16x8x2_t
Definition: arm_mve.h:7225
static __inline__ uint32x4x4_t
Definition: arm_mve.h:7265
static __inline__ uint8x16x2_t
Definition: arm_mve.h:7245
static __inline__ int16_t
Definition: arm_mve.h:303
static __inline__ uint16_t
Definition: arm_mve.h:315
int16x8_t val[2]
Definition: arm_mve.h:26
int16x8_t val[4]
Definition: arm_mve.h:27
int32x4_t val[2]
Definition: arm_mve.h:29
int32x4_t val[4]
Definition: arm_mve.h:30
int64x2_t val[2]
Definition: arm_mve.h:32
int64x2_t val[4]
Definition: arm_mve.h:33
int8x16_t val[2]
Definition: arm_mve.h:35
int8x16_t val[4]
Definition: arm_mve.h:36
uint16x8_t val[2]
Definition: arm_mve.h:38
uint16x8_t val[4]
Definition: arm_mve.h:39
uint32x4_t val[2]
Definition: arm_mve.h:41
uint32x4_t val[4]
Definition: arm_mve.h:42
uint64x2_t val[2]
Definition: arm_mve.h:44
uint64x2_t val[4]
Definition: arm_mve.h:45
uint8x16_t val[2]
Definition: arm_mve.h:47
uint8x16_t val[4]
Definition: arm_mve.h:48