sketchbook
 All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
arm_neon.h
Go to the documentation of this file.
1 /*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2  *
3  * Permission is hereby granted, free of charge, to any person obtaining a copy
4  * of this software and associated documentation files (the "Software"), to deal
5  * in the Software without restriction, including without limitation the rights
6  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7  * copies of the Software, and to permit persons to whom the Software is
8  * furnished to do so, subject to the following conditions:
9  *
10  * The above copyright notice and this permission notice shall be included in
11  * all copies or substantial portions of the Software.
12  *
13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19  * THE SOFTWARE.
20  *
21  *===-----------------------------------------------------------------------===
22  */
23 
24 #ifndef __ARM_NEON_H
25 #define __ARM_NEON_H
26 
27 #if !defined(__ARM_NEON)
28 #error "NEON support not enabled"
29 #endif
30 
31 #include <stdint.h>
32 
33 typedef float float32_t;
34 typedef __fp16 float16_t;
35 #ifdef __aarch64__
36 typedef double float64_t;
37 #endif
38 
39 #ifdef __aarch64__
40 typedef uint8_t poly8_t;
41 typedef uint16_t poly16_t;
42 typedef uint64_t poly64_t;
43 typedef __uint128_t poly128_t;
44 #else
45 typedef int8_t poly8_t;
46 typedef int16_t poly16_t;
47 #endif
48 typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
49 typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
50 typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
51 typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
52 typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
53 typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
54 typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
55 typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
56 typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
57 typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
58 typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
59 typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
60 typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
61 typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
62 typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
63 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
64 typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
65 typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
66 typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
67 typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
68 #ifdef __aarch64__
69 typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
70 typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
71 #endif
72 typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
73 typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
74 typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
75 typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
76 #ifdef __aarch64__
77 typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
78 typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
79 #endif
80 
81 typedef struct int8x8x2_t {
82  int8x8_t val[2];
83 } int8x8x2_t;
84 
85 typedef struct int8x16x2_t {
86  int8x16_t val[2];
87 } int8x16x2_t;
88 
89 typedef struct int16x4x2_t {
90  int16x4_t val[2];
91 } int16x4x2_t;
92 
93 typedef struct int16x8x2_t {
94  int16x8_t val[2];
95 } int16x8x2_t;
96 
97 typedef struct int32x2x2_t {
98  int32x2_t val[2];
99 } int32x2x2_t;
100 
101 typedef struct int32x4x2_t {
102  int32x4_t val[2];
103 } int32x4x2_t;
104 
105 typedef struct int64x1x2_t {
106  int64x1_t val[2];
107 } int64x1x2_t;
108 
109 typedef struct int64x2x2_t {
110  int64x2_t val[2];
111 } int64x2x2_t;
112 
113 typedef struct uint8x8x2_t {
114  uint8x8_t val[2];
115 } uint8x8x2_t;
116 
117 typedef struct uint8x16x2_t {
118  uint8x16_t val[2];
119 } uint8x16x2_t;
120 
121 typedef struct uint16x4x2_t {
122  uint16x4_t val[2];
123 } uint16x4x2_t;
124 
125 typedef struct uint16x8x2_t {
126  uint16x8_t val[2];
127 } uint16x8x2_t;
128 
129 typedef struct uint32x2x2_t {
130  uint32x2_t val[2];
131 } uint32x2x2_t;
132 
133 typedef struct uint32x4x2_t {
134  uint32x4_t val[2];
135 } uint32x4x2_t;
136 
137 typedef struct uint64x1x2_t {
138  uint64x1_t val[2];
139 } uint64x1x2_t;
140 
141 typedef struct uint64x2x2_t {
142  uint64x2_t val[2];
143 } uint64x2x2_t;
144 
145 typedef struct float16x4x2_t {
146  float16x4_t val[2];
147 } float16x4x2_t;
148 
149 typedef struct float16x8x2_t {
150  float16x8_t val[2];
151 } float16x8x2_t;
152 
153 typedef struct float32x2x2_t {
154  float32x2_t val[2];
155 } float32x2x2_t;
156 
157 typedef struct float32x4x2_t {
158  float32x4_t val[2];
159 } float32x4x2_t;
160 
161 #ifdef __aarch64__
162 typedef struct float64x1x2_t {
163  float64x1_t val[2];
164 } float64x1x2_t;
165 
166 typedef struct float64x2x2_t {
167  float64x2_t val[2];
168 } float64x2x2_t;
169 
170 #endif
171 typedef struct poly8x8x2_t {
172  poly8x8_t val[2];
173 } poly8x8x2_t;
174 
175 typedef struct poly8x16x2_t {
176  poly8x16_t val[2];
177 } poly8x16x2_t;
178 
179 typedef struct poly16x4x2_t {
180  poly16x4_t val[2];
181 } poly16x4x2_t;
182 
183 typedef struct poly16x8x2_t {
184  poly16x8_t val[2];
185 } poly16x8x2_t;
186 
187 #ifdef __aarch64__
188 typedef struct poly64x1x2_t {
189  poly64x1_t val[2];
190 } poly64x1x2_t;
191 
192 typedef struct poly64x2x2_t {
193  poly64x2_t val[2];
194 } poly64x2x2_t;
195 
196 #endif
197 typedef struct int8x8x3_t {
198  int8x8_t val[3];
199 } int8x8x3_t;
200 
201 typedef struct int8x16x3_t {
202  int8x16_t val[3];
203 } int8x16x3_t;
204 
205 typedef struct int16x4x3_t {
206  int16x4_t val[3];
207 } int16x4x3_t;
208 
209 typedef struct int16x8x3_t {
210  int16x8_t val[3];
211 } int16x8x3_t;
212 
213 typedef struct int32x2x3_t {
214  int32x2_t val[3];
215 } int32x2x3_t;
216 
217 typedef struct int32x4x3_t {
218  int32x4_t val[3];
219 } int32x4x3_t;
220 
221 typedef struct int64x1x3_t {
222  int64x1_t val[3];
223 } int64x1x3_t;
224 
225 typedef struct int64x2x3_t {
226  int64x2_t val[3];
227 } int64x2x3_t;
228 
229 typedef struct uint8x8x3_t {
230  uint8x8_t val[3];
231 } uint8x8x3_t;
232 
233 typedef struct uint8x16x3_t {
234  uint8x16_t val[3];
235 } uint8x16x3_t;
236 
237 typedef struct uint16x4x3_t {
238  uint16x4_t val[3];
239 } uint16x4x3_t;
240 
241 typedef struct uint16x8x3_t {
242  uint16x8_t val[3];
243 } uint16x8x3_t;
244 
245 typedef struct uint32x2x3_t {
246  uint32x2_t val[3];
247 } uint32x2x3_t;
248 
249 typedef struct uint32x4x3_t {
250  uint32x4_t val[3];
251 } uint32x4x3_t;
252 
253 typedef struct uint64x1x3_t {
254  uint64x1_t val[3];
255 } uint64x1x3_t;
256 
257 typedef struct uint64x2x3_t {
258  uint64x2_t val[3];
259 } uint64x2x3_t;
260 
261 typedef struct float16x4x3_t {
262  float16x4_t val[3];
263 } float16x4x3_t;
264 
265 typedef struct float16x8x3_t {
266  float16x8_t val[3];
267 } float16x8x3_t;
268 
269 typedef struct float32x2x3_t {
270  float32x2_t val[3];
271 } float32x2x3_t;
272 
273 typedef struct float32x4x3_t {
274  float32x4_t val[3];
275 } float32x4x3_t;
276 
277 #ifdef __aarch64__
278 typedef struct float64x1x3_t {
279  float64x1_t val[3];
280 } float64x1x3_t;
281 
282 typedef struct float64x2x3_t {
283  float64x2_t val[3];
284 } float64x2x3_t;
285 
286 #endif
287 typedef struct poly8x8x3_t {
288  poly8x8_t val[3];
289 } poly8x8x3_t;
290 
291 typedef struct poly8x16x3_t {
292  poly8x16_t val[3];
293 } poly8x16x3_t;
294 
295 typedef struct poly16x4x3_t {
296  poly16x4_t val[3];
297 } poly16x4x3_t;
298 
299 typedef struct poly16x8x3_t {
300  poly16x8_t val[3];
301 } poly16x8x3_t;
302 
303 #ifdef __aarch64__
304 typedef struct poly64x1x3_t {
305  poly64x1_t val[3];
306 } poly64x1x3_t;
307 
308 typedef struct poly64x2x3_t {
309  poly64x2_t val[3];
310 } poly64x2x3_t;
311 
312 #endif
313 typedef struct int8x8x4_t {
314  int8x8_t val[4];
315 } int8x8x4_t;
316 
317 typedef struct int8x16x4_t {
318  int8x16_t val[4];
319 } int8x16x4_t;
320 
321 typedef struct int16x4x4_t {
322  int16x4_t val[4];
323 } int16x4x4_t;
324 
325 typedef struct int16x8x4_t {
326  int16x8_t val[4];
327 } int16x8x4_t;
328 
329 typedef struct int32x2x4_t {
330  int32x2_t val[4];
331 } int32x2x4_t;
332 
333 typedef struct int32x4x4_t {
334  int32x4_t val[4];
335 } int32x4x4_t;
336 
337 typedef struct int64x1x4_t {
338  int64x1_t val[4];
339 } int64x1x4_t;
340 
341 typedef struct int64x2x4_t {
342  int64x2_t val[4];
343 } int64x2x4_t;
344 
345 typedef struct uint8x8x4_t {
346  uint8x8_t val[4];
347 } uint8x8x4_t;
348 
349 typedef struct uint8x16x4_t {
350  uint8x16_t val[4];
351 } uint8x16x4_t;
352 
353 typedef struct uint16x4x4_t {
354  uint16x4_t val[4];
355 } uint16x4x4_t;
356 
357 typedef struct uint16x8x4_t {
358  uint16x8_t val[4];
359 } uint16x8x4_t;
360 
361 typedef struct uint32x2x4_t {
362  uint32x2_t val[4];
363 } uint32x2x4_t;
364 
365 typedef struct uint32x4x4_t {
366  uint32x4_t val[4];
367 } uint32x4x4_t;
368 
369 typedef struct uint64x1x4_t {
370  uint64x1_t val[4];
371 } uint64x1x4_t;
372 
373 typedef struct uint64x2x4_t {
374  uint64x2_t val[4];
375 } uint64x2x4_t;
376 
377 typedef struct float16x4x4_t {
378  float16x4_t val[4];
379 } float16x4x4_t;
380 
381 typedef struct float16x8x4_t {
382  float16x8_t val[4];
383 } float16x8x4_t;
384 
385 typedef struct float32x2x4_t {
386  float32x2_t val[4];
387 } float32x2x4_t;
388 
389 typedef struct float32x4x4_t {
390  float32x4_t val[4];
391 } float32x4x4_t;
392 
393 #ifdef __aarch64__
394 typedef struct float64x1x4_t {
395  float64x1_t val[4];
396 } float64x1x4_t;
397 
398 typedef struct float64x2x4_t {
399  float64x2_t val[4];
400 } float64x2x4_t;
401 
402 #endif
403 typedef struct poly8x8x4_t {
404  poly8x8_t val[4];
405 } poly8x8x4_t;
406 
407 typedef struct poly8x16x4_t {
408  poly8x16_t val[4];
409 } poly8x16x4_t;
410 
411 typedef struct poly16x4x4_t {
412  poly16x4_t val[4];
413 } poly16x4x4_t;
414 
415 typedef struct poly16x8x4_t {
416  poly16x8_t val[4];
417 } poly16x8x4_t;
418 
419 #ifdef __aarch64__
420 typedef struct poly64x1x4_t {
421  poly64x1_t val[4];
422 } poly64x1x4_t;
423 
424 typedef struct poly64x2x4_t {
425  poly64x2_t val[4];
426 } poly64x2x4_t;
427 
428 #endif
429 
430 #define __ai static inline __attribute__((__always_inline__, __nodebug__))
431 
432 #ifdef __LITTLE_ENDIAN__
433 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
434  uint8x16_t __ret;
435  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
436  return __ret;
437 }
438 #else
439 __ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
440  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
441  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
442  uint8x16_t __ret;
443  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
444  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
445  return __ret;
446 }
447 __ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
448  uint8x16_t __ret;
449  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
450  return __ret;
451 }
452 #endif
453 
454 #ifdef __LITTLE_ENDIAN__
455 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
456  uint32x4_t __ret;
457  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
458  return __ret;
459 }
460 #else
461 __ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
462  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
463  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
464  uint32x4_t __ret;
465  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
467  return __ret;
468 }
469 __ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
470  uint32x4_t __ret;
471  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
472  return __ret;
473 }
474 #endif
475 
476 #ifdef __LITTLE_ENDIAN__
477 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
478  uint16x8_t __ret;
479  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
480  return __ret;
481 }
482 #else
483 __ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
484  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
485  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
486  uint16x8_t __ret;
487  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
488  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
489  return __ret;
490 }
491 __ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
492  uint16x8_t __ret;
493  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
494  return __ret;
495 }
496 #endif
497 
498 #ifdef __LITTLE_ENDIAN__
499 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
500  int8x16_t __ret;
501  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
502  return __ret;
503 }
504 #else
505 __ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
506  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
507  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
508  int8x16_t __ret;
509  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
511  return __ret;
512 }
513 __ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
514  int8x16_t __ret;
515  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
516  return __ret;
517 }
518 #endif
519 
520 #ifdef __LITTLE_ENDIAN__
521 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
522  float32x4_t __ret;
523  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
524  return __ret;
525 }
526 #else
527 __ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
528  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
529  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
530  float32x4_t __ret;
531  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
533  return __ret;
534 }
535 #endif
536 
537 #ifdef __LITTLE_ENDIAN__
538 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
539  int32x4_t __ret;
540  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
541  return __ret;
542 }
543 #else
544 __ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
545  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
546  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
547  int32x4_t __ret;
548  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
549  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
550  return __ret;
551 }
552 __ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
553  int32x4_t __ret;
554  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
555  return __ret;
556 }
557 #endif
558 
559 #ifdef __LITTLE_ENDIAN__
560 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
561  int16x8_t __ret;
562  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
563  return __ret;
564 }
565 #else
566 __ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
567  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
568  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
569  int16x8_t __ret;
570  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
571  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
572  return __ret;
573 }
574 __ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
575  int16x8_t __ret;
576  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
577  return __ret;
578 }
579 #endif
580 
581 #ifdef __LITTLE_ENDIAN__
582 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
583  uint8x8_t __ret;
584  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
585  return __ret;
586 }
587 #else
588 __ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
589  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
590  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
591  uint8x8_t __ret;
592  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
593  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
594  return __ret;
595 }
596 __ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
597  uint8x8_t __ret;
598  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
599  return __ret;
600 }
601 #endif
602 
603 #ifdef __LITTLE_ENDIAN__
604 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
605  uint32x2_t __ret;
606  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
607  return __ret;
608 }
609 #else
610 __ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
611  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
612  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
613  uint32x2_t __ret;
614  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
616  return __ret;
617 }
618 __ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
619  uint32x2_t __ret;
620  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
621  return __ret;
622 }
623 #endif
624 
625 #ifdef __LITTLE_ENDIAN__
626 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
627  uint16x4_t __ret;
628  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
629  return __ret;
630 }
631 #else
632 __ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
633  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
634  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
635  uint16x4_t __ret;
636  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
638  return __ret;
639 }
640 __ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
641  uint16x4_t __ret;
642  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
643  return __ret;
644 }
645 #endif
646 
647 #ifdef __LITTLE_ENDIAN__
648 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
649  int8x8_t __ret;
650  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
651  return __ret;
652 }
653 #else
654 __ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
655  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
656  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
657  int8x8_t __ret;
658  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
659  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
660  return __ret;
661 }
662 __ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
663  int8x8_t __ret;
664  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
665  return __ret;
666 }
667 #endif
668 
669 #ifdef __LITTLE_ENDIAN__
670 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
671  float32x2_t __ret;
672  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
673  return __ret;
674 }
675 #else
676 __ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
677  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
678  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
679  float32x2_t __ret;
680  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
682  return __ret;
683 }
684 #endif
685 
686 #ifdef __LITTLE_ENDIAN__
687 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
688  int32x2_t __ret;
689  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
690  return __ret;
691 }
692 #else
693 __ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
694  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
695  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
696  int32x2_t __ret;
697  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
699  return __ret;
700 }
701 __ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
702  int32x2_t __ret;
703  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
704  return __ret;
705 }
706 #endif
707 
708 #ifdef __LITTLE_ENDIAN__
709 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
710  int16x4_t __ret;
711  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
712  return __ret;
713 }
714 #else
715 __ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
716  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
717  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
718  int16x4_t __ret;
719  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
721  return __ret;
722 }
723 __ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
724  int16x4_t __ret;
725  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
726  return __ret;
727 }
728 #endif
729 
730 #ifdef __LITTLE_ENDIAN__
731 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
732  int8x16_t __ret;
733  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
734  return __ret;
735 }
736 #else
737 __ai int8x16_t vabsq_s8(int8x16_t __p0) {
738  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
739  int8x16_t __ret;
740  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
742  return __ret;
743 }
744 #endif
745 
746 #ifdef __LITTLE_ENDIAN__
747 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
748  float32x4_t __ret;
749  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
750  return __ret;
751 }
752 #else
753 __ai float32x4_t vabsq_f32(float32x4_t __p0) {
754  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
755  float32x4_t __ret;
756  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
757  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
758  return __ret;
759 }
760 #endif
761 
762 #ifdef __LITTLE_ENDIAN__
763 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
764  int32x4_t __ret;
765  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
766  return __ret;
767 }
768 #else
769 __ai int32x4_t vabsq_s32(int32x4_t __p0) {
770  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
771  int32x4_t __ret;
772  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
774  return __ret;
775 }
776 #endif
777 
778 #ifdef __LITTLE_ENDIAN__
779 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
780  int16x8_t __ret;
781  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
782  return __ret;
783 }
784 #else
785 __ai int16x8_t vabsq_s16(int16x8_t __p0) {
786  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
787  int16x8_t __ret;
788  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
790  return __ret;
791 }
792 #endif
793 
794 #ifdef __LITTLE_ENDIAN__
795 __ai int8x8_t vabs_s8(int8x8_t __p0) {
796  int8x8_t __ret;
797  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
798  return __ret;
799 }
800 #else
801 __ai int8x8_t vabs_s8(int8x8_t __p0) {
802  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
803  int8x8_t __ret;
804  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
806  return __ret;
807 }
808 #endif
809 
810 #ifdef __LITTLE_ENDIAN__
811 __ai float32x2_t vabs_f32(float32x2_t __p0) {
812  float32x2_t __ret;
813  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
814  return __ret;
815 }
816 #else
817 __ai float32x2_t vabs_f32(float32x2_t __p0) {
818  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
819  float32x2_t __ret;
820  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
822  return __ret;
823 }
824 #endif
825 
826 #ifdef __LITTLE_ENDIAN__
827 __ai int32x2_t vabs_s32(int32x2_t __p0) {
828  int32x2_t __ret;
829  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
830  return __ret;
831 }
832 #else
833 __ai int32x2_t vabs_s32(int32x2_t __p0) {
834  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
835  int32x2_t __ret;
836  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
838  return __ret;
839 }
840 #endif
841 
842 #ifdef __LITTLE_ENDIAN__
843 __ai int16x4_t vabs_s16(int16x4_t __p0) {
844  int16x4_t __ret;
845  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
846  return __ret;
847 }
848 #else
849 __ai int16x4_t vabs_s16(int16x4_t __p0) {
850  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
851  int16x4_t __ret;
852  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
854  return __ret;
855 }
856 #endif
857 
858 #ifdef __LITTLE_ENDIAN__
859 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
860  uint8x16_t __ret;
861  __ret = __p0 + __p1;
862  return __ret;
863 }
864 #else
865 __ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
866  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
867  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
868  uint8x16_t __ret;
869  __ret = __rev0 + __rev1;
870  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
871  return __ret;
872 }
873 #endif
874 
875 #ifdef __LITTLE_ENDIAN__
876 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
877  uint32x4_t __ret;
878  __ret = __p0 + __p1;
879  return __ret;
880 }
881 #else
882 __ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
883  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
884  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
885  uint32x4_t __ret;
886  __ret = __rev0 + __rev1;
887  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
888  return __ret;
889 }
890 #endif
891 
892 #ifdef __LITTLE_ENDIAN__
893 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
894  uint64x2_t __ret;
895  __ret = __p0 + __p1;
896  return __ret;
897 }
898 #else
899 __ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
900  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
901  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
902  uint64x2_t __ret;
903  __ret = __rev0 + __rev1;
904  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
905  return __ret;
906 }
907 #endif
908 
909 #ifdef __LITTLE_ENDIAN__
910 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
911  uint16x8_t __ret;
912  __ret = __p0 + __p1;
913  return __ret;
914 }
915 #else
916 __ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
917  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
918  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
919  uint16x8_t __ret;
920  __ret = __rev0 + __rev1;
921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
922  return __ret;
923 }
924 #endif
925 
926 #ifdef __LITTLE_ENDIAN__
927 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
928  int8x16_t __ret;
929  __ret = __p0 + __p1;
930  return __ret;
931 }
932 #else
933 __ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
934  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
935  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
936  int8x16_t __ret;
937  __ret = __rev0 + __rev1;
938  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
939  return __ret;
940 }
941 #endif
942 
943 #ifdef __LITTLE_ENDIAN__
944 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
945  float32x4_t __ret;
946  __ret = __p0 + __p1;
947  return __ret;
948 }
949 #else
950 __ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
951  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
952  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
953  float32x4_t __ret;
954  __ret = __rev0 + __rev1;
955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
956  return __ret;
957 }
958 #endif
959 
960 #ifdef __LITTLE_ENDIAN__
961 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
962  int32x4_t __ret;
963  __ret = __p0 + __p1;
964  return __ret;
965 }
966 #else
967 __ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
968  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
969  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
970  int32x4_t __ret;
971  __ret = __rev0 + __rev1;
972  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
973  return __ret;
974 }
975 #endif
976 
977 #ifdef __LITTLE_ENDIAN__
978 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
979  int64x2_t __ret;
980  __ret = __p0 + __p1;
981  return __ret;
982 }
983 #else
984 __ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
985  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
986  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
987  int64x2_t __ret;
988  __ret = __rev0 + __rev1;
989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
990  return __ret;
991 }
992 #endif
993 
994 #ifdef __LITTLE_ENDIAN__
995 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
996  int16x8_t __ret;
997  __ret = __p0 + __p1;
998  return __ret;
999 }
1000 #else
1001 __ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1002  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1003  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1004  int16x8_t __ret;
1005  __ret = __rev0 + __rev1;
1006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1007  return __ret;
1008 }
1009 #endif
1010 
1011 #ifdef __LITTLE_ENDIAN__
1012 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1013  uint8x8_t __ret;
1014  __ret = __p0 + __p1;
1015  return __ret;
1016 }
1017 #else
1018 __ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1019  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1020  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1021  uint8x8_t __ret;
1022  __ret = __rev0 + __rev1;
1023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1024  return __ret;
1025 }
1026 #endif
1027 
1028 #ifdef __LITTLE_ENDIAN__
1029 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1030  uint32x2_t __ret;
1031  __ret = __p0 + __p1;
1032  return __ret;
1033 }
1034 #else
1035 __ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1036  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1037  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1038  uint32x2_t __ret;
1039  __ret = __rev0 + __rev1;
1040  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1041  return __ret;
1042 }
1043 #endif
1044 
1045 #ifdef __LITTLE_ENDIAN__
1046 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1047  uint64x1_t __ret;
1048  __ret = __p0 + __p1;
1049  return __ret;
1050 }
1051 #else
1052 __ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1053  uint64x1_t __ret;
1054  __ret = __p0 + __p1;
1055  return __ret;
1056 }
1057 #endif
1058 
1059 #ifdef __LITTLE_ENDIAN__
1060 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1061  uint16x4_t __ret;
1062  __ret = __p0 + __p1;
1063  return __ret;
1064 }
1065 #else
1066 __ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1067  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1068  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1069  uint16x4_t __ret;
1070  __ret = __rev0 + __rev1;
1071  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1072  return __ret;
1073 }
1074 #endif
1075 
1076 #ifdef __LITTLE_ENDIAN__
1077 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1078  int8x8_t __ret;
1079  __ret = __p0 + __p1;
1080  return __ret;
1081 }
1082 #else
1083 __ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1084  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1085  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1086  int8x8_t __ret;
1087  __ret = __rev0 + __rev1;
1088  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1089  return __ret;
1090 }
1091 #endif
1092 
1093 #ifdef __LITTLE_ENDIAN__
1094 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1095  float32x2_t __ret;
1096  __ret = __p0 + __p1;
1097  return __ret;
1098 }
1099 #else
1100 __ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1101  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1102  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1103  float32x2_t __ret;
1104  __ret = __rev0 + __rev1;
1105  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1106  return __ret;
1107 }
1108 #endif
1109 
1110 #ifdef __LITTLE_ENDIAN__
1111 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1112  int32x2_t __ret;
1113  __ret = __p0 + __p1;
1114  return __ret;
1115 }
1116 #else
1117 __ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1118  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1119  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1120  int32x2_t __ret;
1121  __ret = __rev0 + __rev1;
1122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1123  return __ret;
1124 }
1125 #endif
1126 
1127 #ifdef __LITTLE_ENDIAN__
1128 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1129  int64x1_t __ret;
1130  __ret = __p0 + __p1;
1131  return __ret;
1132 }
1133 #else
1134 __ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1135  int64x1_t __ret;
1136  __ret = __p0 + __p1;
1137  return __ret;
1138 }
1139 #endif
1140 
1141 #ifdef __LITTLE_ENDIAN__
1142 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1143  int16x4_t __ret;
1144  __ret = __p0 + __p1;
1145  return __ret;
1146 }
1147 #else
1148 __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1149  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1150  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1151  int16x4_t __ret;
1152  __ret = __rev0 + __rev1;
1153  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1154  return __ret;
1155 }
1156 #endif
1157 
1158 #ifdef __LITTLE_ENDIAN__
1159 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1160  uint16x4_t __ret;
1161  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1162  return __ret;
1163 }
1164 #else
1165 __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1166  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1167  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1168  uint16x4_t __ret;
1169  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
1170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1171  return __ret;
1172 }
1173 __ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1174  uint16x4_t __ret;
1175  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1176  return __ret;
1177 }
1178 #endif
1179 
1180 #ifdef __LITTLE_ENDIAN__
1181 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1182  uint32x2_t __ret;
1183  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1184  return __ret;
1185 }
1186 #else
1187 __ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1188  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1189  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1190  uint32x2_t __ret;
1191  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
1192  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1193  return __ret;
1194 }
1195 __ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1196  uint32x2_t __ret;
1197  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1198  return __ret;
1199 }
1200 #endif
1201 
1202 #ifdef __LITTLE_ENDIAN__
1203 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1204  uint8x8_t __ret;
1205  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1206  return __ret;
1207 }
1208 #else
1209 __ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1210  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1211  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1212  uint8x8_t __ret;
1213  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
1214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1215  return __ret;
1216 }
1217 __ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1218  uint8x8_t __ret;
1219  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1220  return __ret;
1221 }
1222 #endif
1223 
1224 #ifdef __LITTLE_ENDIAN__
1225 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1226  int16x4_t __ret;
1227  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1228  return __ret;
1229 }
1230 #else
1231 __ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1232  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1233  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1234  int16x4_t __ret;
1235  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
1236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1237  return __ret;
1238 }
1239 __ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1240  int16x4_t __ret;
1241  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1242  return __ret;
1243 }
1244 #endif
1245 
1246 #ifdef __LITTLE_ENDIAN__
1247 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1248  int32x2_t __ret;
1249  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1250  return __ret;
1251 }
1252 #else
1253 __ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1254  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1255  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1256  int32x2_t __ret;
1257  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
1258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1259  return __ret;
1260 }
1261 __ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1262  int32x2_t __ret;
1263  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1264  return __ret;
1265 }
1266 #endif
1267 
1268 #ifdef __LITTLE_ENDIAN__
1269 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1270  int8x8_t __ret;
1271  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1272  return __ret;
1273 }
1274 #else
1275 __ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1276  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1277  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1278  int8x8_t __ret;
1279  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
1280  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1281  return __ret;
1282 }
1283 __ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1284  int8x8_t __ret;
1285  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1286  return __ret;
1287 }
1288 #endif
1289 
1290 #ifdef __LITTLE_ENDIAN__
1291 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1292  uint8x16_t __ret;
1293  __ret = __p0 & __p1;
1294  return __ret;
1295 }
1296 #else
1297 __ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1298  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1299  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1300  uint8x16_t __ret;
1301  __ret = __rev0 & __rev1;
1302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1303  return __ret;
1304 }
1305 #endif
1306 
1307 #ifdef __LITTLE_ENDIAN__
1308 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1309  uint32x4_t __ret;
1310  __ret = __p0 & __p1;
1311  return __ret;
1312 }
1313 #else
1314 __ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1315  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1316  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1317  uint32x4_t __ret;
1318  __ret = __rev0 & __rev1;
1319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1320  return __ret;
1321 }
1322 #endif
1323 
1324 #ifdef __LITTLE_ENDIAN__
1325 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1326  uint64x2_t __ret;
1327  __ret = __p0 & __p1;
1328  return __ret;
1329 }
1330 #else
1331 __ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1332  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1333  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1334  uint64x2_t __ret;
1335  __ret = __rev0 & __rev1;
1336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1337  return __ret;
1338 }
1339 #endif
1340 
1341 #ifdef __LITTLE_ENDIAN__
1342 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1343  uint16x8_t __ret;
1344  __ret = __p0 & __p1;
1345  return __ret;
1346 }
1347 #else
1348 __ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1349  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1350  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1351  uint16x8_t __ret;
1352  __ret = __rev0 & __rev1;
1353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1354  return __ret;
1355 }
1356 #endif
1357 
1358 #ifdef __LITTLE_ENDIAN__
1359 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1360  int8x16_t __ret;
1361  __ret = __p0 & __p1;
1362  return __ret;
1363 }
1364 #else
1365 __ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1366  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1367  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1368  int8x16_t __ret;
1369  __ret = __rev0 & __rev1;
1370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1371  return __ret;
1372 }
1373 #endif
1374 
1375 #ifdef __LITTLE_ENDIAN__
1376 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1377  int32x4_t __ret;
1378  __ret = __p0 & __p1;
1379  return __ret;
1380 }
1381 #else
1382 __ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1383  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1384  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1385  int32x4_t __ret;
1386  __ret = __rev0 & __rev1;
1387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1388  return __ret;
1389 }
1390 #endif
1391 
1392 #ifdef __LITTLE_ENDIAN__
1393 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1394  int64x2_t __ret;
1395  __ret = __p0 & __p1;
1396  return __ret;
1397 }
1398 #else
1399 __ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1400  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1401  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1402  int64x2_t __ret;
1403  __ret = __rev0 & __rev1;
1404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1405  return __ret;
1406 }
1407 #endif
1408 
1409 #ifdef __LITTLE_ENDIAN__
1410 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1411  int16x8_t __ret;
1412  __ret = __p0 & __p1;
1413  return __ret;
1414 }
1415 #else
1416 __ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1417  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1418  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1419  int16x8_t __ret;
1420  __ret = __rev0 & __rev1;
1421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1422  return __ret;
1423 }
1424 #endif
1425 
1426 #ifdef __LITTLE_ENDIAN__
1427 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1428  uint8x8_t __ret;
1429  __ret = __p0 & __p1;
1430  return __ret;
1431 }
1432 #else
1433 __ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1434  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1435  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1436  uint8x8_t __ret;
1437  __ret = __rev0 & __rev1;
1438  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1439  return __ret;
1440 }
1441 #endif
1442 
1443 #ifdef __LITTLE_ENDIAN__
1444 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1445  uint32x2_t __ret;
1446  __ret = __p0 & __p1;
1447  return __ret;
1448 }
1449 #else
1450 __ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1451  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1452  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1453  uint32x2_t __ret;
1454  __ret = __rev0 & __rev1;
1455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1456  return __ret;
1457 }
1458 #endif
1459 
1460 #ifdef __LITTLE_ENDIAN__
1461 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1462  uint64x1_t __ret;
1463  __ret = __p0 & __p1;
1464  return __ret;
1465 }
1466 #else
1467 __ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1468  uint64x1_t __ret;
1469  __ret = __p0 & __p1;
1470  return __ret;
1471 }
1472 #endif
1473 
1474 #ifdef __LITTLE_ENDIAN__
1475 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1476  uint16x4_t __ret;
1477  __ret = __p0 & __p1;
1478  return __ret;
1479 }
1480 #else
1481 __ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1482  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1483  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1484  uint16x4_t __ret;
1485  __ret = __rev0 & __rev1;
1486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1487  return __ret;
1488 }
1489 #endif
1490 
1491 #ifdef __LITTLE_ENDIAN__
1492 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1493  int8x8_t __ret;
1494  __ret = __p0 & __p1;
1495  return __ret;
1496 }
1497 #else
1498 __ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1499  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1500  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1501  int8x8_t __ret;
1502  __ret = __rev0 & __rev1;
1503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1504  return __ret;
1505 }
1506 #endif
1507 
1508 #ifdef __LITTLE_ENDIAN__
1509 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1510  int32x2_t __ret;
1511  __ret = __p0 & __p1;
1512  return __ret;
1513 }
1514 #else
1515 __ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1516  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1517  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1518  int32x2_t __ret;
1519  __ret = __rev0 & __rev1;
1520  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1521  return __ret;
1522 }
1523 #endif
1524 
1525 #ifdef __LITTLE_ENDIAN__
1526 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1527  int64x1_t __ret;
1528  __ret = __p0 & __p1;
1529  return __ret;
1530 }
1531 #else
1532 __ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1533  int64x1_t __ret;
1534  __ret = __p0 & __p1;
1535  return __ret;
1536 }
1537 #endif
1538 
1539 #ifdef __LITTLE_ENDIAN__
1540 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1541  int16x4_t __ret;
1542  __ret = __p0 & __p1;
1543  return __ret;
1544 }
1545 #else
1546 __ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1547  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1548  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1549  int16x4_t __ret;
1550  __ret = __rev0 & __rev1;
1551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1552  return __ret;
1553 }
1554 #endif
1555 
1556 #ifdef __LITTLE_ENDIAN__
1557 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1558  uint8x16_t __ret;
1559  __ret = __p0 & ~__p1;
1560  return __ret;
1561 }
1562 #else
1563 __ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1564  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1565  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1566  uint8x16_t __ret;
1567  __ret = __rev0 & ~__rev1;
1568  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1569  return __ret;
1570 }
1571 #endif
1572 
1573 #ifdef __LITTLE_ENDIAN__
1574 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1575  uint32x4_t __ret;
1576  __ret = __p0 & ~__p1;
1577  return __ret;
1578 }
1579 #else
1580 __ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1581  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1582  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1583  uint32x4_t __ret;
1584  __ret = __rev0 & ~__rev1;
1585  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1586  return __ret;
1587 }
1588 #endif
1589 
1590 #ifdef __LITTLE_ENDIAN__
1591 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1592  uint64x2_t __ret;
1593  __ret = __p0 & ~__p1;
1594  return __ret;
1595 }
1596 #else
1597 __ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1598  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1599  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1600  uint64x2_t __ret;
1601  __ret = __rev0 & ~__rev1;
1602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1603  return __ret;
1604 }
1605 #endif
1606 
1607 #ifdef __LITTLE_ENDIAN__
1608 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1609  uint16x8_t __ret;
1610  __ret = __p0 & ~__p1;
1611  return __ret;
1612 }
1613 #else
1614 __ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1615  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1616  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1617  uint16x8_t __ret;
1618  __ret = __rev0 & ~__rev1;
1619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1620  return __ret;
1621 }
1622 #endif
1623 
1624 #ifdef __LITTLE_ENDIAN__
1625 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1626  int8x16_t __ret;
1627  __ret = __p0 & ~__p1;
1628  return __ret;
1629 }
1630 #else
1631 __ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1632  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1633  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1634  int8x16_t __ret;
1635  __ret = __rev0 & ~__rev1;
1636  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1637  return __ret;
1638 }
1639 #endif
1640 
1641 #ifdef __LITTLE_ENDIAN__
1642 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1643  int32x4_t __ret;
1644  __ret = __p0 & ~__p1;
1645  return __ret;
1646 }
1647 #else
1648 __ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1649  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1650  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1651  int32x4_t __ret;
1652  __ret = __rev0 & ~__rev1;
1653  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1654  return __ret;
1655 }
1656 #endif
1657 
1658 #ifdef __LITTLE_ENDIAN__
1659 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1660  int64x2_t __ret;
1661  __ret = __p0 & ~__p1;
1662  return __ret;
1663 }
1664 #else
1665 __ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1666  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1667  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1668  int64x2_t __ret;
1669  __ret = __rev0 & ~__rev1;
1670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1671  return __ret;
1672 }
1673 #endif
1674 
1675 #ifdef __LITTLE_ENDIAN__
1676 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1677  int16x8_t __ret;
1678  __ret = __p0 & ~__p1;
1679  return __ret;
1680 }
1681 #else
1682 __ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1683  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1684  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1685  int16x8_t __ret;
1686  __ret = __rev0 & ~__rev1;
1687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1688  return __ret;
1689 }
1690 #endif
1691 
1692 #ifdef __LITTLE_ENDIAN__
1693 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1694  uint8x8_t __ret;
1695  __ret = __p0 & ~__p1;
1696  return __ret;
1697 }
1698 #else
1699 __ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1700  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1701  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1702  uint8x8_t __ret;
1703  __ret = __rev0 & ~__rev1;
1704  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1705  return __ret;
1706 }
1707 #endif
1708 
1709 #ifdef __LITTLE_ENDIAN__
1710 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1711  uint32x2_t __ret;
1712  __ret = __p0 & ~__p1;
1713  return __ret;
1714 }
1715 #else
1716 __ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1717  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1718  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1719  uint32x2_t __ret;
1720  __ret = __rev0 & ~__rev1;
1721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1722  return __ret;
1723 }
1724 #endif
1725 
1726 #ifdef __LITTLE_ENDIAN__
1727 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1728  uint64x1_t __ret;
1729  __ret = __p0 & ~__p1;
1730  return __ret;
1731 }
1732 #else
1733 __ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1734  uint64x1_t __ret;
1735  __ret = __p0 & ~__p1;
1736  return __ret;
1737 }
1738 #endif
1739 
1740 #ifdef __LITTLE_ENDIAN__
1741 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1742  uint16x4_t __ret;
1743  __ret = __p0 & ~__p1;
1744  return __ret;
1745 }
1746 #else
1747 __ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1748  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1749  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1750  uint16x4_t __ret;
1751  __ret = __rev0 & ~__rev1;
1752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1753  return __ret;
1754 }
1755 #endif
1756 
1757 #ifdef __LITTLE_ENDIAN__
1758 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1759  int8x8_t __ret;
1760  __ret = __p0 & ~__p1;
1761  return __ret;
1762 }
1763 #else
1764 __ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1765  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1766  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1767  int8x8_t __ret;
1768  __ret = __rev0 & ~__rev1;
1769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1770  return __ret;
1771 }
1772 #endif
1773 
1774 #ifdef __LITTLE_ENDIAN__
1775 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1776  int32x2_t __ret;
1777  __ret = __p0 & ~__p1;
1778  return __ret;
1779 }
1780 #else
1781 __ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1782  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1783  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1784  int32x2_t __ret;
1785  __ret = __rev0 & ~__rev1;
1786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1787  return __ret;
1788 }
1789 #endif
1790 
1791 #ifdef __LITTLE_ENDIAN__
1792 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1793  int64x1_t __ret;
1794  __ret = __p0 & ~__p1;
1795  return __ret;
1796 }
1797 #else
1798 __ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1799  int64x1_t __ret;
1800  __ret = __p0 & ~__p1;
1801  return __ret;
1802 }
1803 #endif
1804 
1805 #ifdef __LITTLE_ENDIAN__
1806 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1807  int16x4_t __ret;
1808  __ret = __p0 & ~__p1;
1809  return __ret;
1810 }
1811 #else
1812 __ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1813  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1814  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1815  int16x4_t __ret;
1816  __ret = __rev0 & ~__rev1;
1817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1818  return __ret;
1819 }
1820 #endif
1821 
1822 #ifdef __LITTLE_ENDIAN__
1823 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1824  poly8x8_t __ret;
1825  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
1826  return __ret;
1827 }
1828 #else
1829 __ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1830  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1831  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1832  poly8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1833  poly8x8_t __ret;
1834  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
1835  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1836  return __ret;
1837 }
1838 #endif
1839 
1840 #ifdef __LITTLE_ENDIAN__
1841 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1842  poly16x4_t __ret;
1843  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
1844  return __ret;
1845 }
1846 #else
1847 __ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1848  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1849  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1850  poly16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1851  poly16x4_t __ret;
1852  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
1853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1854  return __ret;
1855 }
1856 #endif
1857 
1858 #ifdef __LITTLE_ENDIAN__
1859 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1860  poly8x16_t __ret;
1861  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
1862  return __ret;
1863 }
1864 #else
1865 __ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1866  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1867  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1868  poly8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1869  poly8x16_t __ret;
1870  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
1871  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1872  return __ret;
1873 }
1874 #endif
1875 
1876 #ifdef __LITTLE_ENDIAN__
1877 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1878  poly16x8_t __ret;
1879  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
1880  return __ret;
1881 }
1882 #else
1883 __ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1884  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1885  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1886  poly16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1887  poly16x8_t __ret;
1888  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
1889  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1890  return __ret;
1891 }
1892 #endif
1893 
1894 #ifdef __LITTLE_ENDIAN__
1895 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1896  uint8x16_t __ret;
1897  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
1898  return __ret;
1899 }
1900 #else
1901 __ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1902  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1903  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1904  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1905  uint8x16_t __ret;
1906  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
1907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1908  return __ret;
1909 }
1910 #endif
1911 
1912 #ifdef __LITTLE_ENDIAN__
1913 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1914  uint32x4_t __ret;
1915  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
1916  return __ret;
1917 }
1918 #else
1919 __ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1920  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1921  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1922  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1923  uint32x4_t __ret;
1924  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
1925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1926  return __ret;
1927 }
1928 #endif
1929 
1930 #ifdef __LITTLE_ENDIAN__
1931 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1932  uint64x2_t __ret;
1933  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
1934  return __ret;
1935 }
1936 #else
1937 __ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1938  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1939  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1940  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
1941  uint64x2_t __ret;
1942  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
1943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1944  return __ret;
1945 }
1946 #endif
1947 
1948 #ifdef __LITTLE_ENDIAN__
1949 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1950  uint16x8_t __ret;
1951  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
1952  return __ret;
1953 }
1954 #else
1955 __ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1956  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1957  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1958  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1959  uint16x8_t __ret;
1960  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
1961  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1962  return __ret;
1963 }
1964 #endif
1965 
1966 #ifdef __LITTLE_ENDIAN__
1967 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1968  int8x16_t __ret;
1969  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
1970  return __ret;
1971 }
1972 #else
1973 __ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1974  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1975  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1976  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1977  int8x16_t __ret;
1978  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
1979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1980  return __ret;
1981 }
1982 #endif
1983 
1984 #ifdef __LITTLE_ENDIAN__
1985 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1986  float32x4_t __ret;
1987  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
1988  return __ret;
1989 }
1990 #else
1991 __ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1992  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1993  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1994  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1995  float32x4_t __ret;
1996  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
1997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1998  return __ret;
1999 }
2000 #endif
2001 
2002 #ifdef __LITTLE_ENDIAN__
2003 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2004  int32x4_t __ret;
2005  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
2006  return __ret;
2007 }
2008 #else
2009 __ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2010  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2011  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2012  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2013  int32x4_t __ret;
2014  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
2015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2016  return __ret;
2017 }
2018 #endif
2019 
2020 #ifdef __LITTLE_ENDIAN__
2021 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2022  int64x2_t __ret;
2023  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
2024  return __ret;
2025 }
2026 #else
2027 __ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2028  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2029  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2030  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2031  int64x2_t __ret;
2032  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
2033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2034  return __ret;
2035 }
2036 #endif
2037 
2038 #ifdef __LITTLE_ENDIAN__
2039 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2040  int16x8_t __ret;
2041  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
2042  return __ret;
2043 }
2044 #else
2045 __ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2046  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2047  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2048  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2049  int16x8_t __ret;
2050  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
2051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2052  return __ret;
2053 }
2054 #endif
2055 
2056 #ifdef __LITTLE_ENDIAN__
2057 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2058  uint8x8_t __ret;
2059  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
2060  return __ret;
2061 }
2062 #else
2063 __ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2064  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2065  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2066  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2067  uint8x8_t __ret;
2068  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
2069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2070  return __ret;
2071 }
2072 #endif
2073 
2074 #ifdef __LITTLE_ENDIAN__
2075 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2076  uint32x2_t __ret;
2077  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
2078  return __ret;
2079 }
2080 #else
2081 __ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2082  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2083  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2084  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2085  uint32x2_t __ret;
2086  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
2087  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2088  return __ret;
2089 }
2090 #endif
2091 
2092 #ifdef __LITTLE_ENDIAN__
2093 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2094  uint64x1_t __ret;
2095  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2096  return __ret;
2097 }
2098 #else
2099 __ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2100  uint64x1_t __ret;
2101  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2102  return __ret;
2103 }
2104 #endif
2105 
2106 #ifdef __LITTLE_ENDIAN__
2107 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2108  uint16x4_t __ret;
2109  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
2110  return __ret;
2111 }
2112 #else
2113 __ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2114  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2115  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2116  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2117  uint16x4_t __ret;
2118  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
2119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2120  return __ret;
2121 }
2122 #endif
2123 
2124 #ifdef __LITTLE_ENDIAN__
2125 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2126  int8x8_t __ret;
2127  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
2128  return __ret;
2129 }
2130 #else
2131 __ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2132  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2133  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2134  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2135  int8x8_t __ret;
2136  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
2137  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2138  return __ret;
2139 }
2140 #endif
2141 
2142 #ifdef __LITTLE_ENDIAN__
2143 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2144  float32x2_t __ret;
2145  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
2146  return __ret;
2147 }
2148 #else
2149 __ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2150  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2151  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2152  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2153  float32x2_t __ret;
2154  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
2155  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2156  return __ret;
2157 }
2158 #endif
2159 
2160 #ifdef __LITTLE_ENDIAN__
2161 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2162  int32x2_t __ret;
2163  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
2164  return __ret;
2165 }
2166 #else
2167 __ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2168  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2169  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2170  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2171  int32x2_t __ret;
2172  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
2173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2174  return __ret;
2175 }
2176 #endif
2177 
2178 #ifdef __LITTLE_ENDIAN__
2179 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2180  int64x1_t __ret;
2181  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2182  return __ret;
2183 }
2184 #else
2185 __ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2186  int64x1_t __ret;
2187  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2188  return __ret;
2189 }
2190 #endif
2191 
2192 #ifdef __LITTLE_ENDIAN__
2193 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2194  int16x4_t __ret;
2195  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
2196  return __ret;
2197 }
2198 #else
2199 __ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2200  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2201  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2202  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2203  int16x4_t __ret;
2204  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
2205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2206  return __ret;
2207 }
2208 #endif
2209 
2210 #ifdef __LITTLE_ENDIAN__
2211 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2212  uint32x4_t __ret;
2213  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2214  return __ret;
2215 }
2216 #else
2217 __ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2218  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2219  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2220  uint32x4_t __ret;
2221  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2223  return __ret;
2224 }
2225 #endif
2226 
2227 #ifdef __LITTLE_ENDIAN__
2228 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2229  uint32x2_t __ret;
2230  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2231  return __ret;
2232 }
2233 #else
2234 __ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2235  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2236  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2237  uint32x2_t __ret;
2238  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2240  return __ret;
2241 }
2242 #endif
2243 
2244 #ifdef __LITTLE_ENDIAN__
2245 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2246  uint32x4_t __ret;
2247  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2248  return __ret;
2249 }
2250 #else
2251 __ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2252  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2253  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2254  uint32x4_t __ret;
2255  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2257  return __ret;
2258 }
2259 #endif
2260 
2261 #ifdef __LITTLE_ENDIAN__
2262 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2263  uint32x2_t __ret;
2264  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2265  return __ret;
2266 }
2267 #else
2268 __ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2269  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2270  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2271  uint32x2_t __ret;
2272  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2273  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2274  return __ret;
2275 }
2276 #endif
2277 
2278 #ifdef __LITTLE_ENDIAN__
2279 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2280  uint32x4_t __ret;
2281  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2282  return __ret;
2283 }
2284 #else
2285 __ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2286  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2287  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2288  uint32x4_t __ret;
2289  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2290  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2291  return __ret;
2292 }
2293 #endif
2294 
2295 #ifdef __LITTLE_ENDIAN__
2296 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2297  uint32x2_t __ret;
2298  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2299  return __ret;
2300 }
2301 #else
2302 __ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2303  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2304  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2305  uint32x2_t __ret;
2306  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2308  return __ret;
2309 }
2310 #endif
2311 
2312 #ifdef __LITTLE_ENDIAN__
2313 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2314  uint32x4_t __ret;
2315  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2316  return __ret;
2317 }
2318 #else
2319 __ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2320  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2321  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2322  uint32x4_t __ret;
2323  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2325  return __ret;
2326 }
2327 #endif
2328 
2329 #ifdef __LITTLE_ENDIAN__
2330 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2331  uint32x2_t __ret;
2332  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2333  return __ret;
2334 }
2335 #else
2336 __ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2337  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2338  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2339  uint32x2_t __ret;
2340  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2342  return __ret;
2343 }
2344 #endif
2345 
2346 #ifdef __LITTLE_ENDIAN__
2347 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2348  uint8x8_t __ret;
2349  __ret = (uint8x8_t)(__p0 == __p1);
2350  return __ret;
2351 }
2352 #else
2353 __ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2354  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2355  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2356  uint8x8_t __ret;
2357  __ret = (uint8x8_t)(__rev0 == __rev1);
2358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2359  return __ret;
2360 }
2361 #endif
2362 
2363 #ifdef __LITTLE_ENDIAN__
2364 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2365  uint8x16_t __ret;
2366  __ret = (uint8x16_t)(__p0 == __p1);
2367  return __ret;
2368 }
2369 #else
2370 __ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2371  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2372  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2373  uint8x16_t __ret;
2374  __ret = (uint8x16_t)(__rev0 == __rev1);
2375  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2376  return __ret;
2377 }
2378 #endif
2379 
2380 #ifdef __LITTLE_ENDIAN__
2381 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2382  uint8x16_t __ret;
2383  __ret = (uint8x16_t)(__p0 == __p1);
2384  return __ret;
2385 }
2386 #else
2387 __ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2388  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2389  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2390  uint8x16_t __ret;
2391  __ret = (uint8x16_t)(__rev0 == __rev1);
2392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2393  return __ret;
2394 }
2395 #endif
2396 
2397 #ifdef __LITTLE_ENDIAN__
2398 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2399  uint32x4_t __ret;
2400  __ret = (uint32x4_t)(__p0 == __p1);
2401  return __ret;
2402 }
2403 #else
2404 __ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2405  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2406  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2407  uint32x4_t __ret;
2408  __ret = (uint32x4_t)(__rev0 == __rev1);
2409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2410  return __ret;
2411 }
2412 #endif
2413 
2414 #ifdef __LITTLE_ENDIAN__
2415 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2416  uint16x8_t __ret;
2417  __ret = (uint16x8_t)(__p0 == __p1);
2418  return __ret;
2419 }
2420 #else
2421 __ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2422  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2423  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2424  uint16x8_t __ret;
2425  __ret = (uint16x8_t)(__rev0 == __rev1);
2426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2427  return __ret;
2428 }
2429 #endif
2430 
2431 #ifdef __LITTLE_ENDIAN__
2432 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2433  uint8x16_t __ret;
2434  __ret = (uint8x16_t)(__p0 == __p1);
2435  return __ret;
2436 }
2437 #else
2438 __ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2439  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2440  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2441  uint8x16_t __ret;
2442  __ret = (uint8x16_t)(__rev0 == __rev1);
2443  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2444  return __ret;
2445 }
2446 #endif
2447 
2448 #ifdef __LITTLE_ENDIAN__
2449 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2450  uint32x4_t __ret;
2451  __ret = (uint32x4_t)(__p0 == __p1);
2452  return __ret;
2453 }
2454 #else
2455 __ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2456  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2457  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2458  uint32x4_t __ret;
2459  __ret = (uint32x4_t)(__rev0 == __rev1);
2460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2461  return __ret;
2462 }
2463 #endif
2464 
2465 #ifdef __LITTLE_ENDIAN__
2466 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2467  uint32x4_t __ret;
2468  __ret = (uint32x4_t)(__p0 == __p1);
2469  return __ret;
2470 }
2471 #else
2472 __ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2473  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2474  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2475  uint32x4_t __ret;
2476  __ret = (uint32x4_t)(__rev0 == __rev1);
2477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2478  return __ret;
2479 }
2480 #endif
2481 
2482 #ifdef __LITTLE_ENDIAN__
2483 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2484  uint16x8_t __ret;
2485  __ret = (uint16x8_t)(__p0 == __p1);
2486  return __ret;
2487 }
2488 #else
2489 __ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2490  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2491  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2492  uint16x8_t __ret;
2493  __ret = (uint16x8_t)(__rev0 == __rev1);
2494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2495  return __ret;
2496 }
2497 #endif
2498 
2499 #ifdef __LITTLE_ENDIAN__
2500 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2501  uint8x8_t __ret;
2502  __ret = (uint8x8_t)(__p0 == __p1);
2503  return __ret;
2504 }
2505 #else
2506 __ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2507  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2508  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2509  uint8x8_t __ret;
2510  __ret = (uint8x8_t)(__rev0 == __rev1);
2511  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2512  return __ret;
2513 }
2514 #endif
2515 
2516 #ifdef __LITTLE_ENDIAN__
2517 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2518  uint32x2_t __ret;
2519  __ret = (uint32x2_t)(__p0 == __p1);
2520  return __ret;
2521 }
2522 #else
2523 __ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2524  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2525  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2526  uint32x2_t __ret;
2527  __ret = (uint32x2_t)(__rev0 == __rev1);
2528  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2529  return __ret;
2530 }
2531 #endif
2532 
2533 #ifdef __LITTLE_ENDIAN__
2534 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2535  uint16x4_t __ret;
2536  __ret = (uint16x4_t)(__p0 == __p1);
2537  return __ret;
2538 }
2539 #else
2540 __ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2541  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2542  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2543  uint16x4_t __ret;
2544  __ret = (uint16x4_t)(__rev0 == __rev1);
2545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2546  return __ret;
2547 }
2548 #endif
2549 
2550 #ifdef __LITTLE_ENDIAN__
2551 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2552  uint8x8_t __ret;
2553  __ret = (uint8x8_t)(__p0 == __p1);
2554  return __ret;
2555 }
2556 #else
2557 __ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2558  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2559  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2560  uint8x8_t __ret;
2561  __ret = (uint8x8_t)(__rev0 == __rev1);
2562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2563  return __ret;
2564 }
2565 #endif
2566 
2567 #ifdef __LITTLE_ENDIAN__
2568 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2569  uint32x2_t __ret;
2570  __ret = (uint32x2_t)(__p0 == __p1);
2571  return __ret;
2572 }
2573 #else
2574 __ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2575  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2576  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2577  uint32x2_t __ret;
2578  __ret = (uint32x2_t)(__rev0 == __rev1);
2579  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2580  return __ret;
2581 }
2582 #endif
2583 
2584 #ifdef __LITTLE_ENDIAN__
2585 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2586  uint32x2_t __ret;
2587  __ret = (uint32x2_t)(__p0 == __p1);
2588  return __ret;
2589 }
2590 #else
2591 __ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2592  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2593  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2594  uint32x2_t __ret;
2595  __ret = (uint32x2_t)(__rev0 == __rev1);
2596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2597  return __ret;
2598 }
2599 #endif
2600 
2601 #ifdef __LITTLE_ENDIAN__
2602 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2603  uint16x4_t __ret;
2604  __ret = (uint16x4_t)(__p0 == __p1);
2605  return __ret;
2606 }
2607 #else
2608 __ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2609  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2610  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2611  uint16x4_t __ret;
2612  __ret = (uint16x4_t)(__rev0 == __rev1);
2613  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2614  return __ret;
2615 }
2616 #endif
2617 
2618 #ifdef __LITTLE_ENDIAN__
2619 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2620  uint8x16_t __ret;
2621  __ret = (uint8x16_t)(__p0 >= __p1);
2622  return __ret;
2623 }
2624 #else
2625 __ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2626  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2627  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2628  uint8x16_t __ret;
2629  __ret = (uint8x16_t)(__rev0 >= __rev1);
2630  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2631  return __ret;
2632 }
2633 #endif
2634 
2635 #ifdef __LITTLE_ENDIAN__
2636 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2637  uint32x4_t __ret;
2638  __ret = (uint32x4_t)(__p0 >= __p1);
2639  return __ret;
2640 }
2641 #else
2642 __ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2643  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2644  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2645  uint32x4_t __ret;
2646  __ret = (uint32x4_t)(__rev0 >= __rev1);
2647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2648  return __ret;
2649 }
2650 #endif
2651 
2652 #ifdef __LITTLE_ENDIAN__
2653 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2654  uint16x8_t __ret;
2655  __ret = (uint16x8_t)(__p0 >= __p1);
2656  return __ret;
2657 }
2658 #else
2659 __ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2660  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2661  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2662  uint16x8_t __ret;
2663  __ret = (uint16x8_t)(__rev0 >= __rev1);
2664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2665  return __ret;
2666 }
2667 #endif
2668 
2669 #ifdef __LITTLE_ENDIAN__
2670 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2671  uint8x16_t __ret;
2672  __ret = (uint8x16_t)(__p0 >= __p1);
2673  return __ret;
2674 }
2675 #else
2676 __ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2677  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2678  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2679  uint8x16_t __ret;
2680  __ret = (uint8x16_t)(__rev0 >= __rev1);
2681  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2682  return __ret;
2683 }
2684 #endif
2685 
2686 #ifdef __LITTLE_ENDIAN__
2687 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2688  uint32x4_t __ret;
2689  __ret = (uint32x4_t)(__p0 >= __p1);
2690  return __ret;
2691 }
2692 #else
2693 __ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2694  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2695  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2696  uint32x4_t __ret;
2697  __ret = (uint32x4_t)(__rev0 >= __rev1);
2698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2699  return __ret;
2700 }
2701 #endif
2702 
2703 #ifdef __LITTLE_ENDIAN__
2704 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2705  uint32x4_t __ret;
2706  __ret = (uint32x4_t)(__p0 >= __p1);
2707  return __ret;
2708 }
2709 #else
2710 __ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2711  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2712  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2713  uint32x4_t __ret;
2714  __ret = (uint32x4_t)(__rev0 >= __rev1);
2715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2716  return __ret;
2717 }
2718 #endif
2719 
2720 #ifdef __LITTLE_ENDIAN__
2721 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2722  uint16x8_t __ret;
2723  __ret = (uint16x8_t)(__p0 >= __p1);
2724  return __ret;
2725 }
2726 #else
2727 __ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2728  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2729  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2730  uint16x8_t __ret;
2731  __ret = (uint16x8_t)(__rev0 >= __rev1);
2732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2733  return __ret;
2734 }
2735 #endif
2736 
2737 #ifdef __LITTLE_ENDIAN__
2738 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2739  uint8x8_t __ret;
2740  __ret = (uint8x8_t)(__p0 >= __p1);
2741  return __ret;
2742 }
2743 #else
2744 __ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2745  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2746  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2747  uint8x8_t __ret;
2748  __ret = (uint8x8_t)(__rev0 >= __rev1);
2749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2750  return __ret;
2751 }
2752 #endif
2753 
2754 #ifdef __LITTLE_ENDIAN__
2755 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2756  uint32x2_t __ret;
2757  __ret = (uint32x2_t)(__p0 >= __p1);
2758  return __ret;
2759 }
2760 #else
2761 __ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2762  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2763  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2764  uint32x2_t __ret;
2765  __ret = (uint32x2_t)(__rev0 >= __rev1);
2766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2767  return __ret;
2768 }
2769 #endif
2770 
2771 #ifdef __LITTLE_ENDIAN__
2772 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2773  uint16x4_t __ret;
2774  __ret = (uint16x4_t)(__p0 >= __p1);
2775  return __ret;
2776 }
2777 #else
2778 __ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2779  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2780  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2781  uint16x4_t __ret;
2782  __ret = (uint16x4_t)(__rev0 >= __rev1);
2783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2784  return __ret;
2785 }
2786 #endif
2787 
2788 #ifdef __LITTLE_ENDIAN__
2789 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2790  uint8x8_t __ret;
2791  __ret = (uint8x8_t)(__p0 >= __p1);
2792  return __ret;
2793 }
2794 #else
2795 __ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2796  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2797  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2798  uint8x8_t __ret;
2799  __ret = (uint8x8_t)(__rev0 >= __rev1);
2800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2801  return __ret;
2802 }
2803 #endif
2804 
2805 #ifdef __LITTLE_ENDIAN__
2806 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2807  uint32x2_t __ret;
2808  __ret = (uint32x2_t)(__p0 >= __p1);
2809  return __ret;
2810 }
2811 #else
2812 __ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2813  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2814  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2815  uint32x2_t __ret;
2816  __ret = (uint32x2_t)(__rev0 >= __rev1);
2817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2818  return __ret;
2819 }
2820 #endif
2821 
2822 #ifdef __LITTLE_ENDIAN__
2823 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2824  uint32x2_t __ret;
2825  __ret = (uint32x2_t)(__p0 >= __p1);
2826  return __ret;
2827 }
2828 #else
2829 __ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2830  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2831  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2832  uint32x2_t __ret;
2833  __ret = (uint32x2_t)(__rev0 >= __rev1);
2834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2835  return __ret;
2836 }
2837 #endif
2838 
2839 #ifdef __LITTLE_ENDIAN__
2840 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2841  uint16x4_t __ret;
2842  __ret = (uint16x4_t)(__p0 >= __p1);
2843  return __ret;
2844 }
2845 #else
2846 __ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2847  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2848  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2849  uint16x4_t __ret;
2850  __ret = (uint16x4_t)(__rev0 >= __rev1);
2851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2852  return __ret;
2853 }
2854 #endif
2855 
2856 #ifdef __LITTLE_ENDIAN__
2857 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2858  uint8x16_t __ret;
2859  __ret = (uint8x16_t)(__p0 > __p1);
2860  return __ret;
2861 }
2862 #else
2863 __ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2864  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2865  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2866  uint8x16_t __ret;
2867  __ret = (uint8x16_t)(__rev0 > __rev1);
2868  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2869  return __ret;
2870 }
2871 #endif
2872 
2873 #ifdef __LITTLE_ENDIAN__
2874 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2875  uint32x4_t __ret;
2876  __ret = (uint32x4_t)(__p0 > __p1);
2877  return __ret;
2878 }
2879 #else
2880 __ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2881  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2882  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2883  uint32x4_t __ret;
2884  __ret = (uint32x4_t)(__rev0 > __rev1);
2885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2886  return __ret;
2887 }
2888 #endif
2889 
2890 #ifdef __LITTLE_ENDIAN__
2891 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2892  uint16x8_t __ret;
2893  __ret = (uint16x8_t)(__p0 > __p1);
2894  return __ret;
2895 }
2896 #else
2897 __ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2898  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2899  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2900  uint16x8_t __ret;
2901  __ret = (uint16x8_t)(__rev0 > __rev1);
2902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2903  return __ret;
2904 }
2905 #endif
2906 
2907 #ifdef __LITTLE_ENDIAN__
2908 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2909  uint8x16_t __ret;
2910  __ret = (uint8x16_t)(__p0 > __p1);
2911  return __ret;
2912 }
2913 #else
2914 __ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2915  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2916  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2917  uint8x16_t __ret;
2918  __ret = (uint8x16_t)(__rev0 > __rev1);
2919  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2920  return __ret;
2921 }
2922 #endif
2923 
2924 #ifdef __LITTLE_ENDIAN__
2925 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2926  uint32x4_t __ret;
2927  __ret = (uint32x4_t)(__p0 > __p1);
2928  return __ret;
2929 }
2930 #else
2931 __ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2932  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2933  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2934  uint32x4_t __ret;
2935  __ret = (uint32x4_t)(__rev0 > __rev1);
2936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2937  return __ret;
2938 }
2939 #endif
2940 
2941 #ifdef __LITTLE_ENDIAN__
2942 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2943  uint32x4_t __ret;
2944  __ret = (uint32x4_t)(__p0 > __p1);
2945  return __ret;
2946 }
2947 #else
2948 __ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2949  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2950  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2951  uint32x4_t __ret;
2952  __ret = (uint32x4_t)(__rev0 > __rev1);
2953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2954  return __ret;
2955 }
2956 #endif
2957 
2958 #ifdef __LITTLE_ENDIAN__
2959 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2960  uint16x8_t __ret;
2961  __ret = (uint16x8_t)(__p0 > __p1);
2962  return __ret;
2963 }
2964 #else
2965 __ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2966  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2967  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2968  uint16x8_t __ret;
2969  __ret = (uint16x8_t)(__rev0 > __rev1);
2970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2971  return __ret;
2972 }
2973 #endif
2974 
2975 #ifdef __LITTLE_ENDIAN__
2976 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2977  uint8x8_t __ret;
2978  __ret = (uint8x8_t)(__p0 > __p1);
2979  return __ret;
2980 }
2981 #else
2982 __ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2983  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2984  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2985  uint8x8_t __ret;
2986  __ret = (uint8x8_t)(__rev0 > __rev1);
2987  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2988  return __ret;
2989 }
2990 #endif
2991 
2992 #ifdef __LITTLE_ENDIAN__
2993 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
2994  uint32x2_t __ret;
2995  __ret = (uint32x2_t)(__p0 > __p1);
2996  return __ret;
2997 }
2998 #else
2999 __ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3000  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3001  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3002  uint32x2_t __ret;
3003  __ret = (uint32x2_t)(__rev0 > __rev1);
3004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3005  return __ret;
3006 }
3007 #endif
3008 
3009 #ifdef __LITTLE_ENDIAN__
3010 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3011  uint16x4_t __ret;
3012  __ret = (uint16x4_t)(__p0 > __p1);
3013  return __ret;
3014 }
3015 #else
3016 __ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3017  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3018  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3019  uint16x4_t __ret;
3020  __ret = (uint16x4_t)(__rev0 > __rev1);
3021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3022  return __ret;
3023 }
3024 #endif
3025 
3026 #ifdef __LITTLE_ENDIAN__
3027 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3028  uint8x8_t __ret;
3029  __ret = (uint8x8_t)(__p0 > __p1);
3030  return __ret;
3031 }
3032 #else
3033 __ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3034  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3035  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3036  uint8x8_t __ret;
3037  __ret = (uint8x8_t)(__rev0 > __rev1);
3038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3039  return __ret;
3040 }
3041 #endif
3042 
3043 #ifdef __LITTLE_ENDIAN__
3044 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3045  uint32x2_t __ret;
3046  __ret = (uint32x2_t)(__p0 > __p1);
3047  return __ret;
3048 }
3049 #else
3050 __ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3051  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3052  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3053  uint32x2_t __ret;
3054  __ret = (uint32x2_t)(__rev0 > __rev1);
3055  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3056  return __ret;
3057 }
3058 #endif
3059 
3060 #ifdef __LITTLE_ENDIAN__
3061 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3062  uint32x2_t __ret;
3063  __ret = (uint32x2_t)(__p0 > __p1);
3064  return __ret;
3065 }
3066 #else
3067 __ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3068  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3069  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3070  uint32x2_t __ret;
3071  __ret = (uint32x2_t)(__rev0 > __rev1);
3072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3073  return __ret;
3074 }
3075 #endif
3076 
3077 #ifdef __LITTLE_ENDIAN__
3078 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3079  uint16x4_t __ret;
3080  __ret = (uint16x4_t)(__p0 > __p1);
3081  return __ret;
3082 }
3083 #else
3084 __ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3085  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3086  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3087  uint16x4_t __ret;
3088  __ret = (uint16x4_t)(__rev0 > __rev1);
3089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3090  return __ret;
3091 }
3092 #endif
3093 
3094 #ifdef __LITTLE_ENDIAN__
3095 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3096  uint8x16_t __ret;
3097  __ret = (uint8x16_t)(__p0 <= __p1);
3098  return __ret;
3099 }
3100 #else
3101 __ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3102  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3103  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3104  uint8x16_t __ret;
3105  __ret = (uint8x16_t)(__rev0 <= __rev1);
3106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3107  return __ret;
3108 }
3109 #endif
3110 
3111 #ifdef __LITTLE_ENDIAN__
3112 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3113  uint32x4_t __ret;
3114  __ret = (uint32x4_t)(__p0 <= __p1);
3115  return __ret;
3116 }
3117 #else
3118 __ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3119  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3120  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3121  uint32x4_t __ret;
3122  __ret = (uint32x4_t)(__rev0 <= __rev1);
3123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3124  return __ret;
3125 }
3126 #endif
3127 
3128 #ifdef __LITTLE_ENDIAN__
3129 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3130  uint16x8_t __ret;
3131  __ret = (uint16x8_t)(__p0 <= __p1);
3132  return __ret;
3133 }
3134 #else
3135 __ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3136  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3137  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3138  uint16x8_t __ret;
3139  __ret = (uint16x8_t)(__rev0 <= __rev1);
3140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3141  return __ret;
3142 }
3143 #endif
3144 
3145 #ifdef __LITTLE_ENDIAN__
3146 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3147  uint8x16_t __ret;
3148  __ret = (uint8x16_t)(__p0 <= __p1);
3149  return __ret;
3150 }
3151 #else
3152 __ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3153  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3154  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3155  uint8x16_t __ret;
3156  __ret = (uint8x16_t)(__rev0 <= __rev1);
3157  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3158  return __ret;
3159 }
3160 #endif
3161 
3162 #ifdef __LITTLE_ENDIAN__
3163 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3164  uint32x4_t __ret;
3165  __ret = (uint32x4_t)(__p0 <= __p1);
3166  return __ret;
3167 }
3168 #else
3169 __ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3170  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3171  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3172  uint32x4_t __ret;
3173  __ret = (uint32x4_t)(__rev0 <= __rev1);
3174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3175  return __ret;
3176 }
3177 #endif
3178 
3179 #ifdef __LITTLE_ENDIAN__
3180 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3181  uint32x4_t __ret;
3182  __ret = (uint32x4_t)(__p0 <= __p1);
3183  return __ret;
3184 }
3185 #else
3186 __ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3187  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3188  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3189  uint32x4_t __ret;
3190  __ret = (uint32x4_t)(__rev0 <= __rev1);
3191  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3192  return __ret;
3193 }
3194 #endif
3195 
3196 #ifdef __LITTLE_ENDIAN__
3197 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3198  uint16x8_t __ret;
3199  __ret = (uint16x8_t)(__p0 <= __p1);
3200  return __ret;
3201 }
3202 #else
3203 __ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3204  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3205  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3206  uint16x8_t __ret;
3207  __ret = (uint16x8_t)(__rev0 <= __rev1);
3208  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3209  return __ret;
3210 }
3211 #endif
3212 
3213 #ifdef __LITTLE_ENDIAN__
3214 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3215  uint8x8_t __ret;
3216  __ret = (uint8x8_t)(__p0 <= __p1);
3217  return __ret;
3218 }
3219 #else
3220 __ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3221  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3222  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3223  uint8x8_t __ret;
3224  __ret = (uint8x8_t)(__rev0 <= __rev1);
3225  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3226  return __ret;
3227 }
3228 #endif
3229 
3230 #ifdef __LITTLE_ENDIAN__
3231 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3232  uint32x2_t __ret;
3233  __ret = (uint32x2_t)(__p0 <= __p1);
3234  return __ret;
3235 }
3236 #else
3237 __ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3238  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3239  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3240  uint32x2_t __ret;
3241  __ret = (uint32x2_t)(__rev0 <= __rev1);
3242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3243  return __ret;
3244 }
3245 #endif
3246 
3247 #ifdef __LITTLE_ENDIAN__
3248 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3249  uint16x4_t __ret;
3250  __ret = (uint16x4_t)(__p0 <= __p1);
3251  return __ret;
3252 }
3253 #else
3254 __ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3255  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3256  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3257  uint16x4_t __ret;
3258  __ret = (uint16x4_t)(__rev0 <= __rev1);
3259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3260  return __ret;
3261 }
3262 #endif
3263 
3264 #ifdef __LITTLE_ENDIAN__
3265 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3266  uint8x8_t __ret;
3267  __ret = (uint8x8_t)(__p0 <= __p1);
3268  return __ret;
3269 }
3270 #else
3271 __ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3272  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3273  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3274  uint8x8_t __ret;
3275  __ret = (uint8x8_t)(__rev0 <= __rev1);
3276  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3277  return __ret;
3278 }
3279 #endif
3280 
3281 #ifdef __LITTLE_ENDIAN__
3282 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3283  uint32x2_t __ret;
3284  __ret = (uint32x2_t)(__p0 <= __p1);
3285  return __ret;
3286 }
3287 #else
3288 __ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3289  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3290  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3291  uint32x2_t __ret;
3292  __ret = (uint32x2_t)(__rev0 <= __rev1);
3293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3294  return __ret;
3295 }
3296 #endif
3297 
3298 #ifdef __LITTLE_ENDIAN__
3299 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3300  uint32x2_t __ret;
3301  __ret = (uint32x2_t)(__p0 <= __p1);
3302  return __ret;
3303 }
3304 #else
3305 __ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3306  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3307  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3308  uint32x2_t __ret;
3309  __ret = (uint32x2_t)(__rev0 <= __rev1);
3310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3311  return __ret;
3312 }
3313 #endif
3314 
3315 #ifdef __LITTLE_ENDIAN__
3316 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3317  uint16x4_t __ret;
3318  __ret = (uint16x4_t)(__p0 <= __p1);
3319  return __ret;
3320 }
3321 #else
3322 __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3323  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3324  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3325  uint16x4_t __ret;
3326  __ret = (uint16x4_t)(__rev0 <= __rev1);
3327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3328  return __ret;
3329 }
3330 #endif
3331 
3332 #ifdef __LITTLE_ENDIAN__
3333 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
3334  int8x16_t __ret;
3335  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
3336  return __ret;
3337 }
3338 #else
3339 __ai int8x16_t vclsq_s8(int8x16_t __p0) {
3340  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3341  int8x16_t __ret;
3342  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
3343  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3344  return __ret;
3345 }
3346 #endif
3347 
3348 #ifdef __LITTLE_ENDIAN__
3349 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
3350  int32x4_t __ret;
3351  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
3352  return __ret;
3353 }
3354 #else
3355 __ai int32x4_t vclsq_s32(int32x4_t __p0) {
3356  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3357  int32x4_t __ret;
3358  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
3359  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3360  return __ret;
3361 }
3362 #endif
3363 
3364 #ifdef __LITTLE_ENDIAN__
3365 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
3366  int16x8_t __ret;
3367  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
3368  return __ret;
3369 }
3370 #else
3371 __ai int16x8_t vclsq_s16(int16x8_t __p0) {
3372  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3373  int16x8_t __ret;
3374  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
3375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3376  return __ret;
3377 }
3378 #endif
3379 
3380 #ifdef __LITTLE_ENDIAN__
3381 __ai int8x8_t vcls_s8(int8x8_t __p0) {
3382  int8x8_t __ret;
3383  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
3384  return __ret;
3385 }
3386 #else
3387 __ai int8x8_t vcls_s8(int8x8_t __p0) {
3388  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3389  int8x8_t __ret;
3390  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
3391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3392  return __ret;
3393 }
3394 #endif
3395 
3396 #ifdef __LITTLE_ENDIAN__
3397 __ai int32x2_t vcls_s32(int32x2_t __p0) {
3398  int32x2_t __ret;
3399  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
3400  return __ret;
3401 }
3402 #else
3403 __ai int32x2_t vcls_s32(int32x2_t __p0) {
3404  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3405  int32x2_t __ret;
3406  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
3407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3408  return __ret;
3409 }
3410 #endif
3411 
3412 #ifdef __LITTLE_ENDIAN__
3413 __ai int16x4_t vcls_s16(int16x4_t __p0) {
3414  int16x4_t __ret;
3415  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
3416  return __ret;
3417 }
3418 #else
3419 __ai int16x4_t vcls_s16(int16x4_t __p0) {
3420  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3421  int16x4_t __ret;
3422  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
3423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3424  return __ret;
3425 }
3426 #endif
3427 
3428 #ifdef __LITTLE_ENDIAN__
3429 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3430  uint8x16_t __ret;
3431  __ret = (uint8x16_t)(__p0 < __p1);
3432  return __ret;
3433 }
3434 #else
3435 __ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3436  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3437  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3438  uint8x16_t __ret;
3439  __ret = (uint8x16_t)(__rev0 < __rev1);
3440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3441  return __ret;
3442 }
3443 #endif
3444 
3445 #ifdef __LITTLE_ENDIAN__
3446 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3447  uint32x4_t __ret;
3448  __ret = (uint32x4_t)(__p0 < __p1);
3449  return __ret;
3450 }
3451 #else
3452 __ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3453  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3454  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3455  uint32x4_t __ret;
3456  __ret = (uint32x4_t)(__rev0 < __rev1);
3457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3458  return __ret;
3459 }
3460 #endif
3461 
3462 #ifdef __LITTLE_ENDIAN__
3463 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3464  uint16x8_t __ret;
3465  __ret = (uint16x8_t)(__p0 < __p1);
3466  return __ret;
3467 }
3468 #else
3469 __ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3470  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3471  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3472  uint16x8_t __ret;
3473  __ret = (uint16x8_t)(__rev0 < __rev1);
3474  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3475  return __ret;
3476 }
3477 #endif
3478 
3479 #ifdef __LITTLE_ENDIAN__
3480 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3481  uint8x16_t __ret;
3482  __ret = (uint8x16_t)(__p0 < __p1);
3483  return __ret;
3484 }
3485 #else
3486 __ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3487  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3488  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3489  uint8x16_t __ret;
3490  __ret = (uint8x16_t)(__rev0 < __rev1);
3491  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3492  return __ret;
3493 }
3494 #endif
3495 
3496 #ifdef __LITTLE_ENDIAN__
3497 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3498  uint32x4_t __ret;
3499  __ret = (uint32x4_t)(__p0 < __p1);
3500  return __ret;
3501 }
3502 #else
3503 __ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3504  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3505  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3506  uint32x4_t __ret;
3507  __ret = (uint32x4_t)(__rev0 < __rev1);
3508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3509  return __ret;
3510 }
3511 #endif
3512 
3513 #ifdef __LITTLE_ENDIAN__
3514 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3515  uint32x4_t __ret;
3516  __ret = (uint32x4_t)(__p0 < __p1);
3517  return __ret;
3518 }
3519 #else
3520 __ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3521  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3522  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3523  uint32x4_t __ret;
3524  __ret = (uint32x4_t)(__rev0 < __rev1);
3525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3526  return __ret;
3527 }
3528 #endif
3529 
3530 #ifdef __LITTLE_ENDIAN__
3531 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3532  uint16x8_t __ret;
3533  __ret = (uint16x8_t)(__p0 < __p1);
3534  return __ret;
3535 }
3536 #else
3537 __ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3538  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3539  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3540  uint16x8_t __ret;
3541  __ret = (uint16x8_t)(__rev0 < __rev1);
3542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3543  return __ret;
3544 }
3545 #endif
3546 
3547 #ifdef __LITTLE_ENDIAN__
3548 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3549  uint8x8_t __ret;
3550  __ret = (uint8x8_t)(__p0 < __p1);
3551  return __ret;
3552 }
3553 #else
3554 __ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3555  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3556  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3557  uint8x8_t __ret;
3558  __ret = (uint8x8_t)(__rev0 < __rev1);
3559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3560  return __ret;
3561 }
3562 #endif
3563 
3564 #ifdef __LITTLE_ENDIAN__
3565 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3566  uint32x2_t __ret;
3567  __ret = (uint32x2_t)(__p0 < __p1);
3568  return __ret;
3569 }
3570 #else
3571 __ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3572  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3573  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3574  uint32x2_t __ret;
3575  __ret = (uint32x2_t)(__rev0 < __rev1);
3576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3577  return __ret;
3578 }
3579 #endif
3580 
3581 #ifdef __LITTLE_ENDIAN__
3582 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3583  uint16x4_t __ret;
3584  __ret = (uint16x4_t)(__p0 < __p1);
3585  return __ret;
3586 }
3587 #else
3588 __ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3589  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3590  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3591  uint16x4_t __ret;
3592  __ret = (uint16x4_t)(__rev0 < __rev1);
3593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3594  return __ret;
3595 }
3596 #endif
3597 
3598 #ifdef __LITTLE_ENDIAN__
3599 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3600  uint8x8_t __ret;
3601  __ret = (uint8x8_t)(__p0 < __p1);
3602  return __ret;
3603 }
3604 #else
3605 __ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3606  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3607  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3608  uint8x8_t __ret;
3609  __ret = (uint8x8_t)(__rev0 < __rev1);
3610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3611  return __ret;
3612 }
3613 #endif
3614 
3615 #ifdef __LITTLE_ENDIAN__
3616 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3617  uint32x2_t __ret;
3618  __ret = (uint32x2_t)(__p0 < __p1);
3619  return __ret;
3620 }
3621 #else
3622 __ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3623  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3624  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3625  uint32x2_t __ret;
3626  __ret = (uint32x2_t)(__rev0 < __rev1);
3627  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3628  return __ret;
3629 }
3630 #endif
3631 
3632 #ifdef __LITTLE_ENDIAN__
3633 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3634  uint32x2_t __ret;
3635  __ret = (uint32x2_t)(__p0 < __p1);
3636  return __ret;
3637 }
3638 #else
3639 __ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3640  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3641  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3642  uint32x2_t __ret;
3643  __ret = (uint32x2_t)(__rev0 < __rev1);
3644  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3645  return __ret;
3646 }
3647 #endif
3648 
3649 #ifdef __LITTLE_ENDIAN__
3650 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3651  uint16x4_t __ret;
3652  __ret = (uint16x4_t)(__p0 < __p1);
3653  return __ret;
3654 }
3655 #else
3656 __ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3657  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3658  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3659  uint16x4_t __ret;
3660  __ret = (uint16x4_t)(__rev0 < __rev1);
3661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3662  return __ret;
3663 }
3664 #endif
3665 
3666 #ifdef __LITTLE_ENDIAN__
3667 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3668  uint8x16_t __ret;
3669  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
3670  return __ret;
3671 }
3672 #else
3673 __ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3674  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3675  uint8x16_t __ret;
3676  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
3677  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3678  return __ret;
3679 }
3680 #endif
3681 
3682 #ifdef __LITTLE_ENDIAN__
3683 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3684  uint32x4_t __ret;
3685  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
3686  return __ret;
3687 }
3688 #else
3689 __ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3690  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3691  uint32x4_t __ret;
3692  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
3693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3694  return __ret;
3695 }
3696 #endif
3697 
3698 #ifdef __LITTLE_ENDIAN__
3699 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3700  uint16x8_t __ret;
3701  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
3702  return __ret;
3703 }
3704 #else
3705 __ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3706  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3707  uint16x8_t __ret;
3708  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
3709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3710  return __ret;
3711 }
3712 #endif
3713 
3714 #ifdef __LITTLE_ENDIAN__
3715 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
3716  int8x16_t __ret;
3717  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
3718  return __ret;
3719 }
3720 #else
3721 __ai int8x16_t vclzq_s8(int8x16_t __p0) {
3722  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3723  int8x16_t __ret;
3724  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
3725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3726  return __ret;
3727 }
3728 #endif
3729 
3730 #ifdef __LITTLE_ENDIAN__
3731 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
3732  int32x4_t __ret;
3733  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
3734  return __ret;
3735 }
3736 #else
3737 __ai int32x4_t vclzq_s32(int32x4_t __p0) {
3738  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3739  int32x4_t __ret;
3740  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
3741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3742  return __ret;
3743 }
3744 #endif
3745 
3746 #ifdef __LITTLE_ENDIAN__
3747 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
3748  int16x8_t __ret;
3749  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
3750  return __ret;
3751 }
3752 #else
3753 __ai int16x8_t vclzq_s16(int16x8_t __p0) {
3754  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3755  int16x8_t __ret;
3756  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
3757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3758  return __ret;
3759 }
3760 #endif
3761 
3762 #ifdef __LITTLE_ENDIAN__
3763 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3764  uint8x8_t __ret;
3765  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
3766  return __ret;
3767 }
3768 #else
3769 __ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3770  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3771  uint8x8_t __ret;
3772  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
3773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3774  return __ret;
3775 }
3776 #endif
3777 
3778 #ifdef __LITTLE_ENDIAN__
3779 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3780  uint32x2_t __ret;
3781  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
3782  return __ret;
3783 }
3784 #else
3785 __ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3786  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3787  uint32x2_t __ret;
3788  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
3789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3790  return __ret;
3791 }
3792 #endif
3793 
3794 #ifdef __LITTLE_ENDIAN__
3795 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3796  uint16x4_t __ret;
3797  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
3798  return __ret;
3799 }
3800 #else
3801 __ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3802  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3803  uint16x4_t __ret;
3804  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
3805  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3806  return __ret;
3807 }
3808 #endif
3809 
3810 #ifdef __LITTLE_ENDIAN__
3811 __ai int8x8_t vclz_s8(int8x8_t __p0) {
3812  int8x8_t __ret;
3813  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
3814  return __ret;
3815 }
3816 #else
3817 __ai int8x8_t vclz_s8(int8x8_t __p0) {
3818  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3819  int8x8_t __ret;
3820  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
3821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3822  return __ret;
3823 }
3824 #endif
3825 
3826 #ifdef __LITTLE_ENDIAN__
3827 __ai int32x2_t vclz_s32(int32x2_t __p0) {
3828  int32x2_t __ret;
3829  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
3830  return __ret;
3831 }
3832 #else
3833 __ai int32x2_t vclz_s32(int32x2_t __p0) {
3834  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3835  int32x2_t __ret;
3836  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
3837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3838  return __ret;
3839 }
3840 #endif
3841 
3842 #ifdef __LITTLE_ENDIAN__
3843 __ai int16x4_t vclz_s16(int16x4_t __p0) {
3844  int16x4_t __ret;
3845  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
3846  return __ret;
3847 }
3848 #else
3849 __ai int16x4_t vclz_s16(int16x4_t __p0) {
3850  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3851  int16x4_t __ret;
3852  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
3853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3854  return __ret;
3855 }
3856 #endif
3857 
3858 #ifdef __LITTLE_ENDIAN__
3859 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3860  poly8x8_t __ret;
3861  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
3862  return __ret;
3863 }
3864 #else
3865 __ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3866  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3867  poly8x8_t __ret;
3868  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
3869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3870  return __ret;
3871 }
3872 #endif
3873 
3874 #ifdef __LITTLE_ENDIAN__
3875 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3876  poly8x16_t __ret;
3877  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
3878  return __ret;
3879 }
3880 #else
3881 __ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3882  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3883  poly8x16_t __ret;
3884  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
3885  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3886  return __ret;
3887 }
3888 #endif
3889 
3890 #ifdef __LITTLE_ENDIAN__
3891 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3892  uint8x16_t __ret;
3893  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
3894  return __ret;
3895 }
3896 #else
3897 __ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3898  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3899  uint8x16_t __ret;
3900  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
3901  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3902  return __ret;
3903 }
3904 #endif
3905 
3906 #ifdef __LITTLE_ENDIAN__
3907 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
3908  int8x16_t __ret;
3909  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
3910  return __ret;
3911 }
3912 #else
3913 __ai int8x16_t vcntq_s8(int8x16_t __p0) {
3914  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3915  int8x16_t __ret;
3916  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
3917  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3918  return __ret;
3919 }
3920 #endif
3921 
3922 #ifdef __LITTLE_ENDIAN__
3923 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3924  uint8x8_t __ret;
3925  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
3926  return __ret;
3927 }
3928 #else
3929 __ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3930  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3931  uint8x8_t __ret;
3932  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
3933  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3934  return __ret;
3935 }
3936 #endif
3937 
3938 #ifdef __LITTLE_ENDIAN__
3939 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
3940  int8x8_t __ret;
3941  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
3942  return __ret;
3943 }
3944 #else
3945 __ai int8x8_t vcnt_s8(int8x8_t __p0) {
3946  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3947  int8x8_t __ret;
3948  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
3949  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3950  return __ret;
3951 }
3952 #endif
3953 
3954 #ifdef __LITTLE_ENDIAN__
3955 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3956  poly8x16_t __ret;
3957  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3958  return __ret;
3959 }
3960 #else
3961 __ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3962  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3963  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3964  poly8x16_t __ret;
3965  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3966  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3967  return __ret;
3968 }
3969 #endif
3970 
3971 #ifdef __LITTLE_ENDIAN__
3972 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3973  poly16x8_t __ret;
3974  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
3975  return __ret;
3976 }
3977 #else
3978 __ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3979  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3980  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3981  poly16x8_t __ret;
3982  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
3983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3984  return __ret;
3985 }
3986 #endif
3987 
3988 #ifdef __LITTLE_ENDIAN__
3989 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3990  uint8x16_t __ret;
3991  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3992  return __ret;
3993 }
3994 #else
3995 __ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3996  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3997  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3998  uint8x16_t __ret;
3999  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4000  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4001  return __ret;
4002 }
4003 __ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
4004  uint8x16_t __ret;
4005  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4006  return __ret;
4007 }
4008 #endif
4009 
4010 #ifdef __LITTLE_ENDIAN__
4011 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4012  uint32x4_t __ret;
4013  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4014  return __ret;
4015 }
4016 #else
4017 __ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4018  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4019  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4020  uint32x4_t __ret;
4021  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4022  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4023  return __ret;
4024 }
4025 __ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4026  uint32x4_t __ret;
4027  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4028  return __ret;
4029 }
4030 #endif
4031 
4032 #ifdef __LITTLE_ENDIAN__
4033 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4034  uint64x2_t __ret;
4035  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4036  return __ret;
4037 }
4038 #else
4039 __ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4040  uint64x2_t __ret;
4041  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4043  return __ret;
4044 }
4045 #endif
4046 
4047 #ifdef __LITTLE_ENDIAN__
4048 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4049  uint16x8_t __ret;
4050  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4051  return __ret;
4052 }
4053 #else
4054 __ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4055  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4056  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4057  uint16x8_t __ret;
4058  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4060  return __ret;
4061 }
4062 __ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4063  uint16x8_t __ret;
4064  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4065  return __ret;
4066 }
4067 #endif
4068 
4069 #ifdef __LITTLE_ENDIAN__
4070 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4071  int8x16_t __ret;
4072  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4073  return __ret;
4074 }
4075 #else
4076 __ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4077  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4078  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4079  int8x16_t __ret;
4080  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4081  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4082  return __ret;
4083 }
4084 __ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4085  int8x16_t __ret;
4086  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4087  return __ret;
4088 }
4089 #endif
4090 
4091 #ifdef __LITTLE_ENDIAN__
4092 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4093  float32x4_t __ret;
4094  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4095  return __ret;
4096 }
4097 #else
4098 __ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4099  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4100  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4101  float32x4_t __ret;
4102  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4104  return __ret;
4105 }
4106 __ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4107  float32x4_t __ret;
4108  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4109  return __ret;
4110 }
4111 #endif
4112 
4113 #ifdef __LITTLE_ENDIAN__
4114 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4115  float16x8_t __ret;
4116  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4117  return __ret;
4118 }
4119 #else
4120 __ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4121  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4122  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4123  float16x8_t __ret;
4124  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4126  return __ret;
4127 }
4128 __ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4129  float16x8_t __ret;
4130  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4131  return __ret;
4132 }
4133 #endif
4134 
4135 #ifdef __LITTLE_ENDIAN__
4136 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4137  int32x4_t __ret;
4138  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4139  return __ret;
4140 }
4141 #else
4142 __ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4143  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4144  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4145  int32x4_t __ret;
4146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4148  return __ret;
4149 }
4150 __ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4151  int32x4_t __ret;
4152  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4153  return __ret;
4154 }
4155 #endif
4156 
4157 #ifdef __LITTLE_ENDIAN__
4158 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4159  int64x2_t __ret;
4160  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4161  return __ret;
4162 }
4163 #else
4164 __ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4165  int64x2_t __ret;
4166  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4168  return __ret;
4169 }
4170 #endif
4171 
4172 #ifdef __LITTLE_ENDIAN__
4173 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4174  int16x8_t __ret;
4175  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4176  return __ret;
4177 }
4178 #else
4179 __ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4180  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4181  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4182  int16x8_t __ret;
4183  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4185  return __ret;
4186 }
4187 __ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4188  int16x8_t __ret;
4189  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4190  return __ret;
4191 }
4192 #endif
4193 
4194 #ifdef __LITTLE_ENDIAN__
4195 __ai poly8x8_t vcreate_p8(uint64_t __p0) {
4196  poly8x8_t __ret;
4197  __ret = (poly8x8_t)(__p0);
4198  return __ret;
4199 }
4200 #else
4201 __ai poly8x8_t vcreate_p8(uint64_t __p0) {
4202  poly8x8_t __ret;
4203  __ret = (poly8x8_t)(__p0);
4204  return __ret;
4205 }
4206 #endif
4207 
4208 #ifdef __LITTLE_ENDIAN__
4209 __ai poly16x4_t vcreate_p16(uint64_t __p0) {
4210  poly16x4_t __ret;
4211  __ret = (poly16x4_t)(__p0);
4212  return __ret;
4213 }
4214 #else
4215 __ai poly16x4_t vcreate_p16(uint64_t __p0) {
4216  poly16x4_t __ret;
4217  __ret = (poly16x4_t)(__p0);
4218  return __ret;
4219 }
4220 #endif
4221 
4222 #ifdef __LITTLE_ENDIAN__
4223 __ai uint8x8_t vcreate_u8(uint64_t __p0) {
4224  uint8x8_t __ret;
4225  __ret = (uint8x8_t)(__p0);
4226  return __ret;
4227 }
4228 #else
4229 __ai uint8x8_t vcreate_u8(uint64_t __p0) {
4230  uint8x8_t __ret;
4231  __ret = (uint8x8_t)(__p0);
4232  return __ret;
4233 }
4234 #endif
4235 
4236 #ifdef __LITTLE_ENDIAN__
4237 __ai uint32x2_t vcreate_u32(uint64_t __p0) {
4238  uint32x2_t __ret;
4239  __ret = (uint32x2_t)(__p0);
4240  return __ret;
4241 }
4242 #else
4243 __ai uint32x2_t vcreate_u32(uint64_t __p0) {
4244  uint32x2_t __ret;
4245  __ret = (uint32x2_t)(__p0);
4246  return __ret;
4247 }
4248 #endif
4249 
4250 #ifdef __LITTLE_ENDIAN__
4251 __ai uint64x1_t vcreate_u64(uint64_t __p0) {
4252  uint64x1_t __ret;
4253  __ret = (uint64x1_t)(__p0);
4254  return __ret;
4255 }
4256 #else
4257 __ai uint64x1_t vcreate_u64(uint64_t __p0) {
4258  uint64x1_t __ret;
4259  __ret = (uint64x1_t)(__p0);
4260  return __ret;
4261 }
4262 #endif
4263 
4264 #ifdef __LITTLE_ENDIAN__
4265 __ai uint16x4_t vcreate_u16(uint64_t __p0) {
4266  uint16x4_t __ret;
4267  __ret = (uint16x4_t)(__p0);
4268  return __ret;
4269 }
4270 #else
4271 __ai uint16x4_t vcreate_u16(uint64_t __p0) {
4272  uint16x4_t __ret;
4273  __ret = (uint16x4_t)(__p0);
4274  return __ret;
4275 }
4276 #endif
4277 
4278 #ifdef __LITTLE_ENDIAN__
4279 __ai int8x8_t vcreate_s8(uint64_t __p0) {
4280  int8x8_t __ret;
4281  __ret = (int8x8_t)(__p0);
4282  return __ret;
4283 }
4284 #else
4285 __ai int8x8_t vcreate_s8(uint64_t __p0) {
4286  int8x8_t __ret;
4287  __ret = (int8x8_t)(__p0);
4288  return __ret;
4289 }
4290 #endif
4291 
4292 #ifdef __LITTLE_ENDIAN__
4293 __ai float32x2_t vcreate_f32(uint64_t __p0) {
4294  float32x2_t __ret;
4295  __ret = (float32x2_t)(__p0);
4296  return __ret;
4297 }
4298 #else
4299 __ai float32x2_t vcreate_f32(uint64_t __p0) {
4300  float32x2_t __ret;
4301  __ret = (float32x2_t)(__p0);
4302  return __ret;
4303 }
4304 #endif
4305 
4306 #ifdef __LITTLE_ENDIAN__
4307 __ai float16x4_t vcreate_f16(uint64_t __p0) {
4308  float16x4_t __ret;
4309  __ret = (float16x4_t)(__p0);
4310  return __ret;
4311 }
4312 #else
4313 __ai float16x4_t vcreate_f16(uint64_t __p0) {
4314  float16x4_t __ret;
4315  __ret = (float16x4_t)(__p0);
4316  return __ret;
4317 }
4318 #endif
4319 
4320 #ifdef __LITTLE_ENDIAN__
4321 __ai int32x2_t vcreate_s32(uint64_t __p0) {
4322  int32x2_t __ret;
4323  __ret = (int32x2_t)(__p0);
4324  return __ret;
4325 }
4326 #else
4327 __ai int32x2_t vcreate_s32(uint64_t __p0) {
4328  int32x2_t __ret;
4329  __ret = (int32x2_t)(__p0);
4330  return __ret;
4331 }
4332 #endif
4333 
4334 #ifdef __LITTLE_ENDIAN__
4335 __ai int64x1_t vcreate_s64(uint64_t __p0) {
4336  int64x1_t __ret;
4337  __ret = (int64x1_t)(__p0);
4338  return __ret;
4339 }
4340 #else
4341 __ai int64x1_t vcreate_s64(uint64_t __p0) {
4342  int64x1_t __ret;
4343  __ret = (int64x1_t)(__p0);
4344  return __ret;
4345 }
4346 #endif
4347 
4348 #ifdef __LITTLE_ENDIAN__
4349 __ai int16x4_t vcreate_s16(uint64_t __p0) {
4350  int16x4_t __ret;
4351  __ret = (int16x4_t)(__p0);
4352  return __ret;
4353 }
4354 #else
4355 __ai int16x4_t vcreate_s16(uint64_t __p0) {
4356  int16x4_t __ret;
4357  __ret = (int16x4_t)(__p0);
4358  return __ret;
4359 }
4360 #endif
4361 
4362 #ifdef __LITTLE_ENDIAN__
4363 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4364  float16x4_t __ret;
4365  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
4366  return __ret;
4367 }
4368 #else
4369 __ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4370  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4371  float16x4_t __ret;
4372  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 8);
4373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4374  return __ret;
4375 }
4376 __ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
4377  float16x4_t __ret;
4378  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
4379  return __ret;
4380 }
4381 #endif
4382 
4383 #ifdef __LITTLE_ENDIAN__
4384 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4385  float32x4_t __ret;
4386  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
4387  return __ret;
4388 }
4389 #else
4390 __ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4391  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4392  float32x4_t __ret;
4393  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
4394  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4395  return __ret;
4396 }
4397 #endif
4398 
4399 #ifdef __LITTLE_ENDIAN__
4400 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4401  float32x4_t __ret;
4402  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
4403  return __ret;
4404 }
4405 #else
4406 __ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4407  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4408  float32x4_t __ret;
4409  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
4410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4411  return __ret;
4412 }
4413 #endif
4414 
4415 #ifdef __LITTLE_ENDIAN__
4416 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4417  float32x2_t __ret;
4418  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
4419  return __ret;
4420 }
4421 #else
4422 __ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4423  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4424  float32x2_t __ret;
4425  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
4426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4427  return __ret;
4428 }
4429 #endif
4430 
4431 #ifdef __LITTLE_ENDIAN__
4432 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4433  float32x2_t __ret;
4434  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
4435  return __ret;
4436 }
4437 #else
4438 __ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4439  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4440  float32x2_t __ret;
4441  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
4442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4443  return __ret;
4444 }
4445 #endif
4446 
4447 #ifdef __LITTLE_ENDIAN__
4448 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4449  float32x4_t __ret;
4450  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
4451  return __ret;
4452 }
4453 #else
4454 __ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4455  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4456  float32x4_t __ret;
4457  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 41);
4458  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4459  return __ret;
4460 }
4461 __ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
4462  float32x4_t __ret;
4463  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
4464  return __ret;
4465 }
4466 #endif
4467 
4468 #ifdef __LITTLE_ENDIAN__
4469 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4470  uint32x4_t __s0 = __p0; \
4471  float32x4_t __ret; \
4472  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
4473  __ret; \
4474 })
4475 #else
4476 #define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4477  uint32x4_t __s0 = __p0; \
4478  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4479  float32x4_t __ret; \
4480  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
4481  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4482  __ret; \
4483 })
4484 #endif
4485 
4486 #ifdef __LITTLE_ENDIAN__
4487 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4488  int32x4_t __s0 = __p0; \
4489  float32x4_t __ret; \
4490  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
4491  __ret; \
4492 })
4493 #else
4494 #define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4495  int32x4_t __s0 = __p0; \
4496  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4497  float32x4_t __ret; \
4498  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
4499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4500  __ret; \
4501 })
4502 #endif
4503 
4504 #ifdef __LITTLE_ENDIAN__
4505 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4506  uint32x2_t __s0 = __p0; \
4507  float32x2_t __ret; \
4508  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
4509  __ret; \
4510 })
4511 #else
4512 #define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4513  uint32x2_t __s0 = __p0; \
4514  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4515  float32x2_t __ret; \
4516  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
4517  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4518  __ret; \
4519 })
4520 #endif
4521 
4522 #ifdef __LITTLE_ENDIAN__
4523 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4524  int32x2_t __s0 = __p0; \
4525  float32x2_t __ret; \
4526  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
4527  __ret; \
4528 })
4529 #else
4530 #define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4531  int32x2_t __s0 = __p0; \
4532  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4533  float32x2_t __ret; \
4534  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
4535  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4536  __ret; \
4537 })
4538 #endif
4539 
4540 #ifdef __LITTLE_ENDIAN__
4541 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4542  float32x4_t __s0 = __p0; \
4543  int32x4_t __ret; \
4544  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
4545  __ret; \
4546 })
4547 #else
4548 #define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4549  float32x4_t __s0 = __p0; \
4550  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4551  int32x4_t __ret; \
4552  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
4553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4554  __ret; \
4555 })
4556 #endif
4557 
4558 #ifdef __LITTLE_ENDIAN__
4559 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4560  float32x2_t __s0 = __p0; \
4561  int32x2_t __ret; \
4562  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
4563  __ret; \
4564 })
4565 #else
4566 #define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4567  float32x2_t __s0 = __p0; \
4568  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4569  int32x2_t __ret; \
4570  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
4571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4572  __ret; \
4573 })
4574 #endif
4575 
4576 #ifdef __LITTLE_ENDIAN__
4577 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4578  float32x4_t __s0 = __p0; \
4579  uint32x4_t __ret; \
4580  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
4581  __ret; \
4582 })
4583 #else
4584 #define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4585  float32x4_t __s0 = __p0; \
4586  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4587  uint32x4_t __ret; \
4588  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
4589  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4590  __ret; \
4591 })
4592 #endif
4593 
4594 #ifdef __LITTLE_ENDIAN__
4595 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4596  float32x2_t __s0 = __p0; \
4597  uint32x2_t __ret; \
4598  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
4599  __ret; \
4600 })
4601 #else
4602 #define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4603  float32x2_t __s0 = __p0; \
4604  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4605  uint32x2_t __ret; \
4606  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
4607  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4608  __ret; \
4609 })
4610 #endif
4611 
4612 #ifdef __LITTLE_ENDIAN__
4613 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4614  int32x4_t __ret;
4615  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
4616  return __ret;
4617 }
4618 #else
4619 __ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4620  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4621  int32x4_t __ret;
4622  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
4623  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4624  return __ret;
4625 }
4626 #endif
4627 
4628 #ifdef __LITTLE_ENDIAN__
4629 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4630  int32x2_t __ret;
4631  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
4632  return __ret;
4633 }
4634 #else
4635 __ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4636  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4637  int32x2_t __ret;
4638  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
4639  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4640  return __ret;
4641 }
4642 #endif
4643 
4644 #ifdef __LITTLE_ENDIAN__
4645 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4646  uint32x4_t __ret;
4647  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
4648  return __ret;
4649 }
4650 #else
4651 __ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4652  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4653  uint32x4_t __ret;
4654  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
4655  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4656  return __ret;
4657 }
4658 #endif
4659 
4660 #ifdef __LITTLE_ENDIAN__
4661 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4662  uint32x2_t __ret;
4663  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
4664  return __ret;
4665 }
4666 #else
4667 __ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4668  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4669  uint32x2_t __ret;
4670  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
4671  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4672  return __ret;
4673 }
4674 #endif
4675 
4676 #ifdef __LITTLE_ENDIAN__
4677 #define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4678  poly8x8_t __s0 = __p0; \
4679  poly8x8_t __ret; \
4680  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4681  __ret; \
4682 })
4683 #else
4684 #define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4685  poly8x8_t __s0 = __p0; \
4686  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4687  poly8x8_t __ret; \
4688  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4689  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4690  __ret; \
4691 })
4692 #endif
4693 
4694 #ifdef __LITTLE_ENDIAN__
4695 #define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4696  poly16x4_t __s0 = __p0; \
4697  poly16x4_t __ret; \
4698  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4699  __ret; \
4700 })
4701 #else
4702 #define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4703  poly16x4_t __s0 = __p0; \
4704  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4705  poly16x4_t __ret; \
4706  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4707  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4708  __ret; \
4709 })
4710 #endif
4711 
4712 #ifdef __LITTLE_ENDIAN__
4713 #define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4714  poly8x8_t __s0 = __p0; \
4715  poly8x16_t __ret; \
4716  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4717  __ret; \
4718 })
4719 #else
4720 #define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4721  poly8x8_t __s0 = __p0; \
4722  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4723  poly8x16_t __ret; \
4724  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4726  __ret; \
4727 })
4728 #endif
4729 
4730 #ifdef __LITTLE_ENDIAN__
4731 #define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4732  poly16x4_t __s0 = __p0; \
4733  poly16x8_t __ret; \
4734  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4735  __ret; \
4736 })
4737 #else
4738 #define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4739  poly16x4_t __s0 = __p0; \
4740  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4741  poly16x8_t __ret; \
4742  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4743  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4744  __ret; \
4745 })
4746 #endif
4747 
4748 #ifdef __LITTLE_ENDIAN__
4749 #define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4750  uint8x8_t __s0 = __p0; \
4751  uint8x16_t __ret; \
4752  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4753  __ret; \
4754 })
4755 #else
4756 #define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4757  uint8x8_t __s0 = __p0; \
4758  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4759  uint8x16_t __ret; \
4760  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4761  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4762  __ret; \
4763 })
4764 #endif
4765 
4766 #ifdef __LITTLE_ENDIAN__
4767 #define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4768  uint32x2_t __s0 = __p0; \
4769  uint32x4_t __ret; \
4770  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4771  __ret; \
4772 })
4773 #else
4774 #define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4775  uint32x2_t __s0 = __p0; \
4776  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4777  uint32x4_t __ret; \
4778  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4779  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4780  __ret; \
4781 })
4782 #endif
4783 
4784 #ifdef __LITTLE_ENDIAN__
4785 #define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4786  uint64x1_t __s0 = __p0; \
4787  uint64x2_t __ret; \
4788  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4789  __ret; \
4790 })
4791 #else
4792 #define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4793  uint64x1_t __s0 = __p0; \
4794  uint64x2_t __ret; \
4795  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4796  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4797  __ret; \
4798 })
4799 #endif
4800 
4801 #ifdef __LITTLE_ENDIAN__
4802 #define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4803  uint16x4_t __s0 = __p0; \
4804  uint16x8_t __ret; \
4805  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4806  __ret; \
4807 })
4808 #else
4809 #define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4810  uint16x4_t __s0 = __p0; \
4811  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4812  uint16x8_t __ret; \
4813  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4814  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4815  __ret; \
4816 })
4817 #endif
4818 
4819 #ifdef __LITTLE_ENDIAN__
4820 #define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4821  int8x8_t __s0 = __p0; \
4822  int8x16_t __ret; \
4823  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4824  __ret; \
4825 })
4826 #else
4827 #define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4828  int8x8_t __s0 = __p0; \
4829  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4830  int8x16_t __ret; \
4831  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4832  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4833  __ret; \
4834 })
4835 #endif
4836 
4837 #ifdef __LITTLE_ENDIAN__
4838 #define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4839  float32x2_t __s0 = __p0; \
4840  float32x4_t __ret; \
4841  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4842  __ret; \
4843 })
4844 #else
4845 #define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4846  float32x2_t __s0 = __p0; \
4847  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4848  float32x4_t __ret; \
4849  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4850  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4851  __ret; \
4852 })
4853 #endif
4854 
4855 #ifdef __LITTLE_ENDIAN__
4856 #define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4857  int32x2_t __s0 = __p0; \
4858  int32x4_t __ret; \
4859  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4860  __ret; \
4861 })
4862 #else
4863 #define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4864  int32x2_t __s0 = __p0; \
4865  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4866  int32x4_t __ret; \
4867  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4869  __ret; \
4870 })
4871 #endif
4872 
4873 #ifdef __LITTLE_ENDIAN__
4874 #define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4875  int64x1_t __s0 = __p0; \
4876  int64x2_t __ret; \
4877  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4878  __ret; \
4879 })
4880 #else
4881 #define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4882  int64x1_t __s0 = __p0; \
4883  int64x2_t __ret; \
4884  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4886  __ret; \
4887 })
4888 #endif
4889 
4890 #ifdef __LITTLE_ENDIAN__
4891 #define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4892  int16x4_t __s0 = __p0; \
4893  int16x8_t __ret; \
4894  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4895  __ret; \
4896 })
4897 #else
4898 #define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4899  int16x4_t __s0 = __p0; \
4900  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4901  int16x8_t __ret; \
4902  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4904  __ret; \
4905 })
4906 #endif
4907 
4908 #ifdef __LITTLE_ENDIAN__
4909 #define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4910  uint8x8_t __s0 = __p0; \
4911  uint8x8_t __ret; \
4912  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4913  __ret; \
4914 })
4915 #else
4916 #define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4917  uint8x8_t __s0 = __p0; \
4918  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4919  uint8x8_t __ret; \
4920  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4922  __ret; \
4923 })
4924 #endif
4925 
4926 #ifdef __LITTLE_ENDIAN__
4927 #define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4928  uint32x2_t __s0 = __p0; \
4929  uint32x2_t __ret; \
4930  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4931  __ret; \
4932 })
4933 #else
4934 #define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4935  uint32x2_t __s0 = __p0; \
4936  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4937  uint32x2_t __ret; \
4938  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4939  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4940  __ret; \
4941 })
4942 #endif
4943 
4944 #ifdef __LITTLE_ENDIAN__
4945 #define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4946  uint64x1_t __s0 = __p0; \
4947  uint64x1_t __ret; \
4948  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4949  __ret; \
4950 })
4951 #else
4952 #define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4953  uint64x1_t __s0 = __p0; \
4954  uint64x1_t __ret; \
4955  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4956  __ret; \
4957 })
4958 #endif
4959 
4960 #ifdef __LITTLE_ENDIAN__
4961 #define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4962  uint16x4_t __s0 = __p0; \
4963  uint16x4_t __ret; \
4964  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4965  __ret; \
4966 })
4967 #else
4968 #define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4969  uint16x4_t __s0 = __p0; \
4970  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4971  uint16x4_t __ret; \
4972  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4973  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4974  __ret; \
4975 })
4976 #endif
4977 
4978 #ifdef __LITTLE_ENDIAN__
4979 #define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4980  int8x8_t __s0 = __p0; \
4981  int8x8_t __ret; \
4982  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4983  __ret; \
4984 })
4985 #else
4986 #define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4987  int8x8_t __s0 = __p0; \
4988  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4989  int8x8_t __ret; \
4990  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4991  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4992  __ret; \
4993 })
4994 #endif
4995 
4996 #ifdef __LITTLE_ENDIAN__
4997 #define vdup_lane_f32(__p0, __p1) __extension__ ({ \
4998  float32x2_t __s0 = __p0; \
4999  float32x2_t __ret; \
5000  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5001  __ret; \
5002 })
5003 #else
5004 #define vdup_lane_f32(__p0, __p1) __extension__ ({ \
5005  float32x2_t __s0 = __p0; \
5006  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5007  float32x2_t __ret; \
5008  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5010  __ret; \
5011 })
5012 #endif
5013 
5014 #ifdef __LITTLE_ENDIAN__
5015 #define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5016  int32x2_t __s0 = __p0; \
5017  int32x2_t __ret; \
5018  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5019  __ret; \
5020 })
5021 #else
5022 #define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5023  int32x2_t __s0 = __p0; \
5024  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5025  int32x2_t __ret; \
5026  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5028  __ret; \
5029 })
5030 #endif
5031 
5032 #ifdef __LITTLE_ENDIAN__
5033 #define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5034  int64x1_t __s0 = __p0; \
5035  int64x1_t __ret; \
5036  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5037  __ret; \
5038 })
5039 #else
5040 #define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5041  int64x1_t __s0 = __p0; \
5042  int64x1_t __ret; \
5043  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5044  __ret; \
5045 })
5046 #endif
5047 
5048 #ifdef __LITTLE_ENDIAN__
5049 #define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5050  int16x4_t __s0 = __p0; \
5051  int16x4_t __ret; \
5052  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
5053  __ret; \
5054 })
5055 #else
5056 #define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5057  int16x4_t __s0 = __p0; \
5058  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5059  int16x4_t __ret; \
5060  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
5061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5062  __ret; \
5063 })
5064 #endif
5065 
5066 #ifdef __LITTLE_ENDIAN__
5067 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5068  poly8x8_t __ret;
5069  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5070  return __ret;
5071 }
5072 #else
5073 __ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5074  poly8x8_t __ret;
5075  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5076  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5077  return __ret;
5078 }
5079 #endif
5080 
5081 #ifdef __LITTLE_ENDIAN__
5082 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5083  poly16x4_t __ret;
5084  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5085  return __ret;
5086 }
5087 #else
5088 __ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5089  poly16x4_t __ret;
5090  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5091  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5092  return __ret;
5093 }
5094 #endif
5095 
5096 #ifdef __LITTLE_ENDIAN__
5097 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5098  poly8x16_t __ret;
5099  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5100  return __ret;
5101 }
5102 #else
5103 __ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5104  poly8x16_t __ret;
5105  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5107  return __ret;
5108 }
5109 #endif
5110 
5111 #ifdef __LITTLE_ENDIAN__
5112 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5113  poly16x8_t __ret;
5114  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5115  return __ret;
5116 }
5117 #else
5118 __ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5119  poly16x8_t __ret;
5120  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5121  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5122  return __ret;
5123 }
5124 #endif
5125 
5126 #ifdef __LITTLE_ENDIAN__
5127 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5128  uint8x16_t __ret;
5129  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5130  return __ret;
5131 }
5132 #else
5133 __ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5134  uint8x16_t __ret;
5135  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5136  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5137  return __ret;
5138 }
5139 #endif
5140 
5141 #ifdef __LITTLE_ENDIAN__
5142 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5143  uint32x4_t __ret;
5144  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5145  return __ret;
5146 }
5147 #else
5148 __ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5149  uint32x4_t __ret;
5150  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5151  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5152  return __ret;
5153 }
5154 #endif
5155 
5156 #ifdef __LITTLE_ENDIAN__
5157 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5158  uint64x2_t __ret;
5159  __ret = (uint64x2_t) {__p0, __p0};
5160  return __ret;
5161 }
5162 #else
5163 __ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5164  uint64x2_t __ret;
5165  __ret = (uint64x2_t) {__p0, __p0};
5166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5167  return __ret;
5168 }
5169 #endif
5170 
5171 #ifdef __LITTLE_ENDIAN__
5172 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5173  uint16x8_t __ret;
5174  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5175  return __ret;
5176 }
5177 #else
5178 __ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5179  uint16x8_t __ret;
5180  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5181  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5182  return __ret;
5183 }
5184 #endif
5185 
5186 #ifdef __LITTLE_ENDIAN__
5187 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
5188  int8x16_t __ret;
5189  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5190  return __ret;
5191 }
5192 #else
5193 __ai int8x16_t vdupq_n_s8(int8_t __p0) {
5194  int8x16_t __ret;
5195  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5196  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5197  return __ret;
5198 }
5199 #endif
5200 
5201 #ifdef __LITTLE_ENDIAN__
5202 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
5203  float32x4_t __ret;
5204  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5205  return __ret;
5206 }
5207 #else
5208 __ai float32x4_t vdupq_n_f32(float32_t __p0) {
5209  float32x4_t __ret;
5210  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5211  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5212  return __ret;
5213 }
5214 #endif
5215 
5216 #ifdef __LITTLE_ENDIAN__
5217 #define vdupq_n_f16(__p0) __extension__ ({ \
5218  float16_t __s0 = __p0; \
5219  float16x8_t __ret; \
5220  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5221  __ret; \
5222 })
5223 #else
5224 #define vdupq_n_f16(__p0) __extension__ ({ \
5225  float16_t __s0 = __p0; \
5226  float16x8_t __ret; \
5227  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5228  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5229  __ret; \
5230 })
5231 #endif
5232 
5233 #ifdef __LITTLE_ENDIAN__
5234 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
5235  int32x4_t __ret;
5236  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5237  return __ret;
5238 }
5239 #else
5240 __ai int32x4_t vdupq_n_s32(int32_t __p0) {
5241  int32x4_t __ret;
5242  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5244  return __ret;
5245 }
5246 #endif
5247 
5248 #ifdef __LITTLE_ENDIAN__
5249 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
5250  int64x2_t __ret;
5251  __ret = (int64x2_t) {__p0, __p0};
5252  return __ret;
5253 }
5254 #else
5255 __ai int64x2_t vdupq_n_s64(int64_t __p0) {
5256  int64x2_t __ret;
5257  __ret = (int64x2_t) {__p0, __p0};
5258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5259  return __ret;
5260 }
5261 #endif
5262 
5263 #ifdef __LITTLE_ENDIAN__
5264 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
5265  int16x8_t __ret;
5266  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5267  return __ret;
5268 }
5269 #else
5270 __ai int16x8_t vdupq_n_s16(int16_t __p0) {
5271  int16x8_t __ret;
5272  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5273  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5274  return __ret;
5275 }
5276 #endif
5277 
5278 #ifdef __LITTLE_ENDIAN__
5279 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5280  uint8x8_t __ret;
5281  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5282  return __ret;
5283 }
5284 #else
5285 __ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5286  uint8x8_t __ret;
5287  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5288  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5289  return __ret;
5290 }
5291 #endif
5292 
5293 #ifdef __LITTLE_ENDIAN__
5294 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5295  uint32x2_t __ret;
5296  __ret = (uint32x2_t) {__p0, __p0};
5297  return __ret;
5298 }
5299 #else
5300 __ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5301  uint32x2_t __ret;
5302  __ret = (uint32x2_t) {__p0, __p0};
5303  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5304  return __ret;
5305 }
5306 #endif
5307 
5308 #ifdef __LITTLE_ENDIAN__
5309 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5310  uint64x1_t __ret;
5311  __ret = (uint64x1_t) {__p0};
5312  return __ret;
5313 }
5314 #else
5315 __ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5316  uint64x1_t __ret;
5317  __ret = (uint64x1_t) {__p0};
5318  return __ret;
5319 }
5320 #endif
5321 
5322 #ifdef __LITTLE_ENDIAN__
5323 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5324  uint16x4_t __ret;
5325  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5326  return __ret;
5327 }
5328 #else
5329 __ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5330  uint16x4_t __ret;
5331  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5332  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5333  return __ret;
5334 }
5335 #endif
5336 
5337 #ifdef __LITTLE_ENDIAN__
5338 __ai int8x8_t vdup_n_s8(int8_t __p0) {
5339  int8x8_t __ret;
5340  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5341  return __ret;
5342 }
5343 #else
5344 __ai int8x8_t vdup_n_s8(int8_t __p0) {
5345  int8x8_t __ret;
5346  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5347  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5348  return __ret;
5349 }
5350 #endif
5351 
5352 #ifdef __LITTLE_ENDIAN__
5353 __ai float32x2_t vdup_n_f32(float32_t __p0) {
5354  float32x2_t __ret;
5355  __ret = (float32x2_t) {__p0, __p0};
5356  return __ret;
5357 }
5358 #else
5359 __ai float32x2_t vdup_n_f32(float32_t __p0) {
5360  float32x2_t __ret;
5361  __ret = (float32x2_t) {__p0, __p0};
5362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5363  return __ret;
5364 }
5365 #endif
5366 
5367 #ifdef __LITTLE_ENDIAN__
5368 #define vdup_n_f16(__p0) __extension__ ({ \
5369  float16_t __s0 = __p0; \
5370  float16x4_t __ret; \
5371  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5372  __ret; \
5373 })
5374 #else
5375 #define vdup_n_f16(__p0) __extension__ ({ \
5376  float16_t __s0 = __p0; \
5377  float16x4_t __ret; \
5378  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5379  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5380  __ret; \
5381 })
5382 #endif
5383 
5384 #ifdef __LITTLE_ENDIAN__
5385 __ai int32x2_t vdup_n_s32(int32_t __p0) {
5386  int32x2_t __ret;
5387  __ret = (int32x2_t) {__p0, __p0};
5388  return __ret;
5389 }
5390 #else
5391 __ai int32x2_t vdup_n_s32(int32_t __p0) {
5392  int32x2_t __ret;
5393  __ret = (int32x2_t) {__p0, __p0};
5394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5395  return __ret;
5396 }
5397 #endif
5398 
5399 #ifdef __LITTLE_ENDIAN__
5400 __ai int64x1_t vdup_n_s64(int64_t __p0) {
5401  int64x1_t __ret;
5402  __ret = (int64x1_t) {__p0};
5403  return __ret;
5404 }
5405 #else
5406 __ai int64x1_t vdup_n_s64(int64_t __p0) {
5407  int64x1_t __ret;
5408  __ret = (int64x1_t) {__p0};
5409  return __ret;
5410 }
5411 #endif
5412 
5413 #ifdef __LITTLE_ENDIAN__
5414 __ai int16x4_t vdup_n_s16(int16_t __p0) {
5415  int16x4_t __ret;
5416  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5417  return __ret;
5418 }
5419 #else
5420 __ai int16x4_t vdup_n_s16(int16_t __p0) {
5421  int16x4_t __ret;
5422  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5424  return __ret;
5425 }
5426 #endif
5427 
5428 #ifdef __LITTLE_ENDIAN__
5429 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5430  uint8x16_t __ret;
5431  __ret = __p0 ^ __p1;
5432  return __ret;
5433 }
5434 #else
5435 __ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5436  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5437  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5438  uint8x16_t __ret;
5439  __ret = __rev0 ^ __rev1;
5440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5441  return __ret;
5442 }
5443 #endif
5444 
5445 #ifdef __LITTLE_ENDIAN__
5446 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5447  uint32x4_t __ret;
5448  __ret = __p0 ^ __p1;
5449  return __ret;
5450 }
5451 #else
5452 __ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5453  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5454  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5455  uint32x4_t __ret;
5456  __ret = __rev0 ^ __rev1;
5457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5458  return __ret;
5459 }
5460 #endif
5461 
5462 #ifdef __LITTLE_ENDIAN__
5463 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5464  uint64x2_t __ret;
5465  __ret = __p0 ^ __p1;
5466  return __ret;
5467 }
5468 #else
5469 __ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5470  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5471  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5472  uint64x2_t __ret;
5473  __ret = __rev0 ^ __rev1;
5474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5475  return __ret;
5476 }
5477 #endif
5478 
5479 #ifdef __LITTLE_ENDIAN__
5480 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5481  uint16x8_t __ret;
5482  __ret = __p0 ^ __p1;
5483  return __ret;
5484 }
5485 #else
5486 __ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5487  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5488  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5489  uint16x8_t __ret;
5490  __ret = __rev0 ^ __rev1;
5491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5492  return __ret;
5493 }
5494 #endif
5495 
5496 #ifdef __LITTLE_ENDIAN__
5497 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5498  int8x16_t __ret;
5499  __ret = __p0 ^ __p1;
5500  return __ret;
5501 }
5502 #else
5503 __ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5504  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5505  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5506  int8x16_t __ret;
5507  __ret = __rev0 ^ __rev1;
5508  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5509  return __ret;
5510 }
5511 #endif
5512 
5513 #ifdef __LITTLE_ENDIAN__
5514 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5515  int32x4_t __ret;
5516  __ret = __p0 ^ __p1;
5517  return __ret;
5518 }
5519 #else
5520 __ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5521  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5522  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5523  int32x4_t __ret;
5524  __ret = __rev0 ^ __rev1;
5525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5526  return __ret;
5527 }
5528 #endif
5529 
5530 #ifdef __LITTLE_ENDIAN__
5531 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5532  int64x2_t __ret;
5533  __ret = __p0 ^ __p1;
5534  return __ret;
5535 }
5536 #else
5537 __ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5538  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5539  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5540  int64x2_t __ret;
5541  __ret = __rev0 ^ __rev1;
5542  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5543  return __ret;
5544 }
5545 #endif
5546 
5547 #ifdef __LITTLE_ENDIAN__
5548 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5549  int16x8_t __ret;
5550  __ret = __p0 ^ __p1;
5551  return __ret;
5552 }
5553 #else
5554 __ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5555  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5556  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5557  int16x8_t __ret;
5558  __ret = __rev0 ^ __rev1;
5559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5560  return __ret;
5561 }
5562 #endif
5563 
5564 #ifdef __LITTLE_ENDIAN__
5565 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5566  uint8x8_t __ret;
5567  __ret = __p0 ^ __p1;
5568  return __ret;
5569 }
5570 #else
5571 __ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5572  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5573  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5574  uint8x8_t __ret;
5575  __ret = __rev0 ^ __rev1;
5576  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5577  return __ret;
5578 }
5579 #endif
5580 
5581 #ifdef __LITTLE_ENDIAN__
5582 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5583  uint32x2_t __ret;
5584  __ret = __p0 ^ __p1;
5585  return __ret;
5586 }
5587 #else
5588 __ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5589  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5590  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5591  uint32x2_t __ret;
5592  __ret = __rev0 ^ __rev1;
5593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5594  return __ret;
5595 }
5596 #endif
5597 
5598 #ifdef __LITTLE_ENDIAN__
5599 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5600  uint64x1_t __ret;
5601  __ret = __p0 ^ __p1;
5602  return __ret;
5603 }
5604 #else
5605 __ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5606  uint64x1_t __ret;
5607  __ret = __p0 ^ __p1;
5608  return __ret;
5609 }
5610 #endif
5611 
5612 #ifdef __LITTLE_ENDIAN__
5613 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5614  uint16x4_t __ret;
5615  __ret = __p0 ^ __p1;
5616  return __ret;
5617 }
5618 #else
5619 __ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5620  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5621  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5622  uint16x4_t __ret;
5623  __ret = __rev0 ^ __rev1;
5624  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5625  return __ret;
5626 }
5627 #endif
5628 
5629 #ifdef __LITTLE_ENDIAN__
5630 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5631  int8x8_t __ret;
5632  __ret = __p0 ^ __p1;
5633  return __ret;
5634 }
5635 #else
5636 __ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5637  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5638  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5639  int8x8_t __ret;
5640  __ret = __rev0 ^ __rev1;
5641  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5642  return __ret;
5643 }
5644 #endif
5645 
5646 #ifdef __LITTLE_ENDIAN__
5647 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5648  int32x2_t __ret;
5649  __ret = __p0 ^ __p1;
5650  return __ret;
5651 }
5652 #else
5653 __ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5654  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5655  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5656  int32x2_t __ret;
5657  __ret = __rev0 ^ __rev1;
5658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5659  return __ret;
5660 }
5661 #endif
5662 
5663 #ifdef __LITTLE_ENDIAN__
5664 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5665  int64x1_t __ret;
5666  __ret = __p0 ^ __p1;
5667  return __ret;
5668 }
5669 #else
5670 __ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5671  int64x1_t __ret;
5672  __ret = __p0 ^ __p1;
5673  return __ret;
5674 }
5675 #endif
5676 
5677 #ifdef __LITTLE_ENDIAN__
5678 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5679  int16x4_t __ret;
5680  __ret = __p0 ^ __p1;
5681  return __ret;
5682 }
5683 #else
5684 __ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5685  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5686  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5687  int16x4_t __ret;
5688  __ret = __rev0 ^ __rev1;
5689  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5690  return __ret;
5691 }
5692 #endif
5693 
5694 #ifdef __LITTLE_ENDIAN__
5695 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5696  poly8x8_t __s0 = __p0; \
5697  poly8x8_t __s1 = __p1; \
5698  poly8x8_t __ret; \
5699  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
5700  __ret; \
5701 })
5702 #else
5703 #define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5704  poly8x8_t __s0 = __p0; \
5705  poly8x8_t __s1 = __p1; \
5706  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5707  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5708  poly8x8_t __ret; \
5709  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
5710  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5711  __ret; \
5712 })
5713 #endif
5714 
5715 #ifdef __LITTLE_ENDIAN__
5716 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5717  poly16x4_t __s0 = __p0; \
5718  poly16x4_t __s1 = __p1; \
5719  poly16x4_t __ret; \
5720  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
5721  __ret; \
5722 })
5723 #else
5724 #define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5725  poly16x4_t __s0 = __p0; \
5726  poly16x4_t __s1 = __p1; \
5727  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5728  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5729  poly16x4_t __ret; \
5730  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
5731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5732  __ret; \
5733 })
5734 #endif
5735 
5736 #ifdef __LITTLE_ENDIAN__
5737 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5738  poly8x16_t __s0 = __p0; \
5739  poly8x16_t __s1 = __p1; \
5740  poly8x16_t __ret; \
5741  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
5742  __ret; \
5743 })
5744 #else
5745 #define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5746  poly8x16_t __s0 = __p0; \
5747  poly8x16_t __s1 = __p1; \
5748  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5749  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5750  poly8x16_t __ret; \
5751  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
5752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5753  __ret; \
5754 })
5755 #endif
5756 
5757 #ifdef __LITTLE_ENDIAN__
5758 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5759  poly16x8_t __s0 = __p0; \
5760  poly16x8_t __s1 = __p1; \
5761  poly16x8_t __ret; \
5762  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
5763  __ret; \
5764 })
5765 #else
5766 #define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5767  poly16x8_t __s0 = __p0; \
5768  poly16x8_t __s1 = __p1; \
5769  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5770  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5771  poly16x8_t __ret; \
5772  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
5773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5774  __ret; \
5775 })
5776 #endif
5777 
5778 #ifdef __LITTLE_ENDIAN__
5779 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5780  uint8x16_t __s0 = __p0; \
5781  uint8x16_t __s1 = __p1; \
5782  uint8x16_t __ret; \
5783  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
5784  __ret; \
5785 })
5786 #else
5787 #define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5788  uint8x16_t __s0 = __p0; \
5789  uint8x16_t __s1 = __p1; \
5790  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5791  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5792  uint8x16_t __ret; \
5793  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
5794  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5795  __ret; \
5796 })
5797 #endif
5798 
5799 #ifdef __LITTLE_ENDIAN__
5800 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5801  uint32x4_t __s0 = __p0; \
5802  uint32x4_t __s1 = __p1; \
5803  uint32x4_t __ret; \
5804  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
5805  __ret; \
5806 })
5807 #else
5808 #define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5809  uint32x4_t __s0 = __p0; \
5810  uint32x4_t __s1 = __p1; \
5811  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5812  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5813  uint32x4_t __ret; \
5814  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
5815  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5816  __ret; \
5817 })
5818 #endif
5819 
5820 #ifdef __LITTLE_ENDIAN__
5821 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5822  uint64x2_t __s0 = __p0; \
5823  uint64x2_t __s1 = __p1; \
5824  uint64x2_t __ret; \
5825  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
5826  __ret; \
5827 })
5828 #else
5829 #define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5830  uint64x2_t __s0 = __p0; \
5831  uint64x2_t __s1 = __p1; \
5832  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5833  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5834  uint64x2_t __ret; \
5835  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
5836  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5837  __ret; \
5838 })
5839 #endif
5840 
5841 #ifdef __LITTLE_ENDIAN__
5842 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5843  uint16x8_t __s0 = __p0; \
5844  uint16x8_t __s1 = __p1; \
5845  uint16x8_t __ret; \
5846  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
5847  __ret; \
5848 })
5849 #else
5850 #define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5851  uint16x8_t __s0 = __p0; \
5852  uint16x8_t __s1 = __p1; \
5853  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5854  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5855  uint16x8_t __ret; \
5856  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
5857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5858  __ret; \
5859 })
5860 #endif
5861 
5862 #ifdef __LITTLE_ENDIAN__
5863 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5864  int8x16_t __s0 = __p0; \
5865  int8x16_t __s1 = __p1; \
5866  int8x16_t __ret; \
5867  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
5868  __ret; \
5869 })
5870 #else
5871 #define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5872  int8x16_t __s0 = __p0; \
5873  int8x16_t __s1 = __p1; \
5874  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5875  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5876  int8x16_t __ret; \
5877  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
5878  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5879  __ret; \
5880 })
5881 #endif
5882 
5883 #ifdef __LITTLE_ENDIAN__
5884 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5885  float32x4_t __s0 = __p0; \
5886  float32x4_t __s1 = __p1; \
5887  float32x4_t __ret; \
5888  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
5889  __ret; \
5890 })
5891 #else
5892 #define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5893  float32x4_t __s0 = __p0; \
5894  float32x4_t __s1 = __p1; \
5895  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5896  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5897  float32x4_t __ret; \
5898  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
5899  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5900  __ret; \
5901 })
5902 #endif
5903 
5904 #ifdef __LITTLE_ENDIAN__
5905 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5906  int32x4_t __s0 = __p0; \
5907  int32x4_t __s1 = __p1; \
5908  int32x4_t __ret; \
5909  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
5910  __ret; \
5911 })
5912 #else
5913 #define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5914  int32x4_t __s0 = __p0; \
5915  int32x4_t __s1 = __p1; \
5916  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5917  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5918  int32x4_t __ret; \
5919  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
5920  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5921  __ret; \
5922 })
5923 #endif
5924 
5925 #ifdef __LITTLE_ENDIAN__
5926 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5927  int64x2_t __s0 = __p0; \
5928  int64x2_t __s1 = __p1; \
5929  int64x2_t __ret; \
5930  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
5931  __ret; \
5932 })
5933 #else
5934 #define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5935  int64x2_t __s0 = __p0; \
5936  int64x2_t __s1 = __p1; \
5937  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5938  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5939  int64x2_t __ret; \
5940  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
5941  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5942  __ret; \
5943 })
5944 #endif
5945 
5946 #ifdef __LITTLE_ENDIAN__
5947 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5948  int16x8_t __s0 = __p0; \
5949  int16x8_t __s1 = __p1; \
5950  int16x8_t __ret; \
5951  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
5952  __ret; \
5953 })
5954 #else
5955 #define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5956  int16x8_t __s0 = __p0; \
5957  int16x8_t __s1 = __p1; \
5958  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5959  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5960  int16x8_t __ret; \
5961  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
5962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5963  __ret; \
5964 })
5965 #endif
5966 
5967 #ifdef __LITTLE_ENDIAN__
5968 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5969  uint8x8_t __s0 = __p0; \
5970  uint8x8_t __s1 = __p1; \
5971  uint8x8_t __ret; \
5972  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
5973  __ret; \
5974 })
5975 #else
5976 #define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5977  uint8x8_t __s0 = __p0; \
5978  uint8x8_t __s1 = __p1; \
5979  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5980  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5981  uint8x8_t __ret; \
5982  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
5983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5984  __ret; \
5985 })
5986 #endif
5987 
5988 #ifdef __LITTLE_ENDIAN__
5989 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5990  uint32x2_t __s0 = __p0; \
5991  uint32x2_t __s1 = __p1; \
5992  uint32x2_t __ret; \
5993  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
5994  __ret; \
5995 })
5996 #else
5997 #define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5998  uint32x2_t __s0 = __p0; \
5999  uint32x2_t __s1 = __p1; \
6000  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6001  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6002  uint32x2_t __ret; \
6003  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
6004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6005  __ret; \
6006 })
6007 #endif
6008 
6009 #ifdef __LITTLE_ENDIAN__
6010 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6011  uint64x1_t __s0 = __p0; \
6012  uint64x1_t __s1 = __p1; \
6013  uint64x1_t __ret; \
6014  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6015  __ret; \
6016 })
6017 #else
6018 #define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6019  uint64x1_t __s0 = __p0; \
6020  uint64x1_t __s1 = __p1; \
6021  uint64x1_t __ret; \
6022  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6023  __ret; \
6024 })
6025 #endif
6026 
6027 #ifdef __LITTLE_ENDIAN__
6028 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6029  uint16x4_t __s0 = __p0; \
6030  uint16x4_t __s1 = __p1; \
6031  uint16x4_t __ret; \
6032  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
6033  __ret; \
6034 })
6035 #else
6036 #define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6037  uint16x4_t __s0 = __p0; \
6038  uint16x4_t __s1 = __p1; \
6039  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6040  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6041  uint16x4_t __ret; \
6042  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
6043  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6044  __ret; \
6045 })
6046 #endif
6047 
6048 #ifdef __LITTLE_ENDIAN__
6049 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6050  int8x8_t __s0 = __p0; \
6051  int8x8_t __s1 = __p1; \
6052  int8x8_t __ret; \
6053  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
6054  __ret; \
6055 })
6056 #else
6057 #define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6058  int8x8_t __s0 = __p0; \
6059  int8x8_t __s1 = __p1; \
6060  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6061  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6062  int8x8_t __ret; \
6063  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
6064  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6065  __ret; \
6066 })
6067 #endif
6068 
6069 #ifdef __LITTLE_ENDIAN__
6070 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6071  float32x2_t __s0 = __p0; \
6072  float32x2_t __s1 = __p1; \
6073  float32x2_t __ret; \
6074  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
6075  __ret; \
6076 })
6077 #else
6078 #define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6079  float32x2_t __s0 = __p0; \
6080  float32x2_t __s1 = __p1; \
6081  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6082  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6083  float32x2_t __ret; \
6084  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
6085  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6086  __ret; \
6087 })
6088 #endif
6089 
6090 #ifdef __LITTLE_ENDIAN__
6091 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6092  int32x2_t __s0 = __p0; \
6093  int32x2_t __s1 = __p1; \
6094  int32x2_t __ret; \
6095  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
6096  __ret; \
6097 })
6098 #else
6099 #define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6100  int32x2_t __s0 = __p0; \
6101  int32x2_t __s1 = __p1; \
6102  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6103  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6104  int32x2_t __ret; \
6105  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
6106  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6107  __ret; \
6108 })
6109 #endif
6110 
6111 #ifdef __LITTLE_ENDIAN__
6112 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6113  int64x1_t __s0 = __p0; \
6114  int64x1_t __s1 = __p1; \
6115  int64x1_t __ret; \
6116  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6117  __ret; \
6118 })
6119 #else
6120 #define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6121  int64x1_t __s0 = __p0; \
6122  int64x1_t __s1 = __p1; \
6123  int64x1_t __ret; \
6124  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6125  __ret; \
6126 })
6127 #endif
6128 
6129 #ifdef __LITTLE_ENDIAN__
6130 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6131  int16x4_t __s0 = __p0; \
6132  int16x4_t __s1 = __p1; \
6133  int16x4_t __ret; \
6134  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
6135  __ret; \
6136 })
6137 #else
6138 #define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6139  int16x4_t __s0 = __p0; \
6140  int16x4_t __s1 = __p1; \
6141  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6142  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6143  int16x4_t __ret; \
6144  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
6145  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6146  __ret; \
6147 })
6148 #endif
6149 
6150 #ifdef __LITTLE_ENDIAN__
6151 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6152  float32x4_t __ret;
6153  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6154  return __ret;
6155 }
6156 #else
6157 __ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6158  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6159  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6160  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
6161  float32x4_t __ret;
6162  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
6163  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6164  return __ret;
6165 }
6166 __ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6167  float32x4_t __ret;
6168  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6169  return __ret;
6170 }
6171 #endif
6172 
6173 #ifdef __LITTLE_ENDIAN__
6174 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6175  float32x2_t __ret;
6176  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6177  return __ret;
6178 }
6179 #else
6180 __ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6181  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6182  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6183  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
6184  float32x2_t __ret;
6185  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
6186  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6187  return __ret;
6188 }
6189 __ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6190  float32x2_t __ret;
6191  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6192  return __ret;
6193 }
6194 #endif
6195 
6196 #ifdef __LITTLE_ENDIAN__
6197 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6198  poly8x8_t __ret;
6199  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6200  return __ret;
6201 }
6202 #else
6203 __ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6204  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6205  poly8x8_t __ret;
6206  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6207  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6208  return __ret;
6209 }
6210 __ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
6211  poly8x8_t __ret;
6212  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6213  return __ret;
6214 }
6215 #endif
6216 
6217 #ifdef __LITTLE_ENDIAN__
6218 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6219  poly16x4_t __ret;
6220  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6221  return __ret;
6222 }
6223 #else
6224 __ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6225  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6226  poly16x4_t __ret;
6227  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6229  return __ret;
6230 }
6231 #endif
6232 
6233 #ifdef __LITTLE_ENDIAN__
6234 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6235  uint8x8_t __ret;
6236  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6237  return __ret;
6238 }
6239 #else
6240 __ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6241  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6242  uint8x8_t __ret;
6243  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6244  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6245  return __ret;
6246 }
6247 __ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
6248  uint8x8_t __ret;
6249  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6250  return __ret;
6251 }
6252 #endif
6253 
6254 #ifdef __LITTLE_ENDIAN__
6255 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6256  uint32x2_t __ret;
6257  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6258  return __ret;
6259 }
6260 #else
6261 __ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6262  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6263  uint32x2_t __ret;
6264  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6266  return __ret;
6267 }
6268 __ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
6269  uint32x2_t __ret;
6270  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6271  return __ret;
6272 }
6273 #endif
6274 
6275 #ifdef __LITTLE_ENDIAN__
6276 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6277  uint64x1_t __ret;
6278  __ret = __builtin_shufflevector(__p0, __p0, 1);
6279  return __ret;
6280 }
6281 #else
6282 __ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6283  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6284  uint64x1_t __ret;
6285  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6286  return __ret;
6287 }
6288 #endif
6289 
6290 #ifdef __LITTLE_ENDIAN__
6291 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6292  uint16x4_t __ret;
6293  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6294  return __ret;
6295 }
6296 #else
6297 __ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6298  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6299  uint16x4_t __ret;
6300  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6301  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6302  return __ret;
6303 }
6304 __ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
6305  uint16x4_t __ret;
6306  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6307  return __ret;
6308 }
6309 #endif
6310 
6311 #ifdef __LITTLE_ENDIAN__
6312 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
6313  int8x8_t __ret;
6314  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6315  return __ret;
6316 }
6317 #else
6318 __ai int8x8_t vget_high_s8(int8x16_t __p0) {
6319  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6320  int8x8_t __ret;
6321  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6322  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6323  return __ret;
6324 }
6325 __ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
6326  int8x8_t __ret;
6327  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6328  return __ret;
6329 }
6330 #endif
6331 
6332 #ifdef __LITTLE_ENDIAN__
6333 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
6334  float32x2_t __ret;
6335  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6336  return __ret;
6337 }
6338 #else
6339 __ai float32x2_t vget_high_f32(float32x4_t __p0) {
6340  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6341  float32x2_t __ret;
6342  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6343  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6344  return __ret;
6345 }
6346 __ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
6347  float32x2_t __ret;
6348  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6349  return __ret;
6350 }
6351 #endif
6352 
6353 #ifdef __LITTLE_ENDIAN__
6354 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
6355  float16x4_t __ret;
6356  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6357  return __ret;
6358 }
6359 #else
6360 __ai float16x4_t vget_high_f16(float16x8_t __p0) {
6361  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6362  float16x4_t __ret;
6363  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6364  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6365  return __ret;
6366 }
6367 __ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
6368  float16x4_t __ret;
6369  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6370  return __ret;
6371 }
6372 #endif
6373 
6374 #ifdef __LITTLE_ENDIAN__
6375 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
6376  int32x2_t __ret;
6377  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6378  return __ret;
6379 }
6380 #else
6381 __ai int32x2_t vget_high_s32(int32x4_t __p0) {
6382  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6383  int32x2_t __ret;
6384  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6386  return __ret;
6387 }
6388 __ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
6389  int32x2_t __ret;
6390  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6391  return __ret;
6392 }
6393 #endif
6394 
6395 #ifdef __LITTLE_ENDIAN__
6396 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
6397  int64x1_t __ret;
6398  __ret = __builtin_shufflevector(__p0, __p0, 1);
6399  return __ret;
6400 }
6401 #else
6402 __ai int64x1_t vget_high_s64(int64x2_t __p0) {
6403  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6404  int64x1_t __ret;
6405  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6406  return __ret;
6407 }
6408 #endif
6409 
6410 #ifdef __LITTLE_ENDIAN__
6411 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
6412  int16x4_t __ret;
6413  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6414  return __ret;
6415 }
6416 #else
6417 __ai int16x4_t vget_high_s16(int16x8_t __p0) {
6418  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6419  int16x4_t __ret;
6420  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6421  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6422  return __ret;
6423 }
6424 __ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
6425  int16x4_t __ret;
6426  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6427  return __ret;
6428 }
6429 #endif
6430 
6431 #ifdef __LITTLE_ENDIAN__
6432 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
6433  poly8x8_t __s0 = __p0; \
6434  poly8_t __ret; \
6435  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6436  __ret; \
6437 })
6438 #else
6439 #define vget_lane_p8(__p0, __p1) __extension__ ({ \
6440  poly8x8_t __s0 = __p0; \
6441  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6442  poly8_t __ret; \
6443  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6444  __ret; \
6445 })
6446 #define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
6447  poly8x8_t __s0 = __p0; \
6448  poly8_t __ret; \
6449  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6450  __ret; \
6451 })
6452 #endif
6453 
6454 #ifdef __LITTLE_ENDIAN__
6455 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
6456  poly16x4_t __s0 = __p0; \
6457  poly16_t __ret; \
6458  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6459  __ret; \
6460 })
6461 #else
6462 #define vget_lane_p16(__p0, __p1) __extension__ ({ \
6463  poly16x4_t __s0 = __p0; \
6464  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6465  poly16_t __ret; \
6466  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6467  __ret; \
6468 })
6469 #define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
6470  poly16x4_t __s0 = __p0; \
6471  poly16_t __ret; \
6472  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6473  __ret; \
6474 })
6475 #endif
6476 
6477 #ifdef __LITTLE_ENDIAN__
6478 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6479  poly8x16_t __s0 = __p0; \
6480  poly8_t __ret; \
6481  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6482  __ret; \
6483 })
6484 #else
6485 #define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6486  poly8x16_t __s0 = __p0; \
6487  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6488  poly8_t __ret; \
6489  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6490  __ret; \
6491 })
6492 #define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6493  poly8x16_t __s0 = __p0; \
6494  poly8_t __ret; \
6495  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6496  __ret; \
6497 })
6498 #endif
6499 
6500 #ifdef __LITTLE_ENDIAN__
6501 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6502  poly16x8_t __s0 = __p0; \
6503  poly16_t __ret; \
6504  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6505  __ret; \
6506 })
6507 #else
6508 #define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6509  poly16x8_t __s0 = __p0; \
6510  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6511  poly16_t __ret; \
6512  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6513  __ret; \
6514 })
6515 #define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6516  poly16x8_t __s0 = __p0; \
6517  poly16_t __ret; \
6518  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6519  __ret; \
6520 })
6521 #endif
6522 
6523 #ifdef __LITTLE_ENDIAN__
6524 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6525  uint8x16_t __s0 = __p0; \
6526  uint8_t __ret; \
6527  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6528  __ret; \
6529 })
6530 #else
6531 #define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6532  uint8x16_t __s0 = __p0; \
6533  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6534  uint8_t __ret; \
6535  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6536  __ret; \
6537 })
6538 #define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6539  uint8x16_t __s0 = __p0; \
6540  uint8_t __ret; \
6541  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6542  __ret; \
6543 })
6544 #endif
6545 
6546 #ifdef __LITTLE_ENDIAN__
6547 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6548  uint32x4_t __s0 = __p0; \
6549  uint32_t __ret; \
6550  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6551  __ret; \
6552 })
6553 #else
6554 #define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6555  uint32x4_t __s0 = __p0; \
6556  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6557  uint32_t __ret; \
6558  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6559  __ret; \
6560 })
6561 #define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6562  uint32x4_t __s0 = __p0; \
6563  uint32_t __ret; \
6564  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6565  __ret; \
6566 })
6567 #endif
6568 
6569 #ifdef __LITTLE_ENDIAN__
6570 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6571  uint64x2_t __s0 = __p0; \
6572  uint64_t __ret; \
6573  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6574  __ret; \
6575 })
6576 #else
6577 #define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6578  uint64x2_t __s0 = __p0; \
6579  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6580  uint64_t __ret; \
6581  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6582  __ret; \
6583 })
6584 #define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6585  uint64x2_t __s0 = __p0; \
6586  uint64_t __ret; \
6587  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6588  __ret; \
6589 })
6590 #endif
6591 
6592 #ifdef __LITTLE_ENDIAN__
6593 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6594  uint16x8_t __s0 = __p0; \
6595  uint16_t __ret; \
6596  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6597  __ret; \
6598 })
6599 #else
6600 #define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6601  uint16x8_t __s0 = __p0; \
6602  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6603  uint16_t __ret; \
6604  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6605  __ret; \
6606 })
6607 #define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6608  uint16x8_t __s0 = __p0; \
6609  uint16_t __ret; \
6610  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6611  __ret; \
6612 })
6613 #endif
6614 
6615 #ifdef __LITTLE_ENDIAN__
6616 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6617  int8x16_t __s0 = __p0; \
6618  int8_t __ret; \
6619  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6620  __ret; \
6621 })
6622 #else
6623 #define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6624  int8x16_t __s0 = __p0; \
6625  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6626  int8_t __ret; \
6627  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6628  __ret; \
6629 })
6630 #define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6631  int8x16_t __s0 = __p0; \
6632  int8_t __ret; \
6633  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6634  __ret; \
6635 })
6636 #endif
6637 
6638 #ifdef __LITTLE_ENDIAN__
6639 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6640  float32x4_t __s0 = __p0; \
6641  float32_t __ret; \
6642  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6643  __ret; \
6644 })
6645 #else
6646 #define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6647  float32x4_t __s0 = __p0; \
6648  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6649  float32_t __ret; \
6650  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__rev0, __p1); \
6651  __ret; \
6652 })
6653 #define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6654  float32x4_t __s0 = __p0; \
6655  float32_t __ret; \
6656  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6657  __ret; \
6658 })
6659 #endif
6660 
6661 #ifdef __LITTLE_ENDIAN__
6662 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6663  int32x4_t __s0 = __p0; \
6664  int32_t __ret; \
6665  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6666  __ret; \
6667 })
6668 #else
6669 #define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6670  int32x4_t __s0 = __p0; \
6671  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6672  int32_t __ret; \
6673  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6674  __ret; \
6675 })
6676 #define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6677  int32x4_t __s0 = __p0; \
6678  int32_t __ret; \
6679  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6680  __ret; \
6681 })
6682 #endif
6683 
6684 #ifdef __LITTLE_ENDIAN__
6685 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6686  int64x2_t __s0 = __p0; \
6687  int64_t __ret; \
6688  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6689  __ret; \
6690 })
6691 #else
6692 #define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6693  int64x2_t __s0 = __p0; \
6694  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6695  int64_t __ret; \
6696  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6697  __ret; \
6698 })
6699 #define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6700  int64x2_t __s0 = __p0; \
6701  int64_t __ret; \
6702  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6703  __ret; \
6704 })
6705 #endif
6706 
6707 #ifdef __LITTLE_ENDIAN__
6708 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6709  int16x8_t __s0 = __p0; \
6710  int16_t __ret; \
6711  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6712  __ret; \
6713 })
6714 #else
6715 #define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6716  int16x8_t __s0 = __p0; \
6717  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6718  int16_t __ret; \
6719  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6720  __ret; \
6721 })
6722 #define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6723  int16x8_t __s0 = __p0; \
6724  int16_t __ret; \
6725  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6726  __ret; \
6727 })
6728 #endif
6729 
6730 #ifdef __LITTLE_ENDIAN__
6731 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
6732  uint8x8_t __s0 = __p0; \
6733  uint8_t __ret; \
6734  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6735  __ret; \
6736 })
6737 #else
6738 #define vget_lane_u8(__p0, __p1) __extension__ ({ \
6739  uint8x8_t __s0 = __p0; \
6740  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6741  uint8_t __ret; \
6742  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6743  __ret; \
6744 })
6745 #define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
6746  uint8x8_t __s0 = __p0; \
6747  uint8_t __ret; \
6748  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6749  __ret; \
6750 })
6751 #endif
6752 
6753 #ifdef __LITTLE_ENDIAN__
6754 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
6755  uint32x2_t __s0 = __p0; \
6756  uint32_t __ret; \
6757  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6758  __ret; \
6759 })
6760 #else
6761 #define vget_lane_u32(__p0, __p1) __extension__ ({ \
6762  uint32x2_t __s0 = __p0; \
6763  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6764  uint32_t __ret; \
6765  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6766  __ret; \
6767 })
6768 #define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
6769  uint32x2_t __s0 = __p0; \
6770  uint32_t __ret; \
6771  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6772  __ret; \
6773 })
6774 #endif
6775 
6776 #ifdef __LITTLE_ENDIAN__
6777 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
6778  uint64x1_t __s0 = __p0; \
6779  uint64_t __ret; \
6780  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6781  __ret; \
6782 })
6783 #else
6784 #define vget_lane_u64(__p0, __p1) __extension__ ({ \
6785  uint64x1_t __s0 = __p0; \
6786  uint64_t __ret; \
6787  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6788  __ret; \
6789 })
6790 #define __noswap_vget_lane_u64(__p0, __p1) __extension__ ({ \
6791  uint64x1_t __s0 = __p0; \
6792  uint64_t __ret; \
6793  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6794  __ret; \
6795 })
6796 #endif
6797 
6798 #ifdef __LITTLE_ENDIAN__
6799 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
6800  uint16x4_t __s0 = __p0; \
6801  uint16_t __ret; \
6802  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6803  __ret; \
6804 })
6805 #else
6806 #define vget_lane_u16(__p0, __p1) __extension__ ({ \
6807  uint16x4_t __s0 = __p0; \
6808  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6809  uint16_t __ret; \
6810  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6811  __ret; \
6812 })
6813 #define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
6814  uint16x4_t __s0 = __p0; \
6815  uint16_t __ret; \
6816  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6817  __ret; \
6818 })
6819 #endif
6820 
6821 #ifdef __LITTLE_ENDIAN__
6822 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
6823  int8x8_t __s0 = __p0; \
6824  int8_t __ret; \
6825  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6826  __ret; \
6827 })
6828 #else
6829 #define vget_lane_s8(__p0, __p1) __extension__ ({ \
6830  int8x8_t __s0 = __p0; \
6831  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6832  int8_t __ret; \
6833  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6834  __ret; \
6835 })
6836 #define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
6837  int8x8_t __s0 = __p0; \
6838  int8_t __ret; \
6839  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6840  __ret; \
6841 })
6842 #endif
6843 
6844 #ifdef __LITTLE_ENDIAN__
6845 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
6846  float32x2_t __s0 = __p0; \
6847  float32_t __ret; \
6848  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6849  __ret; \
6850 })
6851 #else
6852 #define vget_lane_f32(__p0, __p1) __extension__ ({ \
6853  float32x2_t __s0 = __p0; \
6854  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6855  float32_t __ret; \
6856  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__rev0, __p1); \
6857  __ret; \
6858 })
6859 #define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
6860  float32x2_t __s0 = __p0; \
6861  float32_t __ret; \
6862  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6863  __ret; \
6864 })
6865 #endif
6866 
6867 #ifdef __LITTLE_ENDIAN__
6868 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
6869  int32x2_t __s0 = __p0; \
6870  int32_t __ret; \
6871  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6872  __ret; \
6873 })
6874 #else
6875 #define vget_lane_s32(__p0, __p1) __extension__ ({ \
6876  int32x2_t __s0 = __p0; \
6877  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6878  int32_t __ret; \
6879  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6880  __ret; \
6881 })
6882 #define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
6883  int32x2_t __s0 = __p0; \
6884  int32_t __ret; \
6885  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6886  __ret; \
6887 })
6888 #endif
6889 
6890 #ifdef __LITTLE_ENDIAN__
6891 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
6892  int64x1_t __s0 = __p0; \
6893  int64_t __ret; \
6894  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6895  __ret; \
6896 })
6897 #else
6898 #define vget_lane_s64(__p0, __p1) __extension__ ({ \
6899  int64x1_t __s0 = __p0; \
6900  int64_t __ret; \
6901  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6902  __ret; \
6903 })
6904 #define __noswap_vget_lane_s64(__p0, __p1) __extension__ ({ \
6905  int64x1_t __s0 = __p0; \
6906  int64_t __ret; \
6907  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6908  __ret; \
6909 })
6910 #endif
6911 
6912 #ifdef __LITTLE_ENDIAN__
6913 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
6914  int16x4_t __s0 = __p0; \
6915  int16_t __ret; \
6916  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6917  __ret; \
6918 })
6919 #else
6920 #define vget_lane_s16(__p0, __p1) __extension__ ({ \
6921  int16x4_t __s0 = __p0; \
6922  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6923  int16_t __ret; \
6924  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6925  __ret; \
6926 })
6927 #define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
6928  int16x4_t __s0 = __p0; \
6929  int16_t __ret; \
6930  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6931  __ret; \
6932 })
6933 #endif
6934 
6935 #ifdef __LITTLE_ENDIAN__
6936 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6937  poly8x8_t __ret;
6938  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6939  return __ret;
6940 }
6941 #else
6942 __ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6943  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6944  poly8x8_t __ret;
6945  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6947  return __ret;
6948 }
6949 #endif
6950 
6951 #ifdef __LITTLE_ENDIAN__
6952 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6953  poly16x4_t __ret;
6954  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6955  return __ret;
6956 }
6957 #else
6958 __ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6959  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6960  poly16x4_t __ret;
6961  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6962  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6963  return __ret;
6964 }
6965 #endif
6966 
6967 #ifdef __LITTLE_ENDIAN__
6968 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6969  uint8x8_t __ret;
6970  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6971  return __ret;
6972 }
6973 #else
6974 __ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6975  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6976  uint8x8_t __ret;
6977  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6978  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6979  return __ret;
6980 }
6981 #endif
6982 
6983 #ifdef __LITTLE_ENDIAN__
6984 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6985  uint32x2_t __ret;
6986  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6987  return __ret;
6988 }
6989 #else
6990 __ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6991  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6992  uint32x2_t __ret;
6993  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
6994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6995  return __ret;
6996 }
6997 #endif
6998 
6999 #ifdef __LITTLE_ENDIAN__
7000 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7001  uint64x1_t __ret;
7002  __ret = __builtin_shufflevector(__p0, __p0, 0);
7003  return __ret;
7004 }
7005 #else
7006 __ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7007  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7008  uint64x1_t __ret;
7009  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7010  return __ret;
7011 }
7012 #endif
7013 
7014 #ifdef __LITTLE_ENDIAN__
7015 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7016  uint16x4_t __ret;
7017  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7018  return __ret;
7019 }
7020 #else
7021 __ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7022  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7023  uint16x4_t __ret;
7024  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7025  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7026  return __ret;
7027 }
7028 #endif
7029 
7030 #ifdef __LITTLE_ENDIAN__
7031 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7032  int8x8_t __ret;
7033  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7034  return __ret;
7035 }
7036 #else
7037 __ai int8x8_t vget_low_s8(int8x16_t __p0) {
7038  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7039  int8x8_t __ret;
7040  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7042  return __ret;
7043 }
7044 #endif
7045 
7046 #ifdef __LITTLE_ENDIAN__
7047 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7048  float32x2_t __ret;
7049  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7050  return __ret;
7051 }
7052 #else
7053 __ai float32x2_t vget_low_f32(float32x4_t __p0) {
7054  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7055  float32x2_t __ret;
7056  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7058  return __ret;
7059 }
7060 #endif
7061 
7062 #ifdef __LITTLE_ENDIAN__
7063 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7064  float16x4_t __ret;
7065  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7066  return __ret;
7067 }
7068 #else
7069 __ai float16x4_t vget_low_f16(float16x8_t __p0) {
7070  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7071  float16x4_t __ret;
7072  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7073  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7074  return __ret;
7075 }
7076 #endif
7077 
7078 #ifdef __LITTLE_ENDIAN__
7079 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
7080  int32x2_t __ret;
7081  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7082  return __ret;
7083 }
7084 #else
7085 __ai int32x2_t vget_low_s32(int32x4_t __p0) {
7086  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7087  int32x2_t __ret;
7088  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7089  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7090  return __ret;
7091 }
7092 #endif
7093 
7094 #ifdef __LITTLE_ENDIAN__
7095 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
7096  int64x1_t __ret;
7097  __ret = __builtin_shufflevector(__p0, __p0, 0);
7098  return __ret;
7099 }
7100 #else
7101 __ai int64x1_t vget_low_s64(int64x2_t __p0) {
7102  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7103  int64x1_t __ret;
7104  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7105  return __ret;
7106 }
7107 #endif
7108 
7109 #ifdef __LITTLE_ENDIAN__
7110 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
7111  int16x4_t __ret;
7112  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7113  return __ret;
7114 }
7115 #else
7116 __ai int16x4_t vget_low_s16(int16x8_t __p0) {
7117  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7118  int16x4_t __ret;
7119  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7120  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7121  return __ret;
7122 }
7123 #endif
7124 
7125 #ifdef __LITTLE_ENDIAN__
7126 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7127  uint8x16_t __ret;
7128  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7129  return __ret;
7130 }
7131 #else
7132 __ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7133  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7134  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7135  uint8x16_t __ret;
7136  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7137  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7138  return __ret;
7139 }
7140 #endif
7141 
7142 #ifdef __LITTLE_ENDIAN__
7143 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7144  uint32x4_t __ret;
7145  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7146  return __ret;
7147 }
7148 #else
7149 __ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7150  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7151  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7152  uint32x4_t __ret;
7153  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7154  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7155  return __ret;
7156 }
7157 #endif
7158 
7159 #ifdef __LITTLE_ENDIAN__
7160 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7161  uint16x8_t __ret;
7162  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7163  return __ret;
7164 }
7165 #else
7166 __ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7167  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7168  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7169  uint16x8_t __ret;
7170  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7171  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7172  return __ret;
7173 }
7174 #endif
7175 
7176 #ifdef __LITTLE_ENDIAN__
7177 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7178  int8x16_t __ret;
7179  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7180  return __ret;
7181 }
7182 #else
7183 __ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7184  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7185  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7186  int8x16_t __ret;
7187  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7188  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7189  return __ret;
7190 }
7191 #endif
7192 
7193 #ifdef __LITTLE_ENDIAN__
7194 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7195  int32x4_t __ret;
7196  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7197  return __ret;
7198 }
7199 #else
7200 __ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7201  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7202  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7203  int32x4_t __ret;
7204  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7206  return __ret;
7207 }
7208 #endif
7209 
7210 #ifdef __LITTLE_ENDIAN__
7211 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7212  int16x8_t __ret;
7213  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7214  return __ret;
7215 }
7216 #else
7217 __ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7218  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7219  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7220  int16x8_t __ret;
7221  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7222  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7223  return __ret;
7224 }
7225 #endif
7226 
7227 #ifdef __LITTLE_ENDIAN__
7228 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7229  uint8x8_t __ret;
7230  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7231  return __ret;
7232 }
7233 #else
7234 __ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7235  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7236  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7237  uint8x8_t __ret;
7238  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7240  return __ret;
7241 }
7242 #endif
7243 
7244 #ifdef __LITTLE_ENDIAN__
7245 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7246  uint32x2_t __ret;
7247  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7248  return __ret;
7249 }
7250 #else
7251 __ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7252  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7253  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7254  uint32x2_t __ret;
7255  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7256  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7257  return __ret;
7258 }
7259 #endif
7260 
7261 #ifdef __LITTLE_ENDIAN__
7262 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7263  uint16x4_t __ret;
7264  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7265  return __ret;
7266 }
7267 #else
7268 __ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7269  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7270  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7271  uint16x4_t __ret;
7272  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7273  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7274  return __ret;
7275 }
7276 #endif
7277 
7278 #ifdef __LITTLE_ENDIAN__
7279 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7280  int8x8_t __ret;
7281  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7282  return __ret;
7283 }
7284 #else
7285 __ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7286  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7287  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7288  int8x8_t __ret;
7289  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7290  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7291  return __ret;
7292 }
7293 #endif
7294 
7295 #ifdef __LITTLE_ENDIAN__
7296 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7297  int32x2_t __ret;
7298  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7299  return __ret;
7300 }
7301 #else
7302 __ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7303  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7304  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7305  int32x2_t __ret;
7306  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7308  return __ret;
7309 }
7310 #endif
7311 
7312 #ifdef __LITTLE_ENDIAN__
7313 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7314  int16x4_t __ret;
7315  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7316  return __ret;
7317 }
7318 #else
7319 __ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7320  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7321  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7322  int16x4_t __ret;
7323  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7325  return __ret;
7326 }
7327 #endif
7328 
7329 #ifdef __LITTLE_ENDIAN__
7330 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7331  uint8x16_t __ret;
7332  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7333  return __ret;
7334 }
7335 #else
7336 __ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7337  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7338  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7339  uint8x16_t __ret;
7340  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7341  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7342  return __ret;
7343 }
7344 #endif
7345 
7346 #ifdef __LITTLE_ENDIAN__
7347 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7348  uint32x4_t __ret;
7349  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7350  return __ret;
7351 }
7352 #else
7353 __ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7354  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7355  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7356  uint32x4_t __ret;
7357  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7358  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7359  return __ret;
7360 }
7361 #endif
7362 
7363 #ifdef __LITTLE_ENDIAN__
7364 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7365  uint16x8_t __ret;
7366  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7367  return __ret;
7368 }
7369 #else
7370 __ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7371  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7372  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7373  uint16x8_t __ret;
7374  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7376  return __ret;
7377 }
7378 #endif
7379 
7380 #ifdef __LITTLE_ENDIAN__
7381 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7382  int8x16_t __ret;
7383  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7384  return __ret;
7385 }
7386 #else
7387 __ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7388  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7389  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7390  int8x16_t __ret;
7391  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7393  return __ret;
7394 }
7395 #endif
7396 
7397 #ifdef __LITTLE_ENDIAN__
7398 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7399  int32x4_t __ret;
7400  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7401  return __ret;
7402 }
7403 #else
7404 __ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7405  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7406  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7407  int32x4_t __ret;
7408  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7410  return __ret;
7411 }
7412 #endif
7413 
7414 #ifdef __LITTLE_ENDIAN__
7415 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7416  int16x8_t __ret;
7417  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7418  return __ret;
7419 }
7420 #else
7421 __ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7422  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7423  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7424  int16x8_t __ret;
7425  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7427  return __ret;
7428 }
7429 #endif
7430 
7431 #ifdef __LITTLE_ENDIAN__
7432 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7433  uint8x8_t __ret;
7434  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7435  return __ret;
7436 }
7437 #else
7438 __ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7439  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7440  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7441  uint8x8_t __ret;
7442  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7443  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7444  return __ret;
7445 }
7446 #endif
7447 
7448 #ifdef __LITTLE_ENDIAN__
7449 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7450  uint32x2_t __ret;
7451  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7452  return __ret;
7453 }
7454 #else
7455 __ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7456  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7457  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7458  uint32x2_t __ret;
7459  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7460  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7461  return __ret;
7462 }
7463 #endif
7464 
7465 #ifdef __LITTLE_ENDIAN__
7466 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7467  uint16x4_t __ret;
7468  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7469  return __ret;
7470 }
7471 #else
7472 __ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7473  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7474  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7475  uint16x4_t __ret;
7476  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7478  return __ret;
7479 }
7480 #endif
7481 
7482 #ifdef __LITTLE_ENDIAN__
7483 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7484  int8x8_t __ret;
7485  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7486  return __ret;
7487 }
7488 #else
7489 __ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7490  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7491  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7492  int8x8_t __ret;
7493  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7495  return __ret;
7496 }
7497 #endif
7498 
7499 #ifdef __LITTLE_ENDIAN__
7500 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7501  int32x2_t __ret;
7502  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7503  return __ret;
7504 }
7505 #else
7506 __ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7507  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7508  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7509  int32x2_t __ret;
7510  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7511  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7512  return __ret;
7513 }
7514 #endif
7515 
7516 #ifdef __LITTLE_ENDIAN__
7517 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7518  int16x4_t __ret;
7519  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7520  return __ret;
7521 }
7522 #else
7523 __ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7524  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7525  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7526  int16x4_t __ret;
7527  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7528  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7529  return __ret;
7530 }
7531 #endif
7532 
7533 #ifdef __LITTLE_ENDIAN__
7534 #define vld1_p8(__p0) __extension__ ({ \
7535  poly8x8_t __ret; \
7536  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7537  __ret; \
7538 })
7539 #else
7540 #define vld1_p8(__p0) __extension__ ({ \
7541  poly8x8_t __ret; \
7542  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7543  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7544  __ret; \
7545 })
7546 #endif
7547 
7548 #ifdef __LITTLE_ENDIAN__
7549 #define vld1_p16(__p0) __extension__ ({ \
7550  poly16x4_t __ret; \
7551  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7552  __ret; \
7553 })
7554 #else
7555 #define vld1_p16(__p0) __extension__ ({ \
7556  poly16x4_t __ret; \
7557  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7558  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7559  __ret; \
7560 })
7561 #endif
7562 
7563 #ifdef __LITTLE_ENDIAN__
7564 #define vld1q_p8(__p0) __extension__ ({ \
7565  poly8x16_t __ret; \
7566  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7567  __ret; \
7568 })
7569 #else
7570 #define vld1q_p8(__p0) __extension__ ({ \
7571  poly8x16_t __ret; \
7572  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7573  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7574  __ret; \
7575 })
7576 #endif
7577 
7578 #ifdef __LITTLE_ENDIAN__
7579 #define vld1q_p16(__p0) __extension__ ({ \
7580  poly16x8_t __ret; \
7581  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7582  __ret; \
7583 })
7584 #else
7585 #define vld1q_p16(__p0) __extension__ ({ \
7586  poly16x8_t __ret; \
7587  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7588  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7589  __ret; \
7590 })
7591 #endif
7592 
7593 #ifdef __LITTLE_ENDIAN__
7594 #define vld1q_u8(__p0) __extension__ ({ \
7595  uint8x16_t __ret; \
7596  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7597  __ret; \
7598 })
7599 #else
7600 #define vld1q_u8(__p0) __extension__ ({ \
7601  uint8x16_t __ret; \
7602  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7603  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7604  __ret; \
7605 })
7606 #endif
7607 
7608 #ifdef __LITTLE_ENDIAN__
7609 #define vld1q_u32(__p0) __extension__ ({ \
7610  uint32x4_t __ret; \
7611  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7612  __ret; \
7613 })
7614 #else
7615 #define vld1q_u32(__p0) __extension__ ({ \
7616  uint32x4_t __ret; \
7617  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7618  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7619  __ret; \
7620 })
7621 #endif
7622 
7623 #ifdef __LITTLE_ENDIAN__
7624 #define vld1q_u64(__p0) __extension__ ({ \
7625  uint64x2_t __ret; \
7626  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7627  __ret; \
7628 })
7629 #else
7630 #define vld1q_u64(__p0) __extension__ ({ \
7631  uint64x2_t __ret; \
7632  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7633  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7634  __ret; \
7635 })
7636 #endif
7637 
7638 #ifdef __LITTLE_ENDIAN__
7639 #define vld1q_u16(__p0) __extension__ ({ \
7640  uint16x8_t __ret; \
7641  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7642  __ret; \
7643 })
7644 #else
7645 #define vld1q_u16(__p0) __extension__ ({ \
7646  uint16x8_t __ret; \
7647  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7648  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7649  __ret; \
7650 })
7651 #endif
7652 
7653 #ifdef __LITTLE_ENDIAN__
7654 #define vld1q_s8(__p0) __extension__ ({ \
7655  int8x16_t __ret; \
7656  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7657  __ret; \
7658 })
7659 #else
7660 #define vld1q_s8(__p0) __extension__ ({ \
7661  int8x16_t __ret; \
7662  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7663  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7664  __ret; \
7665 })
7666 #endif
7667 
7668 #ifdef __LITTLE_ENDIAN__
7669 #define vld1q_f32(__p0) __extension__ ({ \
7670  float32x4_t __ret; \
7671  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7672  __ret; \
7673 })
7674 #else
7675 #define vld1q_f32(__p0) __extension__ ({ \
7676  float32x4_t __ret; \
7677  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7678  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7679  __ret; \
7680 })
7681 #endif
7682 
7683 #ifdef __LITTLE_ENDIAN__
7684 #define vld1q_f16(__p0) __extension__ ({ \
7685  float16x8_t __ret; \
7686  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7687  __ret; \
7688 })
7689 #else
7690 #define vld1q_f16(__p0) __extension__ ({ \
7691  float16x8_t __ret; \
7692  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7693  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7694  __ret; \
7695 })
7696 #endif
7697 
7698 #ifdef __LITTLE_ENDIAN__
7699 #define vld1q_s32(__p0) __extension__ ({ \
7700  int32x4_t __ret; \
7701  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7702  __ret; \
7703 })
7704 #else
7705 #define vld1q_s32(__p0) __extension__ ({ \
7706  int32x4_t __ret; \
7707  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7708  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7709  __ret; \
7710 })
7711 #endif
7712 
7713 #ifdef __LITTLE_ENDIAN__
7714 #define vld1q_s64(__p0) __extension__ ({ \
7715  int64x2_t __ret; \
7716  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7717  __ret; \
7718 })
7719 #else
7720 #define vld1q_s64(__p0) __extension__ ({ \
7721  int64x2_t __ret; \
7722  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7723  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7724  __ret; \
7725 })
7726 #endif
7727 
7728 #ifdef __LITTLE_ENDIAN__
7729 #define vld1q_s16(__p0) __extension__ ({ \
7730  int16x8_t __ret; \
7731  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7732  __ret; \
7733 })
7734 #else
7735 #define vld1q_s16(__p0) __extension__ ({ \
7736  int16x8_t __ret; \
7737  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7738  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7739  __ret; \
7740 })
7741 #endif
7742 
7743 #ifdef __LITTLE_ENDIAN__
7744 #define vld1_u8(__p0) __extension__ ({ \
7745  uint8x8_t __ret; \
7746  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7747  __ret; \
7748 })
7749 #else
7750 #define vld1_u8(__p0) __extension__ ({ \
7751  uint8x8_t __ret; \
7752  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7753  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7754  __ret; \
7755 })
7756 #endif
7757 
7758 #ifdef __LITTLE_ENDIAN__
7759 #define vld1_u32(__p0) __extension__ ({ \
7760  uint32x2_t __ret; \
7761  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7762  __ret; \
7763 })
7764 #else
7765 #define vld1_u32(__p0) __extension__ ({ \
7766  uint32x2_t __ret; \
7767  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7768  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7769  __ret; \
7770 })
7771 #endif
7772 
7773 #ifdef __LITTLE_ENDIAN__
7774 #define vld1_u64(__p0) __extension__ ({ \
7775  uint64x1_t __ret; \
7776  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7777  __ret; \
7778 })
7779 #else
7780 #define vld1_u64(__p0) __extension__ ({ \
7781  uint64x1_t __ret; \
7782  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7783  __ret; \
7784 })
7785 #endif
7786 
7787 #ifdef __LITTLE_ENDIAN__
7788 #define vld1_u16(__p0) __extension__ ({ \
7789  uint16x4_t __ret; \
7790  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7791  __ret; \
7792 })
7793 #else
7794 #define vld1_u16(__p0) __extension__ ({ \
7795  uint16x4_t __ret; \
7796  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7798  __ret; \
7799 })
7800 #endif
7801 
7802 #ifdef __LITTLE_ENDIAN__
7803 #define vld1_s8(__p0) __extension__ ({ \
7804  int8x8_t __ret; \
7805  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7806  __ret; \
7807 })
7808 #else
7809 #define vld1_s8(__p0) __extension__ ({ \
7810  int8x8_t __ret; \
7811  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7812  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7813  __ret; \
7814 })
7815 #endif
7816 
7817 #ifdef __LITTLE_ENDIAN__
7818 #define vld1_f32(__p0) __extension__ ({ \
7819  float32x2_t __ret; \
7820  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7821  __ret; \
7822 })
7823 #else
7824 #define vld1_f32(__p0) __extension__ ({ \
7825  float32x2_t __ret; \
7826  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7828  __ret; \
7829 })
7830 #endif
7831 
7832 #ifdef __LITTLE_ENDIAN__
7833 #define vld1_f16(__p0) __extension__ ({ \
7834  float16x4_t __ret; \
7835  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7836  __ret; \
7837 })
7838 #else
7839 #define vld1_f16(__p0) __extension__ ({ \
7840  float16x4_t __ret; \
7841  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7843  __ret; \
7844 })
7845 #endif
7846 
7847 #ifdef __LITTLE_ENDIAN__
7848 #define vld1_s32(__p0) __extension__ ({ \
7849  int32x2_t __ret; \
7850  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7851  __ret; \
7852 })
7853 #else
7854 #define vld1_s32(__p0) __extension__ ({ \
7855  int32x2_t __ret; \
7856  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7857  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7858  __ret; \
7859 })
7860 #endif
7861 
7862 #ifdef __LITTLE_ENDIAN__
7863 #define vld1_s64(__p0) __extension__ ({ \
7864  int64x1_t __ret; \
7865  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7866  __ret; \
7867 })
7868 #else
7869 #define vld1_s64(__p0) __extension__ ({ \
7870  int64x1_t __ret; \
7871  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7872  __ret; \
7873 })
7874 #endif
7875 
7876 #ifdef __LITTLE_ENDIAN__
7877 #define vld1_s16(__p0) __extension__ ({ \
7878  int16x4_t __ret; \
7879  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7880  __ret; \
7881 })
7882 #else
7883 #define vld1_s16(__p0) __extension__ ({ \
7884  int16x4_t __ret; \
7885  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7886  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7887  __ret; \
7888 })
7889 #endif
7890 
7891 #ifdef __LITTLE_ENDIAN__
7892 #define vld1_dup_p8(__p0) __extension__ ({ \
7893  poly8x8_t __ret; \
7894  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7895  __ret; \
7896 })
7897 #else
7898 #define vld1_dup_p8(__p0) __extension__ ({ \
7899  poly8x8_t __ret; \
7900  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7902  __ret; \
7903 })
7904 #endif
7905 
7906 #ifdef __LITTLE_ENDIAN__
7907 #define vld1_dup_p16(__p0) __extension__ ({ \
7908  poly16x4_t __ret; \
7909  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7910  __ret; \
7911 })
7912 #else
7913 #define vld1_dup_p16(__p0) __extension__ ({ \
7914  poly16x4_t __ret; \
7915  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7917  __ret; \
7918 })
7919 #endif
7920 
7921 #ifdef __LITTLE_ENDIAN__
7922 #define vld1q_dup_p8(__p0) __extension__ ({ \
7923  poly8x16_t __ret; \
7924  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7925  __ret; \
7926 })
7927 #else
7928 #define vld1q_dup_p8(__p0) __extension__ ({ \
7929  poly8x16_t __ret; \
7930  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7931  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7932  __ret; \
7933 })
7934 #endif
7935 
7936 #ifdef __LITTLE_ENDIAN__
7937 #define vld1q_dup_p16(__p0) __extension__ ({ \
7938  poly16x8_t __ret; \
7939  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7940  __ret; \
7941 })
7942 #else
7943 #define vld1q_dup_p16(__p0) __extension__ ({ \
7944  poly16x8_t __ret; \
7945  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7947  __ret; \
7948 })
7949 #endif
7950 
7951 #ifdef __LITTLE_ENDIAN__
7952 #define vld1q_dup_u8(__p0) __extension__ ({ \
7953  uint8x16_t __ret; \
7954  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7955  __ret; \
7956 })
7957 #else
7958 #define vld1q_dup_u8(__p0) __extension__ ({ \
7959  uint8x16_t __ret; \
7960  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7961  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7962  __ret; \
7963 })
7964 #endif
7965 
7966 #ifdef __LITTLE_ENDIAN__
7967 #define vld1q_dup_u32(__p0) __extension__ ({ \
7968  uint32x4_t __ret; \
7969  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7970  __ret; \
7971 })
7972 #else
7973 #define vld1q_dup_u32(__p0) __extension__ ({ \
7974  uint32x4_t __ret; \
7975  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7976  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7977  __ret; \
7978 })
7979 #endif
7980 
7981 #ifdef __LITTLE_ENDIAN__
7982 #define vld1q_dup_u64(__p0) __extension__ ({ \
7983  uint64x2_t __ret; \
7984  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7985  __ret; \
7986 })
7987 #else
7988 #define vld1q_dup_u64(__p0) __extension__ ({ \
7989  uint64x2_t __ret; \
7990  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7991  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7992  __ret; \
7993 })
7994 #endif
7995 
7996 #ifdef __LITTLE_ENDIAN__
7997 #define vld1q_dup_u16(__p0) __extension__ ({ \
7998  uint16x8_t __ret; \
7999  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8000  __ret; \
8001 })
8002 #else
8003 #define vld1q_dup_u16(__p0) __extension__ ({ \
8004  uint16x8_t __ret; \
8005  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8007  __ret; \
8008 })
8009 #endif
8010 
8011 #ifdef __LITTLE_ENDIAN__
8012 #define vld1q_dup_s8(__p0) __extension__ ({ \
8013  int8x16_t __ret; \
8014  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8015  __ret; \
8016 })
8017 #else
8018 #define vld1q_dup_s8(__p0) __extension__ ({ \
8019  int8x16_t __ret; \
8020  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8021  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8022  __ret; \
8023 })
8024 #endif
8025 
8026 #ifdef __LITTLE_ENDIAN__
8027 #define vld1q_dup_f32(__p0) __extension__ ({ \
8028  float32x4_t __ret; \
8029  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8030  __ret; \
8031 })
8032 #else
8033 #define vld1q_dup_f32(__p0) __extension__ ({ \
8034  float32x4_t __ret; \
8035  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8036  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8037  __ret; \
8038 })
8039 #endif
8040 
8041 #ifdef __LITTLE_ENDIAN__
8042 #define vld1q_dup_f16(__p0) __extension__ ({ \
8043  float16x8_t __ret; \
8044  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8045  __ret; \
8046 })
8047 #else
8048 #define vld1q_dup_f16(__p0) __extension__ ({ \
8049  float16x8_t __ret; \
8050  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8052  __ret; \
8053 })
8054 #endif
8055 
8056 #ifdef __LITTLE_ENDIAN__
8057 #define vld1q_dup_s32(__p0) __extension__ ({ \
8058  int32x4_t __ret; \
8059  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8060  __ret; \
8061 })
8062 #else
8063 #define vld1q_dup_s32(__p0) __extension__ ({ \
8064  int32x4_t __ret; \
8065  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8066  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8067  __ret; \
8068 })
8069 #endif
8070 
8071 #ifdef __LITTLE_ENDIAN__
8072 #define vld1q_dup_s64(__p0) __extension__ ({ \
8073  int64x2_t __ret; \
8074  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8075  __ret; \
8076 })
8077 #else
8078 #define vld1q_dup_s64(__p0) __extension__ ({ \
8079  int64x2_t __ret; \
8080  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8081  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8082  __ret; \
8083 })
8084 #endif
8085 
8086 #ifdef __LITTLE_ENDIAN__
8087 #define vld1q_dup_s16(__p0) __extension__ ({ \
8088  int16x8_t __ret; \
8089  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8090  __ret; \
8091 })
8092 #else
8093 #define vld1q_dup_s16(__p0) __extension__ ({ \
8094  int16x8_t __ret; \
8095  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8097  __ret; \
8098 })
8099 #endif
8100 
8101 #ifdef __LITTLE_ENDIAN__
8102 #define vld1_dup_u8(__p0) __extension__ ({ \
8103  uint8x8_t __ret; \
8104  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8105  __ret; \
8106 })
8107 #else
8108 #define vld1_dup_u8(__p0) __extension__ ({ \
8109  uint8x8_t __ret; \
8110  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8111  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8112  __ret; \
8113 })
8114 #endif
8115 
8116 #ifdef __LITTLE_ENDIAN__
8117 #define vld1_dup_u32(__p0) __extension__ ({ \
8118  uint32x2_t __ret; \
8119  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8120  __ret; \
8121 })
8122 #else
8123 #define vld1_dup_u32(__p0) __extension__ ({ \
8124  uint32x2_t __ret; \
8125  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8127  __ret; \
8128 })
8129 #endif
8130 
8131 #ifdef __LITTLE_ENDIAN__
8132 #define vld1_dup_u64(__p0) __extension__ ({ \
8133  uint64x1_t __ret; \
8134  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8135  __ret; \
8136 })
8137 #else
8138 #define vld1_dup_u64(__p0) __extension__ ({ \
8139  uint64x1_t __ret; \
8140  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8141  __ret; \
8142 })
8143 #endif
8144 
8145 #ifdef __LITTLE_ENDIAN__
8146 #define vld1_dup_u16(__p0) __extension__ ({ \
8147  uint16x4_t __ret; \
8148  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8149  __ret; \
8150 })
8151 #else
8152 #define vld1_dup_u16(__p0) __extension__ ({ \
8153  uint16x4_t __ret; \
8154  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8156  __ret; \
8157 })
8158 #endif
8159 
8160 #ifdef __LITTLE_ENDIAN__
8161 #define vld1_dup_s8(__p0) __extension__ ({ \
8162  int8x8_t __ret; \
8163  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8164  __ret; \
8165 })
8166 #else
8167 #define vld1_dup_s8(__p0) __extension__ ({ \
8168  int8x8_t __ret; \
8169  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8170  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8171  __ret; \
8172 })
8173 #endif
8174 
8175 #ifdef __LITTLE_ENDIAN__
8176 #define vld1_dup_f32(__p0) __extension__ ({ \
8177  float32x2_t __ret; \
8178  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8179  __ret; \
8180 })
8181 #else
8182 #define vld1_dup_f32(__p0) __extension__ ({ \
8183  float32x2_t __ret; \
8184  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8185  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8186  __ret; \
8187 })
8188 #endif
8189 
8190 #ifdef __LITTLE_ENDIAN__
8191 #define vld1_dup_f16(__p0) __extension__ ({ \
8192  float16x4_t __ret; \
8193  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8194  __ret; \
8195 })
8196 #else
8197 #define vld1_dup_f16(__p0) __extension__ ({ \
8198  float16x4_t __ret; \
8199  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8201  __ret; \
8202 })
8203 #endif
8204 
8205 #ifdef __LITTLE_ENDIAN__
8206 #define vld1_dup_s32(__p0) __extension__ ({ \
8207  int32x2_t __ret; \
8208  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8209  __ret; \
8210 })
8211 #else
8212 #define vld1_dup_s32(__p0) __extension__ ({ \
8213  int32x2_t __ret; \
8214  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8215  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8216  __ret; \
8217 })
8218 #endif
8219 
8220 #ifdef __LITTLE_ENDIAN__
8221 #define vld1_dup_s64(__p0) __extension__ ({ \
8222  int64x1_t __ret; \
8223  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8224  __ret; \
8225 })
8226 #else
8227 #define vld1_dup_s64(__p0) __extension__ ({ \
8228  int64x1_t __ret; \
8229  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8230  __ret; \
8231 })
8232 #endif
8233 
8234 #ifdef __LITTLE_ENDIAN__
8235 #define vld1_dup_s16(__p0) __extension__ ({ \
8236  int16x4_t __ret; \
8237  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8238  __ret; \
8239 })
8240 #else
8241 #define vld1_dup_s16(__p0) __extension__ ({ \
8242  int16x4_t __ret; \
8243  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8244  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8245  __ret; \
8246 })
8247 #endif
8248 
8249 #ifdef __LITTLE_ENDIAN__
8250 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8251  poly8x8_t __s1 = __p1; \
8252  poly8x8_t __ret; \
8253  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
8254  __ret; \
8255 })
8256 #else
8257 #define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8258  poly8x8_t __s1 = __p1; \
8259  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8260  poly8x8_t __ret; \
8261  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
8262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8263  __ret; \
8264 })
8265 #endif
8266 
8267 #ifdef __LITTLE_ENDIAN__
8268 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8269  poly16x4_t __s1 = __p1; \
8270  poly16x4_t __ret; \
8271  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
8272  __ret; \
8273 })
8274 #else
8275 #define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8276  poly16x4_t __s1 = __p1; \
8277  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8278  poly16x4_t __ret; \
8279  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
8280  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8281  __ret; \
8282 })
8283 #endif
8284 
8285 #ifdef __LITTLE_ENDIAN__
8286 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8287  poly8x16_t __s1 = __p1; \
8288  poly8x16_t __ret; \
8289  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
8290  __ret; \
8291 })
8292 #else
8293 #define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8294  poly8x16_t __s1 = __p1; \
8295  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8296  poly8x16_t __ret; \
8297  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
8298  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8299  __ret; \
8300 })
8301 #endif
8302 
8303 #ifdef __LITTLE_ENDIAN__
8304 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8305  poly16x8_t __s1 = __p1; \
8306  poly16x8_t __ret; \
8307  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
8308  __ret; \
8309 })
8310 #else
8311 #define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8312  poly16x8_t __s1 = __p1; \
8313  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8314  poly16x8_t __ret; \
8315  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
8316  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8317  __ret; \
8318 })
8319 #endif
8320 
8321 #ifdef __LITTLE_ENDIAN__
8322 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8323  uint8x16_t __s1 = __p1; \
8324  uint8x16_t __ret; \
8325  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
8326  __ret; \
8327 })
8328 #else
8329 #define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8330  uint8x16_t __s1 = __p1; \
8331  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8332  uint8x16_t __ret; \
8333  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
8334  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8335  __ret; \
8336 })
8337 #endif
8338 
8339 #ifdef __LITTLE_ENDIAN__
8340 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8341  uint32x4_t __s1 = __p1; \
8342  uint32x4_t __ret; \
8343  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
8344  __ret; \
8345 })
8346 #else
8347 #define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8348  uint32x4_t __s1 = __p1; \
8349  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8350  uint32x4_t __ret; \
8351  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
8352  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8353  __ret; \
8354 })
8355 #endif
8356 
8357 #ifdef __LITTLE_ENDIAN__
8358 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8359  uint64x2_t __s1 = __p1; \
8360  uint64x2_t __ret; \
8361  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
8362  __ret; \
8363 })
8364 #else
8365 #define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8366  uint64x2_t __s1 = __p1; \
8367  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8368  uint64x2_t __ret; \
8369  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
8370  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8371  __ret; \
8372 })
8373 #endif
8374 
8375 #ifdef __LITTLE_ENDIAN__
8376 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8377  uint16x8_t __s1 = __p1; \
8378  uint16x8_t __ret; \
8379  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
8380  __ret; \
8381 })
8382 #else
8383 #define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8384  uint16x8_t __s1 = __p1; \
8385  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8386  uint16x8_t __ret; \
8387  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
8388  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8389  __ret; \
8390 })
8391 #endif
8392 
8393 #ifdef __LITTLE_ENDIAN__
8394 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8395  int8x16_t __s1 = __p1; \
8396  int8x16_t __ret; \
8397  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
8398  __ret; \
8399 })
8400 #else
8401 #define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8402  int8x16_t __s1 = __p1; \
8403  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8404  int8x16_t __ret; \
8405  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
8406  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8407  __ret; \
8408 })
8409 #endif
8410 
8411 #ifdef __LITTLE_ENDIAN__
8412 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8413  float32x4_t __s1 = __p1; \
8414  float32x4_t __ret; \
8415  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
8416  __ret; \
8417 })
8418 #else
8419 #define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8420  float32x4_t __s1 = __p1; \
8421  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8422  float32x4_t __ret; \
8423  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
8424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8425  __ret; \
8426 })
8427 #endif
8428 
8429 #ifdef __LITTLE_ENDIAN__
8430 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8431  float16x8_t __s1 = __p1; \
8432  float16x8_t __ret; \
8433  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
8434  __ret; \
8435 })
8436 #else
8437 #define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8438  float16x8_t __s1 = __p1; \
8439  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8440  float16x8_t __ret; \
8441  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
8442  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8443  __ret; \
8444 })
8445 #endif
8446 
8447 #ifdef __LITTLE_ENDIAN__
8448 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8449  int32x4_t __s1 = __p1; \
8450  int32x4_t __ret; \
8451  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
8452  __ret; \
8453 })
8454 #else
8455 #define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8456  int32x4_t __s1 = __p1; \
8457  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8458  int32x4_t __ret; \
8459  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
8460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8461  __ret; \
8462 })
8463 #endif
8464 
8465 #ifdef __LITTLE_ENDIAN__
8466 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8467  int64x2_t __s1 = __p1; \
8468  int64x2_t __ret; \
8469  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
8470  __ret; \
8471 })
8472 #else
8473 #define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8474  int64x2_t __s1 = __p1; \
8475  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8476  int64x2_t __ret; \
8477  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
8478  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8479  __ret; \
8480 })
8481 #endif
8482 
8483 #ifdef __LITTLE_ENDIAN__
8484 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8485  int16x8_t __s1 = __p1; \
8486  int16x8_t __ret; \
8487  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
8488  __ret; \
8489 })
8490 #else
8491 #define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8492  int16x8_t __s1 = __p1; \
8493  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8494  int16x8_t __ret; \
8495  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
8496  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8497  __ret; \
8498 })
8499 #endif
8500 
8501 #ifdef __LITTLE_ENDIAN__
8502 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8503  uint8x8_t __s1 = __p1; \
8504  uint8x8_t __ret; \
8505  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
8506  __ret; \
8507 })
8508 #else
8509 #define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8510  uint8x8_t __s1 = __p1; \
8511  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8512  uint8x8_t __ret; \
8513  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
8514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8515  __ret; \
8516 })
8517 #endif
8518 
8519 #ifdef __LITTLE_ENDIAN__
8520 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8521  uint32x2_t __s1 = __p1; \
8522  uint32x2_t __ret; \
8523  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
8524  __ret; \
8525 })
8526 #else
8527 #define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8528  uint32x2_t __s1 = __p1; \
8529  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8530  uint32x2_t __ret; \
8531  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
8532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8533  __ret; \
8534 })
8535 #endif
8536 
8537 #ifdef __LITTLE_ENDIAN__
8538 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8539  uint64x1_t __s1 = __p1; \
8540  uint64x1_t __ret; \
8541  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8542  __ret; \
8543 })
8544 #else
8545 #define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8546  uint64x1_t __s1 = __p1; \
8547  uint64x1_t __ret; \
8548  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8549  __ret; \
8550 })
8551 #endif
8552 
8553 #ifdef __LITTLE_ENDIAN__
8554 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8555  uint16x4_t __s1 = __p1; \
8556  uint16x4_t __ret; \
8557  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
8558  __ret; \
8559 })
8560 #else
8561 #define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8562  uint16x4_t __s1 = __p1; \
8563  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8564  uint16x4_t __ret; \
8565  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
8566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8567  __ret; \
8568 })
8569 #endif
8570 
8571 #ifdef __LITTLE_ENDIAN__
8572 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8573  int8x8_t __s1 = __p1; \
8574  int8x8_t __ret; \
8575  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
8576  __ret; \
8577 })
8578 #else
8579 #define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8580  int8x8_t __s1 = __p1; \
8581  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8582  int8x8_t __ret; \
8583  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
8584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8585  __ret; \
8586 })
8587 #endif
8588 
8589 #ifdef __LITTLE_ENDIAN__
8590 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8591  float32x2_t __s1 = __p1; \
8592  float32x2_t __ret; \
8593  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
8594  __ret; \
8595 })
8596 #else
8597 #define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8598  float32x2_t __s1 = __p1; \
8599  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8600  float32x2_t __ret; \
8601  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
8602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8603  __ret; \
8604 })
8605 #endif
8606 
8607 #ifdef __LITTLE_ENDIAN__
8608 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8609  float16x4_t __s1 = __p1; \
8610  float16x4_t __ret; \
8611  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
8612  __ret; \
8613 })
8614 #else
8615 #define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8616  float16x4_t __s1 = __p1; \
8617  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8618  float16x4_t __ret; \
8619  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
8620  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8621  __ret; \
8622 })
8623 #endif
8624 
8625 #ifdef __LITTLE_ENDIAN__
8626 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8627  int32x2_t __s1 = __p1; \
8628  int32x2_t __ret; \
8629  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
8630  __ret; \
8631 })
8632 #else
8633 #define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8634  int32x2_t __s1 = __p1; \
8635  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8636  int32x2_t __ret; \
8637  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
8638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8639  __ret; \
8640 })
8641 #endif
8642 
8643 #ifdef __LITTLE_ENDIAN__
8644 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8645  int64x1_t __s1 = __p1; \
8646  int64x1_t __ret; \
8647  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8648  __ret; \
8649 })
8650 #else
8651 #define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8652  int64x1_t __s1 = __p1; \
8653  int64x1_t __ret; \
8654  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8655  __ret; \
8656 })
8657 #endif
8658 
8659 #ifdef __LITTLE_ENDIAN__
8660 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8661  int16x4_t __s1 = __p1; \
8662  int16x4_t __ret; \
8663  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
8664  __ret; \
8665 })
8666 #else
8667 #define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8668  int16x4_t __s1 = __p1; \
8669  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8670  int16x4_t __ret; \
8671  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
8672  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8673  __ret; \
8674 })
8675 #endif
8676 
8677 #ifdef __LITTLE_ENDIAN__
8678 #define vld2_p8(__p0) __extension__ ({ \
8679  poly8x8x2_t __ret; \
8680  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8681  __ret; \
8682 })
8683 #else
8684 #define vld2_p8(__p0) __extension__ ({ \
8685  poly8x8x2_t __ret; \
8686  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8687  \
8688  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8689  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8690  __ret; \
8691 })
8692 #endif
8693 
8694 #ifdef __LITTLE_ENDIAN__
8695 #define vld2_p16(__p0) __extension__ ({ \
8696  poly16x4x2_t __ret; \
8697  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8698  __ret; \
8699 })
8700 #else
8701 #define vld2_p16(__p0) __extension__ ({ \
8702  poly16x4x2_t __ret; \
8703  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8704  \
8705  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8706  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8707  __ret; \
8708 })
8709 #endif
8710 
8711 #ifdef __LITTLE_ENDIAN__
8712 #define vld2q_p8(__p0) __extension__ ({ \
8713  poly8x16x2_t __ret; \
8714  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8715  __ret; \
8716 })
8717 #else
8718 #define vld2q_p8(__p0) __extension__ ({ \
8719  poly8x16x2_t __ret; \
8720  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8721  \
8722  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8723  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8724  __ret; \
8725 })
8726 #endif
8727 
8728 #ifdef __LITTLE_ENDIAN__
8729 #define vld2q_p16(__p0) __extension__ ({ \
8730  poly16x8x2_t __ret; \
8731  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8732  __ret; \
8733 })
8734 #else
8735 #define vld2q_p16(__p0) __extension__ ({ \
8736  poly16x8x2_t __ret; \
8737  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8738  \
8739  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8740  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8741  __ret; \
8742 })
8743 #endif
8744 
8745 #ifdef __LITTLE_ENDIAN__
8746 #define vld2q_u8(__p0) __extension__ ({ \
8747  uint8x16x2_t __ret; \
8748  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8749  __ret; \
8750 })
8751 #else
8752 #define vld2q_u8(__p0) __extension__ ({ \
8753  uint8x16x2_t __ret; \
8754  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8755  \
8756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8758  __ret; \
8759 })
8760 #endif
8761 
8762 #ifdef __LITTLE_ENDIAN__
8763 #define vld2q_u32(__p0) __extension__ ({ \
8764  uint32x4x2_t __ret; \
8765  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8766  __ret; \
8767 })
8768 #else
8769 #define vld2q_u32(__p0) __extension__ ({ \
8770  uint32x4x2_t __ret; \
8771  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8772  \
8773  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8774  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8775  __ret; \
8776 })
8777 #endif
8778 
8779 #ifdef __LITTLE_ENDIAN__
8780 #define vld2q_u16(__p0) __extension__ ({ \
8781  uint16x8x2_t __ret; \
8782  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8783  __ret; \
8784 })
8785 #else
8786 #define vld2q_u16(__p0) __extension__ ({ \
8787  uint16x8x2_t __ret; \
8788  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8789  \
8790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8792  __ret; \
8793 })
8794 #endif
8795 
8796 #ifdef __LITTLE_ENDIAN__
8797 #define vld2q_s8(__p0) __extension__ ({ \
8798  int8x16x2_t __ret; \
8799  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8800  __ret; \
8801 })
8802 #else
8803 #define vld2q_s8(__p0) __extension__ ({ \
8804  int8x16x2_t __ret; \
8805  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8806  \
8807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8809  __ret; \
8810 })
8811 #endif
8812 
8813 #ifdef __LITTLE_ENDIAN__
8814 #define vld2q_f32(__p0) __extension__ ({ \
8815  float32x4x2_t __ret; \
8816  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8817  __ret; \
8818 })
8819 #else
8820 #define vld2q_f32(__p0) __extension__ ({ \
8821  float32x4x2_t __ret; \
8822  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8823  \
8824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8826  __ret; \
8827 })
8828 #endif
8829 
8830 #ifdef __LITTLE_ENDIAN__
8831 #define vld2q_f16(__p0) __extension__ ({ \
8832  float16x8x2_t __ret; \
8833  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8834  __ret; \
8835 })
8836 #else
8837 #define vld2q_f16(__p0) __extension__ ({ \
8838  float16x8x2_t __ret; \
8839  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8840  \
8841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8843  __ret; \
8844 })
8845 #endif
8846 
8847 #ifdef __LITTLE_ENDIAN__
8848 #define vld2q_s32(__p0) __extension__ ({ \
8849  int32x4x2_t __ret; \
8850  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8851  __ret; \
8852 })
8853 #else
8854 #define vld2q_s32(__p0) __extension__ ({ \
8855  int32x4x2_t __ret; \
8856  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8857  \
8858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8860  __ret; \
8861 })
8862 #endif
8863 
8864 #ifdef __LITTLE_ENDIAN__
8865 #define vld2q_s16(__p0) __extension__ ({ \
8866  int16x8x2_t __ret; \
8867  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8868  __ret; \
8869 })
8870 #else
8871 #define vld2q_s16(__p0) __extension__ ({ \
8872  int16x8x2_t __ret; \
8873  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8874  \
8875  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8876  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8877  __ret; \
8878 })
8879 #endif
8880 
8881 #ifdef __LITTLE_ENDIAN__
8882 #define vld2_u8(__p0) __extension__ ({ \
8883  uint8x8x2_t __ret; \
8884  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8885  __ret; \
8886 })
8887 #else
8888 #define vld2_u8(__p0) __extension__ ({ \
8889  uint8x8x2_t __ret; \
8890  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8891  \
8892  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8893  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8894  __ret; \
8895 })
8896 #endif
8897 
8898 #ifdef __LITTLE_ENDIAN__
8899 #define vld2_u32(__p0) __extension__ ({ \
8900  uint32x2x2_t __ret; \
8901  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8902  __ret; \
8903 })
8904 #else
8905 #define vld2_u32(__p0) __extension__ ({ \
8906  uint32x2x2_t __ret; \
8907  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8908  \
8909  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8910  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8911  __ret; \
8912 })
8913 #endif
8914 
8915 #ifdef __LITTLE_ENDIAN__
8916 #define vld2_u64(__p0) __extension__ ({ \
8917  uint64x1x2_t __ret; \
8918  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8919  __ret; \
8920 })
8921 #else
8922 #define vld2_u64(__p0) __extension__ ({ \
8923  uint64x1x2_t __ret; \
8924  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8925  __ret; \
8926 })
8927 #endif
8928 
8929 #ifdef __LITTLE_ENDIAN__
8930 #define vld2_u16(__p0) __extension__ ({ \
8931  uint16x4x2_t __ret; \
8932  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8933  __ret; \
8934 })
8935 #else
8936 #define vld2_u16(__p0) __extension__ ({ \
8937  uint16x4x2_t __ret; \
8938  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8939  \
8940  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8941  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8942  __ret; \
8943 })
8944 #endif
8945 
8946 #ifdef __LITTLE_ENDIAN__
8947 #define vld2_s8(__p0) __extension__ ({ \
8948  int8x8x2_t __ret; \
8949  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8950  __ret; \
8951 })
8952 #else
8953 #define vld2_s8(__p0) __extension__ ({ \
8954  int8x8x2_t __ret; \
8955  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8956  \
8957  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8958  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8959  __ret; \
8960 })
8961 #endif
8962 
8963 #ifdef __LITTLE_ENDIAN__
8964 #define vld2_f32(__p0) __extension__ ({ \
8965  float32x2x2_t __ret; \
8966  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8967  __ret; \
8968 })
8969 #else
8970 #define vld2_f32(__p0) __extension__ ({ \
8971  float32x2x2_t __ret; \
8972  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8973  \
8974  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8975  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8976  __ret; \
8977 })
8978 #endif
8979 
8980 #ifdef __LITTLE_ENDIAN__
8981 #define vld2_f16(__p0) __extension__ ({ \
8982  float16x4x2_t __ret; \
8983  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8984  __ret; \
8985 })
8986 #else
8987 #define vld2_f16(__p0) __extension__ ({ \
8988  float16x4x2_t __ret; \
8989  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8990  \
8991  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8992  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8993  __ret; \
8994 })
8995 #endif
8996 
8997 #ifdef __LITTLE_ENDIAN__
8998 #define vld2_s32(__p0) __extension__ ({ \
8999  int32x2x2_t __ret; \
9000  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9001  __ret; \
9002 })
9003 #else
9004 #define vld2_s32(__p0) __extension__ ({ \
9005  int32x2x2_t __ret; \
9006  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9007  \
9008  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9009  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9010  __ret; \
9011 })
9012 #endif
9013 
9014 #ifdef __LITTLE_ENDIAN__
9015 #define vld2_s64(__p0) __extension__ ({ \
9016  int64x1x2_t __ret; \
9017  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9018  __ret; \
9019 })
9020 #else
9021 #define vld2_s64(__p0) __extension__ ({ \
9022  int64x1x2_t __ret; \
9023  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9024  __ret; \
9025 })
9026 #endif
9027 
9028 #ifdef __LITTLE_ENDIAN__
9029 #define vld2_s16(__p0) __extension__ ({ \
9030  int16x4x2_t __ret; \
9031  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9032  __ret; \
9033 })
9034 #else
9035 #define vld2_s16(__p0) __extension__ ({ \
9036  int16x4x2_t __ret; \
9037  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9038  \
9039  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9040  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9041  __ret; \
9042 })
9043 #endif
9044 
9045 #ifdef __LITTLE_ENDIAN__
9046 #define vld2_dup_p8(__p0) __extension__ ({ \
9047  poly8x8x2_t __ret; \
9048  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9049  __ret; \
9050 })
9051 #else
9052 #define vld2_dup_p8(__p0) __extension__ ({ \
9053  poly8x8x2_t __ret; \
9054  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9055  \
9056  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9057  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9058  __ret; \
9059 })
9060 #endif
9061 
9062 #ifdef __LITTLE_ENDIAN__
9063 #define vld2_dup_p16(__p0) __extension__ ({ \
9064  poly16x4x2_t __ret; \
9065  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9066  __ret; \
9067 })
9068 #else
9069 #define vld2_dup_p16(__p0) __extension__ ({ \
9070  poly16x4x2_t __ret; \
9071  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9072  \
9073  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9074  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9075  __ret; \
9076 })
9077 #endif
9078 
9079 #ifdef __LITTLE_ENDIAN__
9080 #define vld2_dup_u8(__p0) __extension__ ({ \
9081  uint8x8x2_t __ret; \
9082  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9083  __ret; \
9084 })
9085 #else
9086 #define vld2_dup_u8(__p0) __extension__ ({ \
9087  uint8x8x2_t __ret; \
9088  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9089  \
9090  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9091  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9092  __ret; \
9093 })
9094 #endif
9095 
9096 #ifdef __LITTLE_ENDIAN__
9097 #define vld2_dup_u32(__p0) __extension__ ({ \
9098  uint32x2x2_t __ret; \
9099  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9100  __ret; \
9101 })
9102 #else
9103 #define vld2_dup_u32(__p0) __extension__ ({ \
9104  uint32x2x2_t __ret; \
9105  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9106  \
9107  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9108  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9109  __ret; \
9110 })
9111 #endif
9112 
9113 #ifdef __LITTLE_ENDIAN__
9114 #define vld2_dup_u64(__p0) __extension__ ({ \
9115  uint64x1x2_t __ret; \
9116  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9117  __ret; \
9118 })
9119 #else
9120 #define vld2_dup_u64(__p0) __extension__ ({ \
9121  uint64x1x2_t __ret; \
9122  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9123  __ret; \
9124 })
9125 #endif
9126 
9127 #ifdef __LITTLE_ENDIAN__
9128 #define vld2_dup_u16(__p0) __extension__ ({ \
9129  uint16x4x2_t __ret; \
9130  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9131  __ret; \
9132 })
9133 #else
9134 #define vld2_dup_u16(__p0) __extension__ ({ \
9135  uint16x4x2_t __ret; \
9136  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9137  \
9138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9140  __ret; \
9141 })
9142 #endif
9143 
9144 #ifdef __LITTLE_ENDIAN__
9145 #define vld2_dup_s8(__p0) __extension__ ({ \
9146  int8x8x2_t __ret; \
9147  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9148  __ret; \
9149 })
9150 #else
9151 #define vld2_dup_s8(__p0) __extension__ ({ \
9152  int8x8x2_t __ret; \
9153  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9154  \
9155  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9156  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9157  __ret; \
9158 })
9159 #endif
9160 
9161 #ifdef __LITTLE_ENDIAN__
9162 #define vld2_dup_f32(__p0) __extension__ ({ \
9163  float32x2x2_t __ret; \
9164  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9165  __ret; \
9166 })
9167 #else
9168 #define vld2_dup_f32(__p0) __extension__ ({ \
9169  float32x2x2_t __ret; \
9170  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9171  \
9172  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9173  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9174  __ret; \
9175 })
9176 #endif
9177 
9178 #ifdef __LITTLE_ENDIAN__
9179 #define vld2_dup_f16(__p0) __extension__ ({ \
9180  float16x4x2_t __ret; \
9181  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9182  __ret; \
9183 })
9184 #else
9185 #define vld2_dup_f16(__p0) __extension__ ({ \
9186  float16x4x2_t __ret; \
9187  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9188  \
9189  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9190  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9191  __ret; \
9192 })
9193 #endif
9194 
9195 #ifdef __LITTLE_ENDIAN__
9196 #define vld2_dup_s32(__p0) __extension__ ({ \
9197  int32x2x2_t __ret; \
9198  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9199  __ret; \
9200 })
9201 #else
9202 #define vld2_dup_s32(__p0) __extension__ ({ \
9203  int32x2x2_t __ret; \
9204  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9205  \
9206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9208  __ret; \
9209 })
9210 #endif
9211 
9212 #ifdef __LITTLE_ENDIAN__
9213 #define vld2_dup_s64(__p0) __extension__ ({ \
9214  int64x1x2_t __ret; \
9215  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9216  __ret; \
9217 })
9218 #else
9219 #define vld2_dup_s64(__p0) __extension__ ({ \
9220  int64x1x2_t __ret; \
9221  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9222  __ret; \
9223 })
9224 #endif
9225 
9226 #ifdef __LITTLE_ENDIAN__
9227 #define vld2_dup_s16(__p0) __extension__ ({ \
9228  int16x4x2_t __ret; \
9229  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9230  __ret; \
9231 })
9232 #else
9233 #define vld2_dup_s16(__p0) __extension__ ({ \
9234  int16x4x2_t __ret; \
9235  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9236  \
9237  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9238  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9239  __ret; \
9240 })
9241 #endif
9242 
9243 #ifdef __LITTLE_ENDIAN__
9244 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9245  poly8x8x2_t __s1 = __p1; \
9246  poly8x8x2_t __ret; \
9247  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
9248  __ret; \
9249 })
9250 #else
9251 #define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9252  poly8x8x2_t __s1 = __p1; \
9253  poly8x8x2_t __rev1; \
9254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9256  poly8x8x2_t __ret; \
9257  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
9258  \
9259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9261  __ret; \
9262 })
9263 #endif
9264 
9265 #ifdef __LITTLE_ENDIAN__
9266 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9267  poly16x4x2_t __s1 = __p1; \
9268  poly16x4x2_t __ret; \
9269  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
9270  __ret; \
9271 })
9272 #else
9273 #define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9274  poly16x4x2_t __s1 = __p1; \
9275  poly16x4x2_t __rev1; \
9276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9278  poly16x4x2_t __ret; \
9279  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
9280  \
9281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9283  __ret; \
9284 })
9285 #endif
9286 
9287 #ifdef __LITTLE_ENDIAN__
9288 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9289  poly16x8x2_t __s1 = __p1; \
9290  poly16x8x2_t __ret; \
9291  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
9292  __ret; \
9293 })
9294 #else
9295 #define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9296  poly16x8x2_t __s1 = __p1; \
9297  poly16x8x2_t __rev1; \
9298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9300  poly16x8x2_t __ret; \
9301  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
9302  \
9303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9305  __ret; \
9306 })
9307 #endif
9308 
9309 #ifdef __LITTLE_ENDIAN__
9310 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9311  uint32x4x2_t __s1 = __p1; \
9312  uint32x4x2_t __ret; \
9313  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
9314  __ret; \
9315 })
9316 #else
9317 #define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9318  uint32x4x2_t __s1 = __p1; \
9319  uint32x4x2_t __rev1; \
9320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9322  uint32x4x2_t __ret; \
9323  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
9324  \
9325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9327  __ret; \
9328 })
9329 #endif
9330 
9331 #ifdef __LITTLE_ENDIAN__
9332 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9333  uint16x8x2_t __s1 = __p1; \
9334  uint16x8x2_t __ret; \
9335  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
9336  __ret; \
9337 })
9338 #else
9339 #define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9340  uint16x8x2_t __s1 = __p1; \
9341  uint16x8x2_t __rev1; \
9342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9344  uint16x8x2_t __ret; \
9345  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
9346  \
9347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9349  __ret; \
9350 })
9351 #endif
9352 
9353 #ifdef __LITTLE_ENDIAN__
9354 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9355  float32x4x2_t __s1 = __p1; \
9356  float32x4x2_t __ret; \
9357  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 41); \
9358  __ret; \
9359 })
9360 #else
9361 #define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9362  float32x4x2_t __s1 = __p1; \
9363  float32x4x2_t __rev1; \
9364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9366  float32x4x2_t __ret; \
9367  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
9368  \
9369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9371  __ret; \
9372 })
9373 #endif
9374 
9375 #ifdef __LITTLE_ENDIAN__
9376 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9377  float16x8x2_t __s1 = __p1; \
9378  float16x8x2_t __ret; \
9379  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 40); \
9380  __ret; \
9381 })
9382 #else
9383 #define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9384  float16x8x2_t __s1 = __p1; \
9385  float16x8x2_t __rev1; \
9386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9388  float16x8x2_t __ret; \
9389  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
9390  \
9391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9393  __ret; \
9394 })
9395 #endif
9396 
9397 #ifdef __LITTLE_ENDIAN__
9398 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9399  int32x4x2_t __s1 = __p1; \
9400  int32x4x2_t __ret; \
9401  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 34); \
9402  __ret; \
9403 })
9404 #else
9405 #define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9406  int32x4x2_t __s1 = __p1; \
9407  int32x4x2_t __rev1; \
9408  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9409  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9410  int32x4x2_t __ret; \
9411  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
9412  \
9413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9415  __ret; \
9416 })
9417 #endif
9418 
9419 #ifdef __LITTLE_ENDIAN__
9420 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9421  int16x8x2_t __s1 = __p1; \
9422  int16x8x2_t __ret; \
9423  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 33); \
9424  __ret; \
9425 })
9426 #else
9427 #define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9428  int16x8x2_t __s1 = __p1; \
9429  int16x8x2_t __rev1; \
9430  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9431  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9432  int16x8x2_t __ret; \
9433  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
9434  \
9435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9437  __ret; \
9438 })
9439 #endif
9440 
9441 #ifdef __LITTLE_ENDIAN__
9442 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9443  uint8x8x2_t __s1 = __p1; \
9444  uint8x8x2_t __ret; \
9445  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
9446  __ret; \
9447 })
9448 #else
9449 #define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9450  uint8x8x2_t __s1 = __p1; \
9451  uint8x8x2_t __rev1; \
9452  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9453  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9454  uint8x8x2_t __ret; \
9455  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
9456  \
9457  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9458  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9459  __ret; \
9460 })
9461 #endif
9462 
9463 #ifdef __LITTLE_ENDIAN__
9464 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9465  uint32x2x2_t __s1 = __p1; \
9466  uint32x2x2_t __ret; \
9467  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
9468  __ret; \
9469 })
9470 #else
9471 #define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9472  uint32x2x2_t __s1 = __p1; \
9473  uint32x2x2_t __rev1; \
9474  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9475  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9476  uint32x2x2_t __ret; \
9477  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
9478  \
9479  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9480  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9481  __ret; \
9482 })
9483 #endif
9484 
9485 #ifdef __LITTLE_ENDIAN__
9486 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9487  uint16x4x2_t __s1 = __p1; \
9488  uint16x4x2_t __ret; \
9489  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
9490  __ret; \
9491 })
9492 #else
9493 #define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9494  uint16x4x2_t __s1 = __p1; \
9495  uint16x4x2_t __rev1; \
9496  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9497  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9498  uint16x4x2_t __ret; \
9499  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
9500  \
9501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9503  __ret; \
9504 })
9505 #endif
9506 
9507 #ifdef __LITTLE_ENDIAN__
9508 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9509  int8x8x2_t __s1 = __p1; \
9510  int8x8x2_t __ret; \
9511  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
9512  __ret; \
9513 })
9514 #else
9515 #define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9516  int8x8x2_t __s1 = __p1; \
9517  int8x8x2_t __rev1; \
9518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9520  int8x8x2_t __ret; \
9521  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
9522  \
9523  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9524  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9525  __ret; \
9526 })
9527 #endif
9528 
9529 #ifdef __LITTLE_ENDIAN__
9530 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9531  float32x2x2_t __s1 = __p1; \
9532  float32x2x2_t __ret; \
9533  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 9); \
9534  __ret; \
9535 })
9536 #else
9537 #define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9538  float32x2x2_t __s1 = __p1; \
9539  float32x2x2_t __rev1; \
9540  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9541  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9542  float32x2x2_t __ret; \
9543  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
9544  \
9545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9547  __ret; \
9548 })
9549 #endif
9550 
9551 #ifdef __LITTLE_ENDIAN__
9552 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9553  float16x4x2_t __s1 = __p1; \
9554  float16x4x2_t __ret; \
9555  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 8); \
9556  __ret; \
9557 })
9558 #else
9559 #define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9560  float16x4x2_t __s1 = __p1; \
9561  float16x4x2_t __rev1; \
9562  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9563  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9564  float16x4x2_t __ret; \
9565  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
9566  \
9567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9569  __ret; \
9570 })
9571 #endif
9572 
9573 #ifdef __LITTLE_ENDIAN__
9574 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9575  int32x2x2_t __s1 = __p1; \
9576  int32x2x2_t __ret; \
9577  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 2); \
9578  __ret; \
9579 })
9580 #else
9581 #define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9582  int32x2x2_t __s1 = __p1; \
9583  int32x2x2_t __rev1; \
9584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9586  int32x2x2_t __ret; \
9587  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
9588  \
9589  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9590  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9591  __ret; \
9592 })
9593 #endif
9594 
9595 #ifdef __LITTLE_ENDIAN__
9596 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9597  int16x4x2_t __s1 = __p1; \
9598  int16x4x2_t __ret; \
9599  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 1); \
9600  __ret; \
9601 })
9602 #else
9603 #define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9604  int16x4x2_t __s1 = __p1; \
9605  int16x4x2_t __rev1; \
9606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9608  int16x4x2_t __ret; \
9609  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
9610  \
9611  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9612  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9613  __ret; \
9614 })
9615 #endif
9616 
9617 #ifdef __LITTLE_ENDIAN__
9618 #define vld3_p8(__p0) __extension__ ({ \
9619  poly8x8x3_t __ret; \
9620  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9621  __ret; \
9622 })
9623 #else
9624 #define vld3_p8(__p0) __extension__ ({ \
9625  poly8x8x3_t __ret; \
9626  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9627  \
9628  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9629  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9630  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9631  __ret; \
9632 })
9633 #endif
9634 
9635 #ifdef __LITTLE_ENDIAN__
9636 #define vld3_p16(__p0) __extension__ ({ \
9637  poly16x4x3_t __ret; \
9638  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9639  __ret; \
9640 })
9641 #else
9642 #define vld3_p16(__p0) __extension__ ({ \
9643  poly16x4x3_t __ret; \
9644  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9645  \
9646  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9647  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9648  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9649  __ret; \
9650 })
9651 #endif
9652 
9653 #ifdef __LITTLE_ENDIAN__
9654 #define vld3q_p8(__p0) __extension__ ({ \
9655  poly8x16x3_t __ret; \
9656  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9657  __ret; \
9658 })
9659 #else
9660 #define vld3q_p8(__p0) __extension__ ({ \
9661  poly8x16x3_t __ret; \
9662  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9663  \
9664  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9665  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9666  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9667  __ret; \
9668 })
9669 #endif
9670 
9671 #ifdef __LITTLE_ENDIAN__
9672 #define vld3q_p16(__p0) __extension__ ({ \
9673  poly16x8x3_t __ret; \
9674  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9675  __ret; \
9676 })
9677 #else
9678 #define vld3q_p16(__p0) __extension__ ({ \
9679  poly16x8x3_t __ret; \
9680  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9681  \
9682  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9683  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9684  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9685  __ret; \
9686 })
9687 #endif
9688 
9689 #ifdef __LITTLE_ENDIAN__
9690 #define vld3q_u8(__p0) __extension__ ({ \
9691  uint8x16x3_t __ret; \
9692  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9693  __ret; \
9694 })
9695 #else
9696 #define vld3q_u8(__p0) __extension__ ({ \
9697  uint8x16x3_t __ret; \
9698  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9699  \
9700  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9701  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9702  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9703  __ret; \
9704 })
9705 #endif
9706 
9707 #ifdef __LITTLE_ENDIAN__
9708 #define vld3q_u32(__p0) __extension__ ({ \
9709  uint32x4x3_t __ret; \
9710  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9711  __ret; \
9712 })
9713 #else
9714 #define vld3q_u32(__p0) __extension__ ({ \
9715  uint32x4x3_t __ret; \
9716  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9717  \
9718  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9719  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9720  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9721  __ret; \
9722 })
9723 #endif
9724 
9725 #ifdef __LITTLE_ENDIAN__
9726 #define vld3q_u16(__p0) __extension__ ({ \
9727  uint16x8x3_t __ret; \
9728  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9729  __ret; \
9730 })
9731 #else
9732 #define vld3q_u16(__p0) __extension__ ({ \
9733  uint16x8x3_t __ret; \
9734  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9735  \
9736  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9737  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9738  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9739  __ret; \
9740 })
9741 #endif
9742 
9743 #ifdef __LITTLE_ENDIAN__
9744 #define vld3q_s8(__p0) __extension__ ({ \
9745  int8x16x3_t __ret; \
9746  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9747  __ret; \
9748 })
9749 #else
9750 #define vld3q_s8(__p0) __extension__ ({ \
9751  int8x16x3_t __ret; \
9752  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9753  \
9754  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9755  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9756  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9757  __ret; \
9758 })
9759 #endif
9760 
9761 #ifdef __LITTLE_ENDIAN__
9762 #define vld3q_f32(__p0) __extension__ ({ \
9763  float32x4x3_t __ret; \
9764  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9765  __ret; \
9766 })
9767 #else
9768 #define vld3q_f32(__p0) __extension__ ({ \
9769  float32x4x3_t __ret; \
9770  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9771  \
9772  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9773  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9774  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9775  __ret; \
9776 })
9777 #endif
9778 
9779 #ifdef __LITTLE_ENDIAN__
9780 #define vld3q_f16(__p0) __extension__ ({ \
9781  float16x8x3_t __ret; \
9782  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9783  __ret; \
9784 })
9785 #else
9786 #define vld3q_f16(__p0) __extension__ ({ \
9787  float16x8x3_t __ret; \
9788  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9789  \
9790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9792  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9793  __ret; \
9794 })
9795 #endif
9796 
9797 #ifdef __LITTLE_ENDIAN__
9798 #define vld3q_s32(__p0) __extension__ ({ \
9799  int32x4x3_t __ret; \
9800  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9801  __ret; \
9802 })
9803 #else
9804 #define vld3q_s32(__p0) __extension__ ({ \
9805  int32x4x3_t __ret; \
9806  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9807  \
9808  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9809  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9810  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9811  __ret; \
9812 })
9813 #endif
9814 
9815 #ifdef __LITTLE_ENDIAN__
9816 #define vld3q_s16(__p0) __extension__ ({ \
9817  int16x8x3_t __ret; \
9818  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9819  __ret; \
9820 })
9821 #else
9822 #define vld3q_s16(__p0) __extension__ ({ \
9823  int16x8x3_t __ret; \
9824  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9825  \
9826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9828  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9829  __ret; \
9830 })
9831 #endif
9832 
9833 #ifdef __LITTLE_ENDIAN__
9834 #define vld3_u8(__p0) __extension__ ({ \
9835  uint8x8x3_t __ret; \
9836  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9837  __ret; \
9838 })
9839 #else
9840 #define vld3_u8(__p0) __extension__ ({ \
9841  uint8x8x3_t __ret; \
9842  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9843  \
9844  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9845  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9846  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9847  __ret; \
9848 })
9849 #endif
9850 
9851 #ifdef __LITTLE_ENDIAN__
9852 #define vld3_u32(__p0) __extension__ ({ \
9853  uint32x2x3_t __ret; \
9854  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9855  __ret; \
9856 })
9857 #else
9858 #define vld3_u32(__p0) __extension__ ({ \
9859  uint32x2x3_t __ret; \
9860  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9861  \
9862  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9863  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9864  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9865  __ret; \
9866 })
9867 #endif
9868 
9869 #ifdef __LITTLE_ENDIAN__
9870 #define vld3_u64(__p0) __extension__ ({ \
9871  uint64x1x3_t __ret; \
9872  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9873  __ret; \
9874 })
9875 #else
9876 #define vld3_u64(__p0) __extension__ ({ \
9877  uint64x1x3_t __ret; \
9878  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9879  __ret; \
9880 })
9881 #endif
9882 
9883 #ifdef __LITTLE_ENDIAN__
9884 #define vld3_u16(__p0) __extension__ ({ \
9885  uint16x4x3_t __ret; \
9886  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9887  __ret; \
9888 })
9889 #else
9890 #define vld3_u16(__p0) __extension__ ({ \
9891  uint16x4x3_t __ret; \
9892  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9893  \
9894  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9895  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9896  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9897  __ret; \
9898 })
9899 #endif
9900 
9901 #ifdef __LITTLE_ENDIAN__
9902 #define vld3_s8(__p0) __extension__ ({ \
9903  int8x8x3_t __ret; \
9904  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9905  __ret; \
9906 })
9907 #else
9908 #define vld3_s8(__p0) __extension__ ({ \
9909  int8x8x3_t __ret; \
9910  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9911  \
9912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9915  __ret; \
9916 })
9917 #endif
9918 
9919 #ifdef __LITTLE_ENDIAN__
9920 #define vld3_f32(__p0) __extension__ ({ \
9921  float32x2x3_t __ret; \
9922  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9923  __ret; \
9924 })
9925 #else
9926 #define vld3_f32(__p0) __extension__ ({ \
9927  float32x2x3_t __ret; \
9928  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9929  \
9930  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9931  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9932  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9933  __ret; \
9934 })
9935 #endif
9936 
9937 #ifdef __LITTLE_ENDIAN__
9938 #define vld3_f16(__p0) __extension__ ({ \
9939  float16x4x3_t __ret; \
9940  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9941  __ret; \
9942 })
9943 #else
9944 #define vld3_f16(__p0) __extension__ ({ \
9945  float16x4x3_t __ret; \
9946  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9947  \
9948  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9949  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9950  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9951  __ret; \
9952 })
9953 #endif
9954 
9955 #ifdef __LITTLE_ENDIAN__
9956 #define vld3_s32(__p0) __extension__ ({ \
9957  int32x2x3_t __ret; \
9958  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9959  __ret; \
9960 })
9961 #else
9962 #define vld3_s32(__p0) __extension__ ({ \
9963  int32x2x3_t __ret; \
9964  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9965  \
9966  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9967  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9968  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9969  __ret; \
9970 })
9971 #endif
9972 
9973 #ifdef __LITTLE_ENDIAN__
9974 #define vld3_s64(__p0) __extension__ ({ \
9975  int64x1x3_t __ret; \
9976  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9977  __ret; \
9978 })
9979 #else
9980 #define vld3_s64(__p0) __extension__ ({ \
9981  int64x1x3_t __ret; \
9982  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9983  __ret; \
9984 })
9985 #endif
9986 
9987 #ifdef __LITTLE_ENDIAN__
9988 #define vld3_s16(__p0) __extension__ ({ \
9989  int16x4x3_t __ret; \
9990  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9991  __ret; \
9992 })
9993 #else
9994 #define vld3_s16(__p0) __extension__ ({ \
9995  int16x4x3_t __ret; \
9996  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9997  \
9998  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9999  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10000  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10001  __ret; \
10002 })
10003 #endif
10004 
10005 #ifdef __LITTLE_ENDIAN__
10006 #define vld3_dup_p8(__p0) __extension__ ({ \
10007  poly8x8x3_t __ret; \
10008  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10009  __ret; \
10010 })
10011 #else
10012 #define vld3_dup_p8(__p0) __extension__ ({ \
10013  poly8x8x3_t __ret; \
10014  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10015  \
10016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10018  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10019  __ret; \
10020 })
10021 #endif
10022 
10023 #ifdef __LITTLE_ENDIAN__
10024 #define vld3_dup_p16(__p0) __extension__ ({ \
10025  poly16x4x3_t __ret; \
10026  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10027  __ret; \
10028 })
10029 #else
10030 #define vld3_dup_p16(__p0) __extension__ ({ \
10031  poly16x4x3_t __ret; \
10032  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10033  \
10034  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10035  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10036  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10037  __ret; \
10038 })
10039 #endif
10040 
10041 #ifdef __LITTLE_ENDIAN__
10042 #define vld3_dup_u8(__p0) __extension__ ({ \
10043  uint8x8x3_t __ret; \
10044  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10045  __ret; \
10046 })
10047 #else
10048 #define vld3_dup_u8(__p0) __extension__ ({ \
10049  uint8x8x3_t __ret; \
10050  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10051  \
10052  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10053  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10054  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10055  __ret; \
10056 })
10057 #endif
10058 
10059 #ifdef __LITTLE_ENDIAN__
10060 #define vld3_dup_u32(__p0) __extension__ ({ \
10061  uint32x2x3_t __ret; \
10062  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10063  __ret; \
10064 })
10065 #else
10066 #define vld3_dup_u32(__p0) __extension__ ({ \
10067  uint32x2x3_t __ret; \
10068  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10069  \
10070  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10071  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10072  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10073  __ret; \
10074 })
10075 #endif
10076 
10077 #ifdef __LITTLE_ENDIAN__
10078 #define vld3_dup_u64(__p0) __extension__ ({ \
10079  uint64x1x3_t __ret; \
10080  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10081  __ret; \
10082 })
10083 #else
10084 #define vld3_dup_u64(__p0) __extension__ ({ \
10085  uint64x1x3_t __ret; \
10086  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10087  __ret; \
10088 })
10089 #endif
10090 
10091 #ifdef __LITTLE_ENDIAN__
10092 #define vld3_dup_u16(__p0) __extension__ ({ \
10093  uint16x4x3_t __ret; \
10094  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10095  __ret; \
10096 })
10097 #else
10098 #define vld3_dup_u16(__p0) __extension__ ({ \
10099  uint16x4x3_t __ret; \
10100  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10101  \
10102  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10103  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10104  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10105  __ret; \
10106 })
10107 #endif
10108 
10109 #ifdef __LITTLE_ENDIAN__
10110 #define vld3_dup_s8(__p0) __extension__ ({ \
10111  int8x8x3_t __ret; \
10112  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10113  __ret; \
10114 })
10115 #else
10116 #define vld3_dup_s8(__p0) __extension__ ({ \
10117  int8x8x3_t __ret; \
10118  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10119  \
10120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10123  __ret; \
10124 })
10125 #endif
10126 
10127 #ifdef __LITTLE_ENDIAN__
10128 #define vld3_dup_f32(__p0) __extension__ ({ \
10129  float32x2x3_t __ret; \
10130  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10131  __ret; \
10132 })
10133 #else
10134 #define vld3_dup_f32(__p0) __extension__ ({ \
10135  float32x2x3_t __ret; \
10136  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10137  \
10138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10140  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10141  __ret; \
10142 })
10143 #endif
10144 
10145 #ifdef __LITTLE_ENDIAN__
10146 #define vld3_dup_f16(__p0) __extension__ ({ \
10147  float16x4x3_t __ret; \
10148  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10149  __ret; \
10150 })
10151 #else
10152 #define vld3_dup_f16(__p0) __extension__ ({ \
10153  float16x4x3_t __ret; \
10154  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10155  \
10156  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10157  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10158  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10159  __ret; \
10160 })
10161 #endif
10162 
10163 #ifdef __LITTLE_ENDIAN__
10164 #define vld3_dup_s32(__p0) __extension__ ({ \
10165  int32x2x3_t __ret; \
10166  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10167  __ret; \
10168 })
10169 #else
10170 #define vld3_dup_s32(__p0) __extension__ ({ \
10171  int32x2x3_t __ret; \
10172  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10173  \
10174  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10175  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10176  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10177  __ret; \
10178 })
10179 #endif
10180 
10181 #ifdef __LITTLE_ENDIAN__
10182 #define vld3_dup_s64(__p0) __extension__ ({ \
10183  int64x1x3_t __ret; \
10184  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10185  __ret; \
10186 })
10187 #else
10188 #define vld3_dup_s64(__p0) __extension__ ({ \
10189  int64x1x3_t __ret; \
10190  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10191  __ret; \
10192 })
10193 #endif
10194 
10195 #ifdef __LITTLE_ENDIAN__
10196 #define vld3_dup_s16(__p0) __extension__ ({ \
10197  int16x4x3_t __ret; \
10198  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10199  __ret; \
10200 })
10201 #else
10202 #define vld3_dup_s16(__p0) __extension__ ({ \
10203  int16x4x3_t __ret; \
10204  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10205  \
10206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10209  __ret; \
10210 })
10211 #endif
10212 
10213 #ifdef __LITTLE_ENDIAN__
10214 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10215  poly8x8x3_t __s1 = __p1; \
10216  poly8x8x3_t __ret; \
10217  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
10218  __ret; \
10219 })
10220 #else
10221 #define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10222  poly8x8x3_t __s1 = __p1; \
10223  poly8x8x3_t __rev1; \
10224  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10225  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10226  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10227  poly8x8x3_t __ret; \
10228  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
10229  \
10230  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10231  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10232  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10233  __ret; \
10234 })
10235 #endif
10236 
10237 #ifdef __LITTLE_ENDIAN__
10238 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10239  poly16x4x3_t __s1 = __p1; \
10240  poly16x4x3_t __ret; \
10241  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
10242  __ret; \
10243 })
10244 #else
10245 #define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10246  poly16x4x3_t __s1 = __p1; \
10247  poly16x4x3_t __rev1; \
10248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10251  poly16x4x3_t __ret; \
10252  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
10253  \
10254  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10255  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10256  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10257  __ret; \
10258 })
10259 #endif
10260 
10261 #ifdef __LITTLE_ENDIAN__
10262 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10263  poly16x8x3_t __s1 = __p1; \
10264  poly16x8x3_t __ret; \
10265  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
10266  __ret; \
10267 })
10268 #else
10269 #define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10270  poly16x8x3_t __s1 = __p1; \
10271  poly16x8x3_t __rev1; \
10272  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10273  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10274  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10275  poly16x8x3_t __ret; \
10276  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
10277  \
10278  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10279  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10280  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10281  __ret; \
10282 })
10283 #endif
10284 
10285 #ifdef __LITTLE_ENDIAN__
10286 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10287  uint32x4x3_t __s1 = __p1; \
10288  uint32x4x3_t __ret; \
10289  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
10290  __ret; \
10291 })
10292 #else
10293 #define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10294  uint32x4x3_t __s1 = __p1; \
10295  uint32x4x3_t __rev1; \
10296  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10297  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10298  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10299  uint32x4x3_t __ret; \
10300  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
10301  \
10302  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10303  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10304  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10305  __ret; \
10306 })
10307 #endif
10308 
10309 #ifdef __LITTLE_ENDIAN__
10310 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10311  uint16x8x3_t __s1 = __p1; \
10312  uint16x8x3_t __ret; \
10313  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
10314  __ret; \
10315 })
10316 #else
10317 #define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10318  uint16x8x3_t __s1 = __p1; \
10319  uint16x8x3_t __rev1; \
10320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10322  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10323  uint16x8x3_t __ret; \
10324  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
10325  \
10326  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10327  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10328  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10329  __ret; \
10330 })
10331 #endif
10332 
10333 #ifdef __LITTLE_ENDIAN__
10334 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10335  float32x4x3_t __s1 = __p1; \
10336  float32x4x3_t __ret; \
10337  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
10338  __ret; \
10339 })
10340 #else
10341 #define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10342  float32x4x3_t __s1 = __p1; \
10343  float32x4x3_t __rev1; \
10344  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10345  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10346  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10347  float32x4x3_t __ret; \
10348  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
10349  \
10350  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10351  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10352  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10353  __ret; \
10354 })
10355 #endif
10356 
10357 #ifdef __LITTLE_ENDIAN__
10358 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10359  float16x8x3_t __s1 = __p1; \
10360  float16x8x3_t __ret; \
10361  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
10362  __ret; \
10363 })
10364 #else
10365 #define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10366  float16x8x3_t __s1 = __p1; \
10367  float16x8x3_t __rev1; \
10368  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10369  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10370  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10371  float16x8x3_t __ret; \
10372  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
10373  \
10374  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10375  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10376  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10377  __ret; \
10378 })
10379 #endif
10380 
10381 #ifdef __LITTLE_ENDIAN__
10382 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10383  int32x4x3_t __s1 = __p1; \
10384  int32x4x3_t __ret; \
10385  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
10386  __ret; \
10387 })
10388 #else
10389 #define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10390  int32x4x3_t __s1 = __p1; \
10391  int32x4x3_t __rev1; \
10392  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10393  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10394  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10395  int32x4x3_t __ret; \
10396  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
10397  \
10398  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10399  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10400  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10401  __ret; \
10402 })
10403 #endif
10404 
10405 #ifdef __LITTLE_ENDIAN__
10406 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10407  int16x8x3_t __s1 = __p1; \
10408  int16x8x3_t __ret; \
10409  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
10410  __ret; \
10411 })
10412 #else
10413 #define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10414  int16x8x3_t __s1 = __p1; \
10415  int16x8x3_t __rev1; \
10416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10419  int16x8x3_t __ret; \
10420  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
10421  \
10422  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10423  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10424  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10425  __ret; \
10426 })
10427 #endif
10428 
10429 #ifdef __LITTLE_ENDIAN__
10430 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10431  uint8x8x3_t __s1 = __p1; \
10432  uint8x8x3_t __ret; \
10433  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
10434  __ret; \
10435 })
10436 #else
10437 #define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10438  uint8x8x3_t __s1 = __p1; \
10439  uint8x8x3_t __rev1; \
10440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10443  uint8x8x3_t __ret; \
10444  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
10445  \
10446  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10447  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10448  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10449  __ret; \
10450 })
10451 #endif
10452 
10453 #ifdef __LITTLE_ENDIAN__
10454 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10455  uint32x2x3_t __s1 = __p1; \
10456  uint32x2x3_t __ret; \
10457  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
10458  __ret; \
10459 })
10460 #else
10461 #define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10462  uint32x2x3_t __s1 = __p1; \
10463  uint32x2x3_t __rev1; \
10464  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10465  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10466  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10467  uint32x2x3_t __ret; \
10468  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
10469  \
10470  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10471  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10472  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10473  __ret; \
10474 })
10475 #endif
10476 
10477 #ifdef __LITTLE_ENDIAN__
10478 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10479  uint16x4x3_t __s1 = __p1; \
10480  uint16x4x3_t __ret; \
10481  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
10482  __ret; \
10483 })
10484 #else
10485 #define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10486  uint16x4x3_t __s1 = __p1; \
10487  uint16x4x3_t __rev1; \
10488  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10489  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10490  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10491  uint16x4x3_t __ret; \
10492  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
10493  \
10494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10496  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10497  __ret; \
10498 })
10499 #endif
10500 
10501 #ifdef __LITTLE_ENDIAN__
10502 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10503  int8x8x3_t __s1 = __p1; \
10504  int8x8x3_t __ret; \
10505  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
10506  __ret; \
10507 })
10508 #else
10509 #define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10510  int8x8x3_t __s1 = __p1; \
10511  int8x8x3_t __rev1; \
10512  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10513  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10514  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10515  int8x8x3_t __ret; \
10516  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
10517  \
10518  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10519  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10520  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10521  __ret; \
10522 })
10523 #endif
10524 
10525 #ifdef __LITTLE_ENDIAN__
10526 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10527  float32x2x3_t __s1 = __p1; \
10528  float32x2x3_t __ret; \
10529  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
10530  __ret; \
10531 })
10532 #else
10533 #define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10534  float32x2x3_t __s1 = __p1; \
10535  float32x2x3_t __rev1; \
10536  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10537  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10538  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10539  float32x2x3_t __ret; \
10540  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
10541  \
10542  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10543  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10544  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10545  __ret; \
10546 })
10547 #endif
10548 
10549 #ifdef __LITTLE_ENDIAN__
10550 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10551  float16x4x3_t __s1 = __p1; \
10552  float16x4x3_t __ret; \
10553  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
10554  __ret; \
10555 })
10556 #else
10557 #define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10558  float16x4x3_t __s1 = __p1; \
10559  float16x4x3_t __rev1; \
10560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10563  float16x4x3_t __ret; \
10564  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
10565  \
10566  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10567  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10568  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10569  __ret; \
10570 })
10571 #endif
10572 
10573 #ifdef __LITTLE_ENDIAN__
10574 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10575  int32x2x3_t __s1 = __p1; \
10576  int32x2x3_t __ret; \
10577  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
10578  __ret; \
10579 })
10580 #else
10581 #define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10582  int32x2x3_t __s1 = __p1; \
10583  int32x2x3_t __rev1; \
10584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10586  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10587  int32x2x3_t __ret; \
10588  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
10589  \
10590  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10591  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10592  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10593  __ret; \
10594 })
10595 #endif
10596 
10597 #ifdef __LITTLE_ENDIAN__
10598 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10599  int16x4x3_t __s1 = __p1; \
10600  int16x4x3_t __ret; \
10601  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
10602  __ret; \
10603 })
10604 #else
10605 #define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10606  int16x4x3_t __s1 = __p1; \
10607  int16x4x3_t __rev1; \
10608  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10609  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10610  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10611  int16x4x3_t __ret; \
10612  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
10613  \
10614  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10615  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10616  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10617  __ret; \
10618 })
10619 #endif
10620 
10621 #ifdef __LITTLE_ENDIAN__
10622 #define vld4_p8(__p0) __extension__ ({ \
10623  poly8x8x4_t __ret; \
10624  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10625  __ret; \
10626 })
10627 #else
10628 #define vld4_p8(__p0) __extension__ ({ \
10629  poly8x8x4_t __ret; \
10630  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10631  \
10632  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10633  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10634  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10635  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10636  __ret; \
10637 })
10638 #endif
10639 
10640 #ifdef __LITTLE_ENDIAN__
10641 #define vld4_p16(__p0) __extension__ ({ \
10642  poly16x4x4_t __ret; \
10643  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10644  __ret; \
10645 })
10646 #else
10647 #define vld4_p16(__p0) __extension__ ({ \
10648  poly16x4x4_t __ret; \
10649  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10650  \
10651  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10652  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10653  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10654  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10655  __ret; \
10656 })
10657 #endif
10658 
10659 #ifdef __LITTLE_ENDIAN__
10660 #define vld4q_p8(__p0) __extension__ ({ \
10661  poly8x16x4_t __ret; \
10662  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10663  __ret; \
10664 })
10665 #else
10666 #define vld4q_p8(__p0) __extension__ ({ \
10667  poly8x16x4_t __ret; \
10668  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10669  \
10670  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10671  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10672  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10673  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10674  __ret; \
10675 })
10676 #endif
10677 
10678 #ifdef __LITTLE_ENDIAN__
10679 #define vld4q_p16(__p0) __extension__ ({ \
10680  poly16x8x4_t __ret; \
10681  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10682  __ret; \
10683 })
10684 #else
10685 #define vld4q_p16(__p0) __extension__ ({ \
10686  poly16x8x4_t __ret; \
10687  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10688  \
10689  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10690  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10691  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10692  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10693  __ret; \
10694 })
10695 #endif
10696 
10697 #ifdef __LITTLE_ENDIAN__
10698 #define vld4q_u8(__p0) __extension__ ({ \
10699  uint8x16x4_t __ret; \
10700  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10701  __ret; \
10702 })
10703 #else
10704 #define vld4q_u8(__p0) __extension__ ({ \
10705  uint8x16x4_t __ret; \
10706  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10707  \
10708  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10709  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10710  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10711  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10712  __ret; \
10713 })
10714 #endif
10715 
10716 #ifdef __LITTLE_ENDIAN__
10717 #define vld4q_u32(__p0) __extension__ ({ \
10718  uint32x4x4_t __ret; \
10719  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10720  __ret; \
10721 })
10722 #else
10723 #define vld4q_u32(__p0) __extension__ ({ \
10724  uint32x4x4_t __ret; \
10725  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10726  \
10727  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10728  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10729  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10730  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10731  __ret; \
10732 })
10733 #endif
10734 
10735 #ifdef __LITTLE_ENDIAN__
10736 #define vld4q_u16(__p0) __extension__ ({ \
10737  uint16x8x4_t __ret; \
10738  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10739  __ret; \
10740 })
10741 #else
10742 #define vld4q_u16(__p0) __extension__ ({ \
10743  uint16x8x4_t __ret; \
10744  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10745  \
10746  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10747  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10748  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10749  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10750  __ret; \
10751 })
10752 #endif
10753 
10754 #ifdef __LITTLE_ENDIAN__
10755 #define vld4q_s8(__p0) __extension__ ({ \
10756  int8x16x4_t __ret; \
10757  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10758  __ret; \
10759 })
10760 #else
10761 #define vld4q_s8(__p0) __extension__ ({ \
10762  int8x16x4_t __ret; \
10763  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10764  \
10765  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10766  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10767  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10768  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10769  __ret; \
10770 })
10771 #endif
10772 
10773 #ifdef __LITTLE_ENDIAN__
10774 #define vld4q_f32(__p0) __extension__ ({ \
10775  float32x4x4_t __ret; \
10776  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10777  __ret; \
10778 })
10779 #else
10780 #define vld4q_f32(__p0) __extension__ ({ \
10781  float32x4x4_t __ret; \
10782  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10783  \
10784  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10785  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10786  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10787  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10788  __ret; \
10789 })
10790 #endif
10791 
10792 #ifdef __LITTLE_ENDIAN__
10793 #define vld4q_f16(__p0) __extension__ ({ \
10794  float16x8x4_t __ret; \
10795  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10796  __ret; \
10797 })
10798 #else
10799 #define vld4q_f16(__p0) __extension__ ({ \
10800  float16x8x4_t __ret; \
10801  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10802  \
10803  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10804  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10805  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10806  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10807  __ret; \
10808 })
10809 #endif
10810 
10811 #ifdef __LITTLE_ENDIAN__
10812 #define vld4q_s32(__p0) __extension__ ({ \
10813  int32x4x4_t __ret; \
10814  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10815  __ret; \
10816 })
10817 #else
10818 #define vld4q_s32(__p0) __extension__ ({ \
10819  int32x4x4_t __ret; \
10820  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10821  \
10822  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10823  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10824  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10825  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10826  __ret; \
10827 })
10828 #endif
10829 
10830 #ifdef __LITTLE_ENDIAN__
10831 #define vld4q_s16(__p0) __extension__ ({ \
10832  int16x8x4_t __ret; \
10833  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10834  __ret; \
10835 })
10836 #else
10837 #define vld4q_s16(__p0) __extension__ ({ \
10838  int16x8x4_t __ret; \
10839  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10840  \
10841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10843  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10844  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10845  __ret; \
10846 })
10847 #endif
10848 
10849 #ifdef __LITTLE_ENDIAN__
10850 #define vld4_u8(__p0) __extension__ ({ \
10851  uint8x8x4_t __ret; \
10852  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10853  __ret; \
10854 })
10855 #else
10856 #define vld4_u8(__p0) __extension__ ({ \
10857  uint8x8x4_t __ret; \
10858  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10859  \
10860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10863  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10864  __ret; \
10865 })
10866 #endif
10867 
10868 #ifdef __LITTLE_ENDIAN__
10869 #define vld4_u32(__p0) __extension__ ({ \
10870  uint32x2x4_t __ret; \
10871  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10872  __ret; \
10873 })
10874 #else
10875 #define vld4_u32(__p0) __extension__ ({ \
10876  uint32x2x4_t __ret; \
10877  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10878  \
10879  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10880  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10881  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10882  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10883  __ret; \
10884 })
10885 #endif
10886 
10887 #ifdef __LITTLE_ENDIAN__
10888 #define vld4_u64(__p0) __extension__ ({ \
10889  uint64x1x4_t __ret; \
10890  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10891  __ret; \
10892 })
10893 #else
10894 #define vld4_u64(__p0) __extension__ ({ \
10895  uint64x1x4_t __ret; \
10896  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10897  __ret; \
10898 })
10899 #endif
10900 
10901 #ifdef __LITTLE_ENDIAN__
10902 #define vld4_u16(__p0) __extension__ ({ \
10903  uint16x4x4_t __ret; \
10904  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10905  __ret; \
10906 })
10907 #else
10908 #define vld4_u16(__p0) __extension__ ({ \
10909  uint16x4x4_t __ret; \
10910  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10911  \
10912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10915  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10916  __ret; \
10917 })
10918 #endif
10919 
10920 #ifdef __LITTLE_ENDIAN__
10921 #define vld4_s8(__p0) __extension__ ({ \
10922  int8x8x4_t __ret; \
10923  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10924  __ret; \
10925 })
10926 #else
10927 #define vld4_s8(__p0) __extension__ ({ \
10928  int8x8x4_t __ret; \
10929  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10930  \
10931  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10932  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10933  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10934  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10935  __ret; \
10936 })
10937 #endif
10938 
10939 #ifdef __LITTLE_ENDIAN__
10940 #define vld4_f32(__p0) __extension__ ({ \
10941  float32x2x4_t __ret; \
10942  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10943  __ret; \
10944 })
10945 #else
10946 #define vld4_f32(__p0) __extension__ ({ \
10947  float32x2x4_t __ret; \
10948  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10949  \
10950  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10951  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10952  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10953  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10954  __ret; \
10955 })
10956 #endif
10957 
10958 #ifdef __LITTLE_ENDIAN__
10959 #define vld4_f16(__p0) __extension__ ({ \
10960  float16x4x4_t __ret; \
10961  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10962  __ret; \
10963 })
10964 #else
10965 #define vld4_f16(__p0) __extension__ ({ \
10966  float16x4x4_t __ret; \
10967  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10968  \
10969  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10970  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10971  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10972  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10973  __ret; \
10974 })
10975 #endif
10976 
10977 #ifdef __LITTLE_ENDIAN__
10978 #define vld4_s32(__p0) __extension__ ({ \
10979  int32x2x4_t __ret; \
10980  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10981  __ret; \
10982 })
10983 #else
10984 #define vld4_s32(__p0) __extension__ ({ \
10985  int32x2x4_t __ret; \
10986  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10987  \
10988  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10989  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10990  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10991  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10992  __ret; \
10993 })
10994 #endif
10995 
10996 #ifdef __LITTLE_ENDIAN__
10997 #define vld4_s64(__p0) __extension__ ({ \
10998  int64x1x4_t __ret; \
10999  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11000  __ret; \
11001 })
11002 #else
11003 #define vld4_s64(__p0) __extension__ ({ \
11004  int64x1x4_t __ret; \
11005  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11006  __ret; \
11007 })
11008 #endif
11009 
11010 #ifdef __LITTLE_ENDIAN__
11011 #define vld4_s16(__p0) __extension__ ({ \
11012  int16x4x4_t __ret; \
11013  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11014  __ret; \
11015 })
11016 #else
11017 #define vld4_s16(__p0) __extension__ ({ \
11018  int16x4x4_t __ret; \
11019  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11020  \
11021  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11022  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11023  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11024  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11025  __ret; \
11026 })
11027 #endif
11028 
11029 #ifdef __LITTLE_ENDIAN__
11030 #define vld4_dup_p8(__p0) __extension__ ({ \
11031  poly8x8x4_t __ret; \
11032  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11033  __ret; \
11034 })
11035 #else
11036 #define vld4_dup_p8(__p0) __extension__ ({ \
11037  poly8x8x4_t __ret; \
11038  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11039  \
11040  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11041  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11042  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11043  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11044  __ret; \
11045 })
11046 #endif
11047 
11048 #ifdef __LITTLE_ENDIAN__
11049 #define vld4_dup_p16(__p0) __extension__ ({ \
11050  poly16x4x4_t __ret; \
11051  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11052  __ret; \
11053 })
11054 #else
11055 #define vld4_dup_p16(__p0) __extension__ ({ \
11056  poly16x4x4_t __ret; \
11057  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11058  \
11059  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11060  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11061  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11062  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11063  __ret; \
11064 })
11065 #endif
11066 
11067 #ifdef __LITTLE_ENDIAN__
11068 #define vld4_dup_u8(__p0) __extension__ ({ \
11069  uint8x8x4_t __ret; \
11070  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11071  __ret; \
11072 })
11073 #else
11074 #define vld4_dup_u8(__p0) __extension__ ({ \
11075  uint8x8x4_t __ret; \
11076  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11077  \
11078  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11079  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11080  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11081  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11082  __ret; \
11083 })
11084 #endif
11085 
11086 #ifdef __LITTLE_ENDIAN__
11087 #define vld4_dup_u32(__p0) __extension__ ({ \
11088  uint32x2x4_t __ret; \
11089  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11090  __ret; \
11091 })
11092 #else
11093 #define vld4_dup_u32(__p0) __extension__ ({ \
11094  uint32x2x4_t __ret; \
11095  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11096  \
11097  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11098  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11099  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11100  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11101  __ret; \
11102 })
11103 #endif
11104 
11105 #ifdef __LITTLE_ENDIAN__
11106 #define vld4_dup_u64(__p0) __extension__ ({ \
11107  uint64x1x4_t __ret; \
11108  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11109  __ret; \
11110 })
11111 #else
11112 #define vld4_dup_u64(__p0) __extension__ ({ \
11113  uint64x1x4_t __ret; \
11114  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11115  __ret; \
11116 })
11117 #endif
11118 
11119 #ifdef __LITTLE_ENDIAN__
11120 #define vld4_dup_u16(__p0) __extension__ ({ \
11121  uint16x4x4_t __ret; \
11122  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11123  __ret; \
11124 })
11125 #else
11126 #define vld4_dup_u16(__p0) __extension__ ({ \
11127  uint16x4x4_t __ret; \
11128  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11129  \
11130  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11131  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11132  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11133  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11134  __ret; \
11135 })
11136 #endif
11137 
11138 #ifdef __LITTLE_ENDIAN__
11139 #define vld4_dup_s8(__p0) __extension__ ({ \
11140  int8x8x4_t __ret; \
11141  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11142  __ret; \
11143 })
11144 #else
11145 #define vld4_dup_s8(__p0) __extension__ ({ \
11146  int8x8x4_t __ret; \
11147  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11148  \
11149  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11150  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11151  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11152  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11153  __ret; \
11154 })
11155 #endif
11156 
11157 #ifdef __LITTLE_ENDIAN__
11158 #define vld4_dup_f32(__p0) __extension__ ({ \
11159  float32x2x4_t __ret; \
11160  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11161  __ret; \
11162 })
11163 #else
11164 #define vld4_dup_f32(__p0) __extension__ ({ \
11165  float32x2x4_t __ret; \
11166  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11167  \
11168  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11169  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11170  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11171  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11172  __ret; \
11173 })
11174 #endif
11175 
11176 #ifdef __LITTLE_ENDIAN__
11177 #define vld4_dup_f16(__p0) __extension__ ({ \
11178  float16x4x4_t __ret; \
11179  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11180  __ret; \
11181 })
11182 #else
11183 #define vld4_dup_f16(__p0) __extension__ ({ \
11184  float16x4x4_t __ret; \
11185  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11186  \
11187  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11188  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11189  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11190  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11191  __ret; \
11192 })
11193 #endif
11194 
11195 #ifdef __LITTLE_ENDIAN__
11196 #define vld4_dup_s32(__p0) __extension__ ({ \
11197  int32x2x4_t __ret; \
11198  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11199  __ret; \
11200 })
11201 #else
11202 #define vld4_dup_s32(__p0) __extension__ ({ \
11203  int32x2x4_t __ret; \
11204  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11205  \
11206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11209  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11210  __ret; \
11211 })
11212 #endif
11213 
11214 #ifdef __LITTLE_ENDIAN__
11215 #define vld4_dup_s64(__p0) __extension__ ({ \
11216  int64x1x4_t __ret; \
11217  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11218  __ret; \
11219 })
11220 #else
11221 #define vld4_dup_s64(__p0) __extension__ ({ \
11222  int64x1x4_t __ret; \
11223  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11224  __ret; \
11225 })
11226 #endif
11227 
11228 #ifdef __LITTLE_ENDIAN__
11229 #define vld4_dup_s16(__p0) __extension__ ({ \
11230  int16x4x4_t __ret; \
11231  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11232  __ret; \
11233 })
11234 #else
11235 #define vld4_dup_s16(__p0) __extension__ ({ \
11236  int16x4x4_t __ret; \
11237  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11238  \
11239  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11240  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11241  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11242  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11243  __ret; \
11244 })
11245 #endif
11246 
11247 #ifdef __LITTLE_ENDIAN__
11248 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11249  poly8x8x4_t __s1 = __p1; \
11250  poly8x8x4_t __ret; \
11251  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
11252  __ret; \
11253 })
11254 #else
11255 #define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11256  poly8x8x4_t __s1 = __p1; \
11257  poly8x8x4_t __rev1; \
11258  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11259  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11260  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11261  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11262  poly8x8x4_t __ret; \
11263  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
11264  \
11265  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11266  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11267  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11268  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11269  __ret; \
11270 })
11271 #endif
11272 
11273 #ifdef __LITTLE_ENDIAN__
11274 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11275  poly16x4x4_t __s1 = __p1; \
11276  poly16x4x4_t __ret; \
11277  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
11278  __ret; \
11279 })
11280 #else
11281 #define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11282  poly16x4x4_t __s1 = __p1; \
11283  poly16x4x4_t __rev1; \
11284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11286  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11287  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11288  poly16x4x4_t __ret; \
11289  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
11290  \
11291  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11292  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11293  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11294  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11295  __ret; \
11296 })
11297 #endif
11298 
11299 #ifdef __LITTLE_ENDIAN__
11300 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11301  poly16x8x4_t __s1 = __p1; \
11302  poly16x8x4_t __ret; \
11303  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
11304  __ret; \
11305 })
11306 #else
11307 #define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11308  poly16x8x4_t __s1 = __p1; \
11309  poly16x8x4_t __rev1; \
11310  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11311  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11312  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11313  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11314  poly16x8x4_t __ret; \
11315  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
11316  \
11317  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11318  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11319  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11320  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11321  __ret; \
11322 })
11323 #endif
11324 
11325 #ifdef __LITTLE_ENDIAN__
11326 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11327  uint32x4x4_t __s1 = __p1; \
11328  uint32x4x4_t __ret; \
11329  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
11330  __ret; \
11331 })
11332 #else
11333 #define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11334  uint32x4x4_t __s1 = __p1; \
11335  uint32x4x4_t __rev1; \
11336  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11337  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11338  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11339  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11340  uint32x4x4_t __ret; \
11341  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
11342  \
11343  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11344  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11345  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11346  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11347  __ret; \
11348 })
11349 #endif
11350 
11351 #ifdef __LITTLE_ENDIAN__
11352 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11353  uint16x8x4_t __s1 = __p1; \
11354  uint16x8x4_t __ret; \
11355  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
11356  __ret; \
11357 })
11358 #else
11359 #define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11360  uint16x8x4_t __s1 = __p1; \
11361  uint16x8x4_t __rev1; \
11362  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11363  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11364  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11365  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11366  uint16x8x4_t __ret; \
11367  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
11368  \
11369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11371  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11372  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11373  __ret; \
11374 })
11375 #endif
11376 
11377 #ifdef __LITTLE_ENDIAN__
11378 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11379  float32x4x4_t __s1 = __p1; \
11380  float32x4x4_t __ret; \
11381  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
11382  __ret; \
11383 })
11384 #else
11385 #define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11386  float32x4x4_t __s1 = __p1; \
11387  float32x4x4_t __rev1; \
11388  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11389  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11390  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11391  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11392  float32x4x4_t __ret; \
11393  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
11394  \
11395  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11396  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11397  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11398  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11399  __ret; \
11400 })
11401 #endif
11402 
11403 #ifdef __LITTLE_ENDIAN__
11404 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11405  float16x8x4_t __s1 = __p1; \
11406  float16x8x4_t __ret; \
11407  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
11408  __ret; \
11409 })
11410 #else
11411 #define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11412  float16x8x4_t __s1 = __p1; \
11413  float16x8x4_t __rev1; \
11414  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11415  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11416  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11417  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11418  float16x8x4_t __ret; \
11419  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
11420  \
11421  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11422  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11423  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11424  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11425  __ret; \
11426 })
11427 #endif
11428 
11429 #ifdef __LITTLE_ENDIAN__
11430 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11431  int32x4x4_t __s1 = __p1; \
11432  int32x4x4_t __ret; \
11433  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
11434  __ret; \
11435 })
11436 #else
11437 #define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11438  int32x4x4_t __s1 = __p1; \
11439  int32x4x4_t __rev1; \
11440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11443  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11444  int32x4x4_t __ret; \
11445  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
11446  \
11447  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11448  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11449  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11450  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11451  __ret; \
11452 })
11453 #endif
11454 
11455 #ifdef __LITTLE_ENDIAN__
11456 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11457  int16x8x4_t __s1 = __p1; \
11458  int16x8x4_t __ret; \
11459  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
11460  __ret; \
11461 })
11462 #else
11463 #define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11464  int16x8x4_t __s1 = __p1; \
11465  int16x8x4_t __rev1; \
11466  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11467  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11468  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11469  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11470  int16x8x4_t __ret; \
11471  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
11472  \
11473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11476  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11477  __ret; \
11478 })
11479 #endif
11480 
11481 #ifdef __LITTLE_ENDIAN__
11482 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11483  uint8x8x4_t __s1 = __p1; \
11484  uint8x8x4_t __ret; \
11485  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
11486  __ret; \
11487 })
11488 #else
11489 #define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11490  uint8x8x4_t __s1 = __p1; \
11491  uint8x8x4_t __rev1; \
11492  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11493  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11494  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11495  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11496  uint8x8x4_t __ret; \
11497  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
11498  \
11499  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11500  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11501  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11502  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11503  __ret; \
11504 })
11505 #endif
11506 
11507 #ifdef __LITTLE_ENDIAN__
11508 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11509  uint32x2x4_t __s1 = __p1; \
11510  uint32x2x4_t __ret; \
11511  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
11512  __ret; \
11513 })
11514 #else
11515 #define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11516  uint32x2x4_t __s1 = __p1; \
11517  uint32x2x4_t __rev1; \
11518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11520  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11521  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11522  uint32x2x4_t __ret; \
11523  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
11524  \
11525  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11526  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11527  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11528  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11529  __ret; \
11530 })
11531 #endif
11532 
11533 #ifdef __LITTLE_ENDIAN__
11534 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11535  uint16x4x4_t __s1 = __p1; \
11536  uint16x4x4_t __ret; \
11537  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
11538  __ret; \
11539 })
11540 #else
11541 #define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11542  uint16x4x4_t __s1 = __p1; \
11543  uint16x4x4_t __rev1; \
11544  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11545  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11546  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11547  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11548  uint16x4x4_t __ret; \
11549  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
11550  \
11551  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11552  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11553  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11554  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11555  __ret; \
11556 })
11557 #endif
11558 
11559 #ifdef __LITTLE_ENDIAN__
11560 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11561  int8x8x4_t __s1 = __p1; \
11562  int8x8x4_t __ret; \
11563  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
11564  __ret; \
11565 })
11566 #else
11567 #define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11568  int8x8x4_t __s1 = __p1; \
11569  int8x8x4_t __rev1; \
11570  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11571  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11572  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11573  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11574  int8x8x4_t __ret; \
11575  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
11576  \
11577  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11578  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11579  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11580  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11581  __ret; \
11582 })
11583 #endif
11584 
11585 #ifdef __LITTLE_ENDIAN__
11586 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11587  float32x2x4_t __s1 = __p1; \
11588  float32x2x4_t __ret; \
11589  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
11590  __ret; \
11591 })
11592 #else
11593 #define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11594  float32x2x4_t __s1 = __p1; \
11595  float32x2x4_t __rev1; \
11596  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11597  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11598  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11599  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11600  float32x2x4_t __ret; \
11601  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
11602  \
11603  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11604  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11605  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11606  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11607  __ret; \
11608 })
11609 #endif
11610 
11611 #ifdef __LITTLE_ENDIAN__
11612 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11613  float16x4x4_t __s1 = __p1; \
11614  float16x4x4_t __ret; \
11615  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
11616  __ret; \
11617 })
11618 #else
11619 #define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11620  float16x4x4_t __s1 = __p1; \
11621  float16x4x4_t __rev1; \
11622  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11623  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11624  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11625  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11626  float16x4x4_t __ret; \
11627  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
11628  \
11629  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11630  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11631  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11632  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11633  __ret; \
11634 })
11635 #endif
11636 
11637 #ifdef __LITTLE_ENDIAN__
11638 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11639  int32x2x4_t __s1 = __p1; \
11640  int32x2x4_t __ret; \
11641  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
11642  __ret; \
11643 })
11644 #else
11645 #define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11646  int32x2x4_t __s1 = __p1; \
11647  int32x2x4_t __rev1; \
11648  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11649  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11650  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11651  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11652  int32x2x4_t __ret; \
11653  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
11654  \
11655  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11656  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11657  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11658  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11659  __ret; \
11660 })
11661 #endif
11662 
11663 #ifdef __LITTLE_ENDIAN__
11664 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11665  int16x4x4_t __s1 = __p1; \
11666  int16x4x4_t __ret; \
11667  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
11668  __ret; \
11669 })
11670 #else
11671 #define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11672  int16x4x4_t __s1 = __p1; \
11673  int16x4x4_t __rev1; \
11674  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11675  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11676  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11677  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11678  int16x4x4_t __ret; \
11679  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
11680  \
11681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11683  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11684  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11685  __ret; \
11686 })
11687 #endif
11688 
11689 #ifdef __LITTLE_ENDIAN__
11690 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11691  uint8x16_t __ret;
11692  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11693  return __ret;
11694 }
11695 #else
11696 __ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11697  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11698  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11699  uint8x16_t __ret;
11700  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11701  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11702  return __ret;
11703 }
11704 #endif
11705 
11706 #ifdef __LITTLE_ENDIAN__
11707 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11708  uint32x4_t __ret;
11709  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11710  return __ret;
11711 }
11712 #else
11713 __ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11714  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11715  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11716  uint32x4_t __ret;
11717  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11719  return __ret;
11720 }
11721 #endif
11722 
11723 #ifdef __LITTLE_ENDIAN__
11724 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11725  uint16x8_t __ret;
11726  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11727  return __ret;
11728 }
11729 #else
11730 __ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11731  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11732  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11733  uint16x8_t __ret;
11734  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11735  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11736  return __ret;
11737 }
11738 #endif
11739 
11740 #ifdef __LITTLE_ENDIAN__
11741 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11742  int8x16_t __ret;
11743  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11744  return __ret;
11745 }
11746 #else
11747 __ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11748  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11749  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11750  int8x16_t __ret;
11751  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11753  return __ret;
11754 }
11755 #endif
11756 
11757 #ifdef __LITTLE_ENDIAN__
11758 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11759  float32x4_t __ret;
11760  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11761  return __ret;
11762 }
11763 #else
11764 __ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11765  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11766  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11767  float32x4_t __ret;
11768  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
11769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11770  return __ret;
11771 }
11772 #endif
11773 
11774 #ifdef __LITTLE_ENDIAN__
11775 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11776  int32x4_t __ret;
11777  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
11778  return __ret;
11779 }
11780 #else
11781 __ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11782  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11783  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11784  int32x4_t __ret;
11785  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
11786  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11787  return __ret;
11788 }
11789 #endif
11790 
11791 #ifdef __LITTLE_ENDIAN__
11792 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11793  int16x8_t __ret;
11794  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
11795  return __ret;
11796 }
11797 #else
11798 __ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11799  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11800  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11801  int16x8_t __ret;
11802  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
11803  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11804  return __ret;
11805 }
11806 #endif
11807 
11808 #ifdef __LITTLE_ENDIAN__
11809 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11810  uint8x8_t __ret;
11811  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
11812  return __ret;
11813 }
11814 #else
11815 __ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11816  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11817  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11818  uint8x8_t __ret;
11819  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
11820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11821  return __ret;
11822 }
11823 #endif
11824 
11825 #ifdef __LITTLE_ENDIAN__
11826 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11827  uint32x2_t __ret;
11828  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
11829  return __ret;
11830 }
11831 #else
11832 __ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11833  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11834  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11835  uint32x2_t __ret;
11836  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
11837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11838  return __ret;
11839 }
11840 #endif
11841 
11842 #ifdef __LITTLE_ENDIAN__
11843 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11844  uint16x4_t __ret;
11845  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
11846  return __ret;
11847 }
11848 #else
11849 __ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11850  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11851  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11852  uint16x4_t __ret;
11853  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
11854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11855  return __ret;
11856 }
11857 #endif
11858 
11859 #ifdef __LITTLE_ENDIAN__
11860 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11861  int8x8_t __ret;
11862  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
11863  return __ret;
11864 }
11865 #else
11866 __ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11867  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11868  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11869  int8x8_t __ret;
11870  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
11871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11872  return __ret;
11873 }
11874 #endif
11875 
11876 #ifdef __LITTLE_ENDIAN__
11877 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11878  float32x2_t __ret;
11879  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
11880  return __ret;
11881 }
11882 #else
11883 __ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11884  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11885  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11886  float32x2_t __ret;
11887  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
11888  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11889  return __ret;
11890 }
11891 #endif
11892 
11893 #ifdef __LITTLE_ENDIAN__
11894 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11895  int32x2_t __ret;
11896  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
11897  return __ret;
11898 }
11899 #else
11900 __ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11901  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11902  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11903  int32x2_t __ret;
11904  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
11905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11906  return __ret;
11907 }
11908 #endif
11909 
11910 #ifdef __LITTLE_ENDIAN__
11911 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11912  int16x4_t __ret;
11913  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
11914  return __ret;
11915 }
11916 #else
11917 __ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11918  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11919  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11920  int16x4_t __ret;
11921  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
11922  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11923  return __ret;
11924 }
11925 #endif
11926 
11927 #ifdef __LITTLE_ENDIAN__
11928 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11929  uint8x16_t __ret;
11930  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11931  return __ret;
11932 }
11933 #else
11934 __ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11935  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11936  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11937  uint8x16_t __ret;
11938  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11940  return __ret;
11941 }
11942 #endif
11943 
11944 #ifdef __LITTLE_ENDIAN__
11945 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11946  uint32x4_t __ret;
11947  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11948  return __ret;
11949 }
11950 #else
11951 __ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11952  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11953  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11954  uint32x4_t __ret;
11955  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11957  return __ret;
11958 }
11959 #endif
11960 
11961 #ifdef __LITTLE_ENDIAN__
11962 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11963  uint16x8_t __ret;
11964  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11965  return __ret;
11966 }
11967 #else
11968 __ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11969  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11970  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11971  uint16x8_t __ret;
11972  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11973  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11974  return __ret;
11975 }
11976 #endif
11977 
11978 #ifdef __LITTLE_ENDIAN__
11979 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11980  int8x16_t __ret;
11981  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11982  return __ret;
11983 }
11984 #else
11985 __ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11986  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11987  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11988  int8x16_t __ret;
11989  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11990  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11991  return __ret;
11992 }
11993 #endif
11994 
11995 #ifdef __LITTLE_ENDIAN__
11996 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
11997  float32x4_t __ret;
11998  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11999  return __ret;
12000 }
12001 #else
12002 __ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
12003  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12004  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12005  float32x4_t __ret;
12006  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
12007  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12008  return __ret;
12009 }
12010 #endif
12011 
12012 #ifdef __LITTLE_ENDIAN__
12013 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12014  int32x4_t __ret;
12015  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
12016  return __ret;
12017 }
12018 #else
12019 __ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12020  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12021  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12022  int32x4_t __ret;
12023  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
12024  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12025  return __ret;
12026 }
12027 #endif
12028 
12029 #ifdef __LITTLE_ENDIAN__
12030 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12031  int16x8_t __ret;
12032  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
12033  return __ret;
12034 }
12035 #else
12036 __ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12037  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12038  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12039  int16x8_t __ret;
12040  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
12041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12042  return __ret;
12043 }
12044 #endif
12045 
12046 #ifdef __LITTLE_ENDIAN__
12047 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12048  uint8x8_t __ret;
12049  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
12050  return __ret;
12051 }
12052 #else
12053 __ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12054  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12055  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12056  uint8x8_t __ret;
12057  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
12058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12059  return __ret;
12060 }
12061 #endif
12062 
12063 #ifdef __LITTLE_ENDIAN__
12064 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12065  uint32x2_t __ret;
12066  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
12067  return __ret;
12068 }
12069 #else
12070 __ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12071  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12072  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12073  uint32x2_t __ret;
12074  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
12075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12076  return __ret;
12077 }
12078 #endif
12079 
12080 #ifdef __LITTLE_ENDIAN__
12081 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12082  uint16x4_t __ret;
12083  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
12084  return __ret;
12085 }
12086 #else
12087 __ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12088  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12089  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12090  uint16x4_t __ret;
12091  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
12092  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12093  return __ret;
12094 }
12095 #endif
12096 
12097 #ifdef __LITTLE_ENDIAN__
12098 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12099  int8x8_t __ret;
12100  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
12101  return __ret;
12102 }
12103 #else
12104 __ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12105  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12106  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12107  int8x8_t __ret;
12108  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
12109  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12110  return __ret;
12111 }
12112 #endif
12113 
12114 #ifdef __LITTLE_ENDIAN__
12115 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12116  float32x2_t __ret;
12117  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
12118  return __ret;
12119 }
12120 #else
12121 __ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12122  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12123  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12124  float32x2_t __ret;
12125  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
12126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12127  return __ret;
12128 }
12129 #endif
12130 
12131 #ifdef __LITTLE_ENDIAN__
12132 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12133  int32x2_t __ret;
12134  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
12135  return __ret;
12136 }
12137 #else
12138 __ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12139  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12140  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12141  int32x2_t __ret;
12142  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
12143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12144  return __ret;
12145 }
12146 #endif
12147 
12148 #ifdef __LITTLE_ENDIAN__
12149 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12150  int16x4_t __ret;
12151  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
12152  return __ret;
12153 }
12154 #else
12155 __ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12156  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12157  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12158  int16x4_t __ret;
12159  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
12160  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12161  return __ret;
12162 }
12163 #endif
12164 
12165 #ifdef __LITTLE_ENDIAN__
12166 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12167  uint8x16_t __ret;
12168  __ret = __p0 + __p1 * __p2;
12169  return __ret;
12170 }
12171 #else
12172 __ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12173  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12174  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12175  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12176  uint8x16_t __ret;
12177  __ret = __rev0 + __rev1 * __rev2;
12178  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12179  return __ret;
12180 }
12181 #endif
12182 
12183 #ifdef __LITTLE_ENDIAN__
12184 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12185  uint32x4_t __ret;
12186  __ret = __p0 + __p1 * __p2;
12187  return __ret;
12188 }
12189 #else
12190 __ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12191  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12192  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12193  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12194  uint32x4_t __ret;
12195  __ret = __rev0 + __rev1 * __rev2;
12196  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12197  return __ret;
12198 }
12199 #endif
12200 
12201 #ifdef __LITTLE_ENDIAN__
12202 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12203  uint16x8_t __ret;
12204  __ret = __p0 + __p1 * __p2;
12205  return __ret;
12206 }
12207 #else
12208 __ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12209  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12210  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12211  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12212  uint16x8_t __ret;
12213  __ret = __rev0 + __rev1 * __rev2;
12214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12215  return __ret;
12216 }
12217 #endif
12218 
12219 #ifdef __LITTLE_ENDIAN__
12220 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12221  int8x16_t __ret;
12222  __ret = __p0 + __p1 * __p2;
12223  return __ret;
12224 }
12225 #else
12226 __ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12227  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12228  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12229  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12230  int8x16_t __ret;
12231  __ret = __rev0 + __rev1 * __rev2;
12232  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12233  return __ret;
12234 }
12235 #endif
12236 
12237 #ifdef __LITTLE_ENDIAN__
12238 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12239  float32x4_t __ret;
12240  __ret = __p0 + __p1 * __p2;
12241  return __ret;
12242 }
12243 #else
12244 __ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12245  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12246  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12247  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12248  float32x4_t __ret;
12249  __ret = __rev0 + __rev1 * __rev2;
12250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12251  return __ret;
12252 }
12253 #endif
12254 
12255 #ifdef __LITTLE_ENDIAN__
12256 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12257  int32x4_t __ret;
12258  __ret = __p0 + __p1 * __p2;
12259  return __ret;
12260 }
12261 #else
12262 __ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12263  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12264  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12265  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12266  int32x4_t __ret;
12267  __ret = __rev0 + __rev1 * __rev2;
12268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12269  return __ret;
12270 }
12271 #endif
12272 
12273 #ifdef __LITTLE_ENDIAN__
12274 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12275  int16x8_t __ret;
12276  __ret = __p0 + __p1 * __p2;
12277  return __ret;
12278 }
12279 #else
12280 __ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12281  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12282  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12283  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12284  int16x8_t __ret;
12285  __ret = __rev0 + __rev1 * __rev2;
12286  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12287  return __ret;
12288 }
12289 #endif
12290 
12291 #ifdef __LITTLE_ENDIAN__
12292 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12293  uint8x8_t __ret;
12294  __ret = __p0 + __p1 * __p2;
12295  return __ret;
12296 }
12297 #else
12298 __ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12299  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12300  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12301  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12302  uint8x8_t __ret;
12303  __ret = __rev0 + __rev1 * __rev2;
12304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12305  return __ret;
12306 }
12307 #endif
12308 
12309 #ifdef __LITTLE_ENDIAN__
12310 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12311  uint32x2_t __ret;
12312  __ret = __p0 + __p1 * __p2;
12313  return __ret;
12314 }
12315 #else
12316 __ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12317  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12318  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12319  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12320  uint32x2_t __ret;
12321  __ret = __rev0 + __rev1 * __rev2;
12322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12323  return __ret;
12324 }
12325 #endif
12326 
12327 #ifdef __LITTLE_ENDIAN__
12328 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12329  uint16x4_t __ret;
12330  __ret = __p0 + __p1 * __p2;
12331  return __ret;
12332 }
12333 #else
12334 __ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12335  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12336  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12337  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12338  uint16x4_t __ret;
12339  __ret = __rev0 + __rev1 * __rev2;
12340  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12341  return __ret;
12342 }
12343 #endif
12344 
12345 #ifdef __LITTLE_ENDIAN__
12346 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12347  int8x8_t __ret;
12348  __ret = __p0 + __p1 * __p2;
12349  return __ret;
12350 }
12351 #else
12352 __ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12353  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12354  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12355  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12356  int8x8_t __ret;
12357  __ret = __rev0 + __rev1 * __rev2;
12358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12359  return __ret;
12360 }
12361 #endif
12362 
12363 #ifdef __LITTLE_ENDIAN__
12364 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12365  float32x2_t __ret;
12366  __ret = __p0 + __p1 * __p2;
12367  return __ret;
12368 }
12369 #else
12370 __ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12371  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12372  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12373  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12374  float32x2_t __ret;
12375  __ret = __rev0 + __rev1 * __rev2;
12376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12377  return __ret;
12378 }
12379 #endif
12380 
12381 #ifdef __LITTLE_ENDIAN__
12382 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12383  int32x2_t __ret;
12384  __ret = __p0 + __p1 * __p2;
12385  return __ret;
12386 }
12387 #else
12388 __ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12389  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12390  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12391  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12392  int32x2_t __ret;
12393  __ret = __rev0 + __rev1 * __rev2;
12394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12395  return __ret;
12396 }
12397 #endif
12398 
12399 #ifdef __LITTLE_ENDIAN__
12400 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12401  int16x4_t __ret;
12402  __ret = __p0 + __p1 * __p2;
12403  return __ret;
12404 }
12405 #else
12406 __ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12407  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12408  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12409  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12410  int16x4_t __ret;
12411  __ret = __rev0 + __rev1 * __rev2;
12412  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12413  return __ret;
12414 }
12415 #endif
12416 
12417 #ifdef __LITTLE_ENDIAN__
12418 #define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12419  uint32x4_t __s0 = __p0; \
12420  uint32x4_t __s1 = __p1; \
12421  uint32x2_t __s2 = __p2; \
12422  uint32x4_t __ret; \
12423  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12424  __ret; \
12425 })
12426 #else
12427 #define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12428  uint32x4_t __s0 = __p0; \
12429  uint32x4_t __s1 = __p1; \
12430  uint32x2_t __s2 = __p2; \
12431  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12432  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12433  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12434  uint32x4_t __ret; \
12435  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12436  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12437  __ret; \
12438 })
12439 #endif
12440 
12441 #ifdef __LITTLE_ENDIAN__
12442 #define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12443  uint16x8_t __s0 = __p0; \
12444  uint16x8_t __s1 = __p1; \
12445  uint16x4_t __s2 = __p2; \
12446  uint16x8_t __ret; \
12447  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12448  __ret; \
12449 })
12450 #else
12451 #define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12452  uint16x8_t __s0 = __p0; \
12453  uint16x8_t __s1 = __p1; \
12454  uint16x4_t __s2 = __p2; \
12455  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12456  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12457  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12458  uint16x8_t __ret; \
12459  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12460  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12461  __ret; \
12462 })
12463 #endif
12464 
12465 #ifdef __LITTLE_ENDIAN__
12466 #define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12467  float32x4_t __s0 = __p0; \
12468  float32x4_t __s1 = __p1; \
12469  float32x2_t __s2 = __p2; \
12470  float32x4_t __ret; \
12471  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12472  __ret; \
12473 })
12474 #else
12475 #define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12476  float32x4_t __s0 = __p0; \
12477  float32x4_t __s1 = __p1; \
12478  float32x2_t __s2 = __p2; \
12479  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12480  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12481  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12482  float32x4_t __ret; \
12483  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12485  __ret; \
12486 })
12487 #endif
12488 
12489 #ifdef __LITTLE_ENDIAN__
12490 #define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12491  int32x4_t __s0 = __p0; \
12492  int32x4_t __s1 = __p1; \
12493  int32x2_t __s2 = __p2; \
12494  int32x4_t __ret; \
12495  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12496  __ret; \
12497 })
12498 #else
12499 #define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12500  int32x4_t __s0 = __p0; \
12501  int32x4_t __s1 = __p1; \
12502  int32x2_t __s2 = __p2; \
12503  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12504  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12505  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12506  int32x4_t __ret; \
12507  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12509  __ret; \
12510 })
12511 #endif
12512 
12513 #ifdef __LITTLE_ENDIAN__
12514 #define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12515  int16x8_t __s0 = __p0; \
12516  int16x8_t __s1 = __p1; \
12517  int16x4_t __s2 = __p2; \
12518  int16x8_t __ret; \
12519  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12520  __ret; \
12521 })
12522 #else
12523 #define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12524  int16x8_t __s0 = __p0; \
12525  int16x8_t __s1 = __p1; \
12526  int16x4_t __s2 = __p2; \
12527  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12528  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12529  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12530  int16x8_t __ret; \
12531  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12533  __ret; \
12534 })
12535 #endif
12536 
12537 #ifdef __LITTLE_ENDIAN__
12538 #define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12539  uint32x2_t __s0 = __p0; \
12540  uint32x2_t __s1 = __p1; \
12541  uint32x2_t __s2 = __p2; \
12542  uint32x2_t __ret; \
12543  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12544  __ret; \
12545 })
12546 #else
12547 #define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12548  uint32x2_t __s0 = __p0; \
12549  uint32x2_t __s1 = __p1; \
12550  uint32x2_t __s2 = __p2; \
12551  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12552  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12553  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12554  uint32x2_t __ret; \
12555  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12556  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12557  __ret; \
12558 })
12559 #endif
12560 
12561 #ifdef __LITTLE_ENDIAN__
12562 #define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12563  uint16x4_t __s0 = __p0; \
12564  uint16x4_t __s1 = __p1; \
12565  uint16x4_t __s2 = __p2; \
12566  uint16x4_t __ret; \
12567  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12568  __ret; \
12569 })
12570 #else
12571 #define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12572  uint16x4_t __s0 = __p0; \
12573  uint16x4_t __s1 = __p1; \
12574  uint16x4_t __s2 = __p2; \
12575  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12576  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12577  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12578  uint16x4_t __ret; \
12579  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12581  __ret; \
12582 })
12583 #endif
12584 
12585 #ifdef __LITTLE_ENDIAN__
12586 #define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12587  float32x2_t __s0 = __p0; \
12588  float32x2_t __s1 = __p1; \
12589  float32x2_t __s2 = __p2; \
12590  float32x2_t __ret; \
12591  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12592  __ret; \
12593 })
12594 #else
12595 #define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12596  float32x2_t __s0 = __p0; \
12597  float32x2_t __s1 = __p1; \
12598  float32x2_t __s2 = __p2; \
12599  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12600  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12601  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12602  float32x2_t __ret; \
12603  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12605  __ret; \
12606 })
12607 #endif
12608 
12609 #ifdef __LITTLE_ENDIAN__
12610 #define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12611  int32x2_t __s0 = __p0; \
12612  int32x2_t __s1 = __p1; \
12613  int32x2_t __s2 = __p2; \
12614  int32x2_t __ret; \
12615  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12616  __ret; \
12617 })
12618 #else
12619 #define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12620  int32x2_t __s0 = __p0; \
12621  int32x2_t __s1 = __p1; \
12622  int32x2_t __s2 = __p2; \
12623  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12624  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12625  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12626  int32x2_t __ret; \
12627  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12628  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12629  __ret; \
12630 })
12631 #endif
12632 
12633 #ifdef __LITTLE_ENDIAN__
12634 #define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12635  int16x4_t __s0 = __p0; \
12636  int16x4_t __s1 = __p1; \
12637  int16x4_t __s2 = __p2; \
12638  int16x4_t __ret; \
12639  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12640  __ret; \
12641 })
12642 #else
12643 #define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12644  int16x4_t __s0 = __p0; \
12645  int16x4_t __s1 = __p1; \
12646  int16x4_t __s2 = __p2; \
12647  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12648  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12649  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12650  int16x4_t __ret; \
12651  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12653  __ret; \
12654 })
12655 #endif
12656 
12657 #ifdef __LITTLE_ENDIAN__
12658 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12659  uint32x4_t __ret;
12660  __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12661  return __ret;
12662 }
12663 #else
12664 __ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12665  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12666  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12667  uint32x4_t __ret;
12668  __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12669  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12670  return __ret;
12671 }
12672 #endif
12673 
12674 #ifdef __LITTLE_ENDIAN__
12675 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12676  uint16x8_t __ret;
12677  __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12678  return __ret;
12679 }
12680 #else
12681 __ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12682  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12683  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12684  uint16x8_t __ret;
12685  __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12686  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12687  return __ret;
12688 }
12689 #endif
12690 
12691 #ifdef __LITTLE_ENDIAN__
12692 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12693  float32x4_t __ret;
12694  __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12695  return __ret;
12696 }
12697 #else
12698 __ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12699  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12700  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12701  float32x4_t __ret;
12702  __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12703  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12704  return __ret;
12705 }
12706 #endif
12707 
12708 #ifdef __LITTLE_ENDIAN__
12709 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12710  int32x4_t __ret;
12711  __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12712  return __ret;
12713 }
12714 #else
12715 __ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12716  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12717  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12718  int32x4_t __ret;
12719  __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12721  return __ret;
12722 }
12723 #endif
12724 
12725 #ifdef __LITTLE_ENDIAN__
12726 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12727  int16x8_t __ret;
12728  __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12729  return __ret;
12730 }
12731 #else
12732 __ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12733  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12734  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12735  int16x8_t __ret;
12736  __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12737  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12738  return __ret;
12739 }
12740 #endif
12741 
12742 #ifdef __LITTLE_ENDIAN__
12743 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12744  uint32x2_t __ret;
12745  __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
12746  return __ret;
12747 }
12748 #else
12749 __ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12750  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12751  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12752  uint32x2_t __ret;
12753  __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
12754  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12755  return __ret;
12756 }
12757 #endif
12758 
12759 #ifdef __LITTLE_ENDIAN__
12760 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12761  uint16x4_t __ret;
12762  __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12763  return __ret;
12764 }
12765 #else
12766 __ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12767  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12768  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12769  uint16x4_t __ret;
12770  __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12771  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12772  return __ret;
12773 }
12774 #endif
12775 
12776 #ifdef __LITTLE_ENDIAN__
12777 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12778  float32x2_t __ret;
12779  __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
12780  return __ret;
12781 }
12782 #else
12783 __ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12784  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12785  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12786  float32x2_t __ret;
12787  __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
12788  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12789  return __ret;
12790 }
12791 #endif
12792 
12793 #ifdef __LITTLE_ENDIAN__
12794 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12795  int32x2_t __ret;
12796  __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
12797  return __ret;
12798 }
12799 #else
12800 __ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12801  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12802  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12803  int32x2_t __ret;
12804  __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
12805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12806  return __ret;
12807 }
12808 #endif
12809 
12810 #ifdef __LITTLE_ENDIAN__
12811 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12812  int16x4_t __ret;
12813  __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12814  return __ret;
12815 }
12816 #else
12817 __ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12818  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12819  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12820  int16x4_t __ret;
12821  __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12822  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12823  return __ret;
12824 }
12825 #endif
12826 
12827 #ifdef __LITTLE_ENDIAN__
12828 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12829  uint8x16_t __ret;
12830  __ret = __p0 - __p1 * __p2;
12831  return __ret;
12832 }
12833 #else
12834 __ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12835  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12836  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12837  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12838  uint8x16_t __ret;
12839  __ret = __rev0 - __rev1 * __rev2;
12840  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12841  return __ret;
12842 }
12843 #endif
12844 
12845 #ifdef __LITTLE_ENDIAN__
12846 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12847  uint32x4_t __ret;
12848  __ret = __p0 - __p1 * __p2;
12849  return __ret;
12850 }
12851 #else
12852 __ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12853  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12854  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12855  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12856  uint32x4_t __ret;
12857  __ret = __rev0 - __rev1 * __rev2;
12858  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12859  return __ret;
12860 }
12861 #endif
12862 
12863 #ifdef __LITTLE_ENDIAN__
12864 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12865  uint16x8_t __ret;
12866  __ret = __p0 - __p1 * __p2;
12867  return __ret;
12868 }
12869 #else
12870 __ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12871  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12872  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12873  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12874  uint16x8_t __ret;
12875  __ret = __rev0 - __rev1 * __rev2;
12876  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12877  return __ret;
12878 }
12879 #endif
12880 
12881 #ifdef __LITTLE_ENDIAN__
12882 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12883  int8x16_t __ret;
12884  __ret = __p0 - __p1 * __p2;
12885  return __ret;
12886 }
12887 #else
12888 __ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12889  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12890  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12891  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12892  int8x16_t __ret;
12893  __ret = __rev0 - __rev1 * __rev2;
12894  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12895  return __ret;
12896 }
12897 #endif
12898 
12899 #ifdef __LITTLE_ENDIAN__
12900 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12901  float32x4_t __ret;
12902  __ret = __p0 - __p1 * __p2;
12903  return __ret;
12904 }
12905 #else
12906 __ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12907  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12908  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12909  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12910  float32x4_t __ret;
12911  __ret = __rev0 - __rev1 * __rev2;
12912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12913  return __ret;
12914 }
12915 #endif
12916 
12917 #ifdef __LITTLE_ENDIAN__
12918 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12919  int32x4_t __ret;
12920  __ret = __p0 - __p1 * __p2;
12921  return __ret;
12922 }
12923 #else
12924 __ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12925  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12926  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12927  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12928  int32x4_t __ret;
12929  __ret = __rev0 - __rev1 * __rev2;
12930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12931  return __ret;
12932 }
12933 #endif
12934 
12935 #ifdef __LITTLE_ENDIAN__
12936 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12937  int16x8_t __ret;
12938  __ret = __p0 - __p1 * __p2;
12939  return __ret;
12940 }
12941 #else
12942 __ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12943  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12944  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12945  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12946  int16x8_t __ret;
12947  __ret = __rev0 - __rev1 * __rev2;
12948  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12949  return __ret;
12950 }
12951 #endif
12952 
12953 #ifdef __LITTLE_ENDIAN__
12954 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12955  uint8x8_t __ret;
12956  __ret = __p0 - __p1 * __p2;
12957  return __ret;
12958 }
12959 #else
12960 __ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12961  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12962  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12963  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12964  uint8x8_t __ret;
12965  __ret = __rev0 - __rev1 * __rev2;
12966  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12967  return __ret;
12968 }
12969 #endif
12970 
12971 #ifdef __LITTLE_ENDIAN__
12972 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12973  uint32x2_t __ret;
12974  __ret = __p0 - __p1 * __p2;
12975  return __ret;
12976 }
12977 #else
12978 __ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12979  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12980  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12981  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12982  uint32x2_t __ret;
12983  __ret = __rev0 - __rev1 * __rev2;
12984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12985  return __ret;
12986 }
12987 #endif
12988 
12989 #ifdef __LITTLE_ENDIAN__
12990 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12991  uint16x4_t __ret;
12992  __ret = __p0 - __p1 * __p2;
12993  return __ret;
12994 }
12995 #else
12996 __ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12997  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12998  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12999  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13000  uint16x4_t __ret;
13001  __ret = __rev0 - __rev1 * __rev2;
13002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13003  return __ret;
13004 }
13005 #endif
13006 
13007 #ifdef __LITTLE_ENDIAN__
13008 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13009  int8x8_t __ret;
13010  __ret = __p0 - __p1 * __p2;
13011  return __ret;
13012 }
13013 #else
13014 __ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13015  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13016  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13017  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
13018  int8x8_t __ret;
13019  __ret = __rev0 - __rev1 * __rev2;
13020  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13021  return __ret;
13022 }
13023 #endif
13024 
13025 #ifdef __LITTLE_ENDIAN__
13026 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13027  float32x2_t __ret;
13028  __ret = __p0 - __p1 * __p2;
13029  return __ret;
13030 }
13031 #else
13032 __ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13033  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13034  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13035  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13036  float32x2_t __ret;
13037  __ret = __rev0 - __rev1 * __rev2;
13038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13039  return __ret;
13040 }
13041 #endif
13042 
13043 #ifdef __LITTLE_ENDIAN__
13044 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13045  int32x2_t __ret;
13046  __ret = __p0 - __p1 * __p2;
13047  return __ret;
13048 }
13049 #else
13050 __ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13051  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13052  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13053  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13054  int32x2_t __ret;
13055  __ret = __rev0 - __rev1 * __rev2;
13056  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13057  return __ret;
13058 }
13059 #endif
13060 
13061 #ifdef __LITTLE_ENDIAN__
13062 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13063  int16x4_t __ret;
13064  __ret = __p0 - __p1 * __p2;
13065  return __ret;
13066 }
13067 #else
13068 __ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13069  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13070  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13071  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13072  int16x4_t __ret;
13073  __ret = __rev0 - __rev1 * __rev2;
13074  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13075  return __ret;
13076 }
13077 #endif
13078 
13079 #ifdef __LITTLE_ENDIAN__
13080 #define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13081  uint32x4_t __s0 = __p0; \
13082  uint32x4_t __s1 = __p1; \
13083  uint32x2_t __s2 = __p2; \
13084  uint32x4_t __ret; \
13085  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13086  __ret; \
13087 })
13088 #else
13089 #define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13090  uint32x4_t __s0 = __p0; \
13091  uint32x4_t __s1 = __p1; \
13092  uint32x2_t __s2 = __p2; \
13093  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13094  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13095  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13096  uint32x4_t __ret; \
13097  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13099  __ret; \
13100 })
13101 #endif
13102 
13103 #ifdef __LITTLE_ENDIAN__
13104 #define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13105  uint16x8_t __s0 = __p0; \
13106  uint16x8_t __s1 = __p1; \
13107  uint16x4_t __s2 = __p2; \
13108  uint16x8_t __ret; \
13109  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13110  __ret; \
13111 })
13112 #else
13113 #define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13114  uint16x8_t __s0 = __p0; \
13115  uint16x8_t __s1 = __p1; \
13116  uint16x4_t __s2 = __p2; \
13117  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13118  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13119  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13120  uint16x8_t __ret; \
13121  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13122  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13123  __ret; \
13124 })
13125 #endif
13126 
13127 #ifdef __LITTLE_ENDIAN__
13128 #define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13129  float32x4_t __s0 = __p0; \
13130  float32x4_t __s1 = __p1; \
13131  float32x2_t __s2 = __p2; \
13132  float32x4_t __ret; \
13133  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13134  __ret; \
13135 })
13136 #else
13137 #define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13138  float32x4_t __s0 = __p0; \
13139  float32x4_t __s1 = __p1; \
13140  float32x2_t __s2 = __p2; \
13141  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13142  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13143  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13144  float32x4_t __ret; \
13145  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13146  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13147  __ret; \
13148 })
13149 #endif
13150 
13151 #ifdef __LITTLE_ENDIAN__
13152 #define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13153  int32x4_t __s0 = __p0; \
13154  int32x4_t __s1 = __p1; \
13155  int32x2_t __s2 = __p2; \
13156  int32x4_t __ret; \
13157  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13158  __ret; \
13159 })
13160 #else
13161 #define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13162  int32x4_t __s0 = __p0; \
13163  int32x4_t __s1 = __p1; \
13164  int32x2_t __s2 = __p2; \
13165  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13166  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13167  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13168  int32x4_t __ret; \
13169  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13171  __ret; \
13172 })
13173 #endif
13174 
13175 #ifdef __LITTLE_ENDIAN__
13176 #define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13177  int16x8_t __s0 = __p0; \
13178  int16x8_t __s1 = __p1; \
13179  int16x4_t __s2 = __p2; \
13180  int16x8_t __ret; \
13181  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13182  __ret; \
13183 })
13184 #else
13185 #define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13186  int16x8_t __s0 = __p0; \
13187  int16x8_t __s1 = __p1; \
13188  int16x4_t __s2 = __p2; \
13189  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13190  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13191  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13192  int16x8_t __ret; \
13193  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13194  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13195  __ret; \
13196 })
13197 #endif
13198 
13199 #ifdef __LITTLE_ENDIAN__
13200 #define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13201  uint32x2_t __s0 = __p0; \
13202  uint32x2_t __s1 = __p1; \
13203  uint32x2_t __s2 = __p2; \
13204  uint32x2_t __ret; \
13205  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13206  __ret; \
13207 })
13208 #else
13209 #define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13210  uint32x2_t __s0 = __p0; \
13211  uint32x2_t __s1 = __p1; \
13212  uint32x2_t __s2 = __p2; \
13213  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13214  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13215  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13216  uint32x2_t __ret; \
13217  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13218  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13219  __ret; \
13220 })
13221 #endif
13222 
13223 #ifdef __LITTLE_ENDIAN__
13224 #define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13225  uint16x4_t __s0 = __p0; \
13226  uint16x4_t __s1 = __p1; \
13227  uint16x4_t __s2 = __p2; \
13228  uint16x4_t __ret; \
13229  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13230  __ret; \
13231 })
13232 #else
13233 #define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13234  uint16x4_t __s0 = __p0; \
13235  uint16x4_t __s1 = __p1; \
13236  uint16x4_t __s2 = __p2; \
13237  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13238  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13239  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13240  uint16x4_t __ret; \
13241  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13242  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13243  __ret; \
13244 })
13245 #endif
13246 
13247 #ifdef __LITTLE_ENDIAN__
13248 #define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13249  float32x2_t __s0 = __p0; \
13250  float32x2_t __s1 = __p1; \
13251  float32x2_t __s2 = __p2; \
13252  float32x2_t __ret; \
13253  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13254  __ret; \
13255 })
13256 #else
13257 #define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13258  float32x2_t __s0 = __p0; \
13259  float32x2_t __s1 = __p1; \
13260  float32x2_t __s2 = __p2; \
13261  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13262  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13263  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13264  float32x2_t __ret; \
13265  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13266  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13267  __ret; \
13268 })
13269 #endif
13270 
13271 #ifdef __LITTLE_ENDIAN__
13272 #define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13273  int32x2_t __s0 = __p0; \
13274  int32x2_t __s1 = __p1; \
13275  int32x2_t __s2 = __p2; \
13276  int32x2_t __ret; \
13277  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13278  __ret; \
13279 })
13280 #else
13281 #define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13282  int32x2_t __s0 = __p0; \
13283  int32x2_t __s1 = __p1; \
13284  int32x2_t __s2 = __p2; \
13285  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13286  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13287  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13288  int32x2_t __ret; \
13289  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13291  __ret; \
13292 })
13293 #endif
13294 
13295 #ifdef __LITTLE_ENDIAN__
13296 #define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13297  int16x4_t __s0 = __p0; \
13298  int16x4_t __s1 = __p1; \
13299  int16x4_t __s2 = __p2; \
13300  int16x4_t __ret; \
13301  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13302  __ret; \
13303 })
13304 #else
13305 #define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13306  int16x4_t __s0 = __p0; \
13307  int16x4_t __s1 = __p1; \
13308  int16x4_t __s2 = __p2; \
13309  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13310  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13311  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13312  int16x4_t __ret; \
13313  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13315  __ret; \
13316 })
13317 #endif
13318 
13319 #ifdef __LITTLE_ENDIAN__
13320 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13321  uint32x4_t __ret;
13322  __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13323  return __ret;
13324 }
13325 #else
13326 __ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13327  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13328  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13329  uint32x4_t __ret;
13330  __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13332  return __ret;
13333 }
13334 #endif
13335 
13336 #ifdef __LITTLE_ENDIAN__
13337 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13338  uint16x8_t __ret;
13339  __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13340  return __ret;
13341 }
13342 #else
13343 __ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13344  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13345  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13346  uint16x8_t __ret;
13347  __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13348  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13349  return __ret;
13350 }
13351 #endif
13352 
13353 #ifdef __LITTLE_ENDIAN__
13354 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13355  float32x4_t __ret;
13356  __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13357  return __ret;
13358 }
13359 #else
13360 __ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13361  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13362  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13363  float32x4_t __ret;
13364  __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13366  return __ret;
13367 }
13368 #endif
13369 
13370 #ifdef __LITTLE_ENDIAN__
13371 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13372  int32x4_t __ret;
13373  __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13374  return __ret;
13375 }
13376 #else
13377 __ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13378  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13379  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13380  int32x4_t __ret;
13381  __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13383  return __ret;
13384 }
13385 #endif
13386 
13387 #ifdef __LITTLE_ENDIAN__
13388 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13389  int16x8_t __ret;
13390  __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13391  return __ret;
13392 }
13393 #else
13394 __ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13395  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13396  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13397  int16x8_t __ret;
13398  __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13400  return __ret;
13401 }
13402 #endif
13403 
13404 #ifdef __LITTLE_ENDIAN__
13405 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13406  uint32x2_t __ret;
13407  __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
13408  return __ret;
13409 }
13410 #else
13411 __ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13412  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13413  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13414  uint32x2_t __ret;
13415  __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
13416  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13417  return __ret;
13418 }
13419 #endif
13420 
13421 #ifdef __LITTLE_ENDIAN__
13422 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13423  uint16x4_t __ret;
13424  __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13425  return __ret;
13426 }
13427 #else
13428 __ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13429  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13430  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13431  uint16x4_t __ret;
13432  __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13434  return __ret;
13435 }
13436 #endif
13437 
13438 #ifdef __LITTLE_ENDIAN__
13439 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13440  float32x2_t __ret;
13441  __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
13442  return __ret;
13443 }
13444 #else
13445 __ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13446  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13447  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13448  float32x2_t __ret;
13449  __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
13450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13451  return __ret;
13452 }
13453 #endif
13454 
13455 #ifdef __LITTLE_ENDIAN__
13456 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13457  int32x2_t __ret;
13458  __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
13459  return __ret;
13460 }
13461 #else
13462 __ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13463  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13464  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13465  int32x2_t __ret;
13466  __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
13467  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13468  return __ret;
13469 }
13470 #endif
13471 
13472 #ifdef __LITTLE_ENDIAN__
13473 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13474  int16x4_t __ret;
13475  __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13476  return __ret;
13477 }
13478 #else
13479 __ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13480  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13481  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13482  int16x4_t __ret;
13483  __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13485  return __ret;
13486 }
13487 #endif
13488 
13489 #ifdef __LITTLE_ENDIAN__
13490 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13491  poly8x8_t __ret;
13492  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13493  return __ret;
13494 }
13495 #else
13496 __ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13497  poly8x8_t __ret;
13498  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13500  return __ret;
13501 }
13502 #endif
13503 
13504 #ifdef __LITTLE_ENDIAN__
13505 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13506  poly16x4_t __ret;
13507  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13508  return __ret;
13509 }
13510 #else
13511 __ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13512  poly16x4_t __ret;
13513  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13514  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13515  return __ret;
13516 }
13517 #endif
13518 
13519 #ifdef __LITTLE_ENDIAN__
13520 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13521  poly8x16_t __ret;
13522  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13523  return __ret;
13524 }
13525 #else
13526 __ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13527  poly8x16_t __ret;
13528  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13529  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13530  return __ret;
13531 }
13532 #endif
13533 
13534 #ifdef __LITTLE_ENDIAN__
13535 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13536  poly16x8_t __ret;
13537  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13538  return __ret;
13539 }
13540 #else
13541 __ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13542  poly16x8_t __ret;
13543  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13544  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13545  return __ret;
13546 }
13547 #endif
13548 
13549 #ifdef __LITTLE_ENDIAN__
13550 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13551  uint8x16_t __ret;
13552  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13553  return __ret;
13554 }
13555 #else
13556 __ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13557  uint8x16_t __ret;
13558  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13559  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13560  return __ret;
13561 }
13562 #endif
13563 
13564 #ifdef __LITTLE_ENDIAN__
13565 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13566  uint32x4_t __ret;
13567  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13568  return __ret;
13569 }
13570 #else
13571 __ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13572  uint32x4_t __ret;
13573  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13575  return __ret;
13576 }
13577 #endif
13578 
13579 #ifdef __LITTLE_ENDIAN__
13580 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13581  uint64x2_t __ret;
13582  __ret = (uint64x2_t) {__p0, __p0};
13583  return __ret;
13584 }
13585 #else
13586 __ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13587  uint64x2_t __ret;
13588  __ret = (uint64x2_t) {__p0, __p0};
13589  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13590  return __ret;
13591 }
13592 #endif
13593 
13594 #ifdef __LITTLE_ENDIAN__
13595 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13596  uint16x8_t __ret;
13597  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13598  return __ret;
13599 }
13600 #else
13601 __ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13602  uint16x8_t __ret;
13603  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13605  return __ret;
13606 }
13607 #endif
13608 
13609 #ifdef __LITTLE_ENDIAN__
13610 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
13611  int8x16_t __ret;
13612  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13613  return __ret;
13614 }
13615 #else
13616 __ai int8x16_t vmovq_n_s8(int8_t __p0) {
13617  int8x16_t __ret;
13618  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13619  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13620  return __ret;
13621 }
13622 #endif
13623 
13624 #ifdef __LITTLE_ENDIAN__
13625 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
13626  float32x4_t __ret;
13627  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13628  return __ret;
13629 }
13630 #else
13631 __ai float32x4_t vmovq_n_f32(float32_t __p0) {
13632  float32x4_t __ret;
13633  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13635  return __ret;
13636 }
13637 #endif
13638 
13639 #ifdef __LITTLE_ENDIAN__
13640 #define vmovq_n_f16(__p0) __extension__ ({ \
13641  float16_t __s0 = __p0; \
13642  float16x8_t __ret; \
13643  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13644  __ret; \
13645 })
13646 #else
13647 #define vmovq_n_f16(__p0) __extension__ ({ \
13648  float16_t __s0 = __p0; \
13649  float16x8_t __ret; \
13650  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13651  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13652  __ret; \
13653 })
13654 #endif
13655 
13656 #ifdef __LITTLE_ENDIAN__
13657 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
13658  int32x4_t __ret;
13659  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13660  return __ret;
13661 }
13662 #else
13663 __ai int32x4_t vmovq_n_s32(int32_t __p0) {
13664  int32x4_t __ret;
13665  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13666  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13667  return __ret;
13668 }
13669 #endif
13670 
13671 #ifdef __LITTLE_ENDIAN__
13672 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
13673  int64x2_t __ret;
13674  __ret = (int64x2_t) {__p0, __p0};
13675  return __ret;
13676 }
13677 #else
13678 __ai int64x2_t vmovq_n_s64(int64_t __p0) {
13679  int64x2_t __ret;
13680  __ret = (int64x2_t) {__p0, __p0};
13681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13682  return __ret;
13683 }
13684 #endif
13685 
13686 #ifdef __LITTLE_ENDIAN__
13687 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
13688  int16x8_t __ret;
13689  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13690  return __ret;
13691 }
13692 #else
13693 __ai int16x8_t vmovq_n_s16(int16_t __p0) {
13694  int16x8_t __ret;
13695  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13696  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13697  return __ret;
13698 }
13699 #endif
13700 
13701 #ifdef __LITTLE_ENDIAN__
13702 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13703  uint8x8_t __ret;
13704  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13705  return __ret;
13706 }
13707 #else
13708 __ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13709  uint8x8_t __ret;
13710  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13711  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13712  return __ret;
13713 }
13714 #endif
13715 
13716 #ifdef __LITTLE_ENDIAN__
13717 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13718  uint32x2_t __ret;
13719  __ret = (uint32x2_t) {__p0, __p0};
13720  return __ret;
13721 }
13722 #else
13723 __ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13724  uint32x2_t __ret;
13725  __ret = (uint32x2_t) {__p0, __p0};
13726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13727  return __ret;
13728 }
13729 #endif
13730 
13731 #ifdef __LITTLE_ENDIAN__
13732 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13733  uint64x1_t __ret;
13734  __ret = (uint64x1_t) {__p0};
13735  return __ret;
13736 }
13737 #else
13738 __ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13739  uint64x1_t __ret;
13740  __ret = (uint64x1_t) {__p0};
13741  return __ret;
13742 }
13743 #endif
13744 
13745 #ifdef __LITTLE_ENDIAN__
13746 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13747  uint16x4_t __ret;
13748  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13749  return __ret;
13750 }
13751 #else
13752 __ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13753  uint16x4_t __ret;
13754  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13755  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13756  return __ret;
13757 }
13758 #endif
13759 
13760 #ifdef __LITTLE_ENDIAN__
13761 __ai int8x8_t vmov_n_s8(int8_t __p0) {
13762  int8x8_t __ret;
13763  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13764  return __ret;
13765 }
13766 #else
13767 __ai int8x8_t vmov_n_s8(int8_t __p0) {
13768  int8x8_t __ret;
13769  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13770  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13771  return __ret;
13772 }
13773 #endif
13774 
13775 #ifdef __LITTLE_ENDIAN__
13776 __ai float32x2_t vmov_n_f32(float32_t __p0) {
13777  float32x2_t __ret;
13778  __ret = (float32x2_t) {__p0, __p0};
13779  return __ret;
13780 }
13781 #else
13782 __ai float32x2_t vmov_n_f32(float32_t __p0) {
13783  float32x2_t __ret;
13784  __ret = (float32x2_t) {__p0, __p0};
13785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13786  return __ret;
13787 }
13788 #endif
13789 
13790 #ifdef __LITTLE_ENDIAN__
13791 #define vmov_n_f16(__p0) __extension__ ({ \
13792  float16_t __s0 = __p0; \
13793  float16x4_t __ret; \
13794  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13795  __ret; \
13796 })
13797 #else
13798 #define vmov_n_f16(__p0) __extension__ ({ \
13799  float16_t __s0 = __p0; \
13800  float16x4_t __ret; \
13801  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13803  __ret; \
13804 })
13805 #endif
13806 
13807 #ifdef __LITTLE_ENDIAN__
13808 __ai int32x2_t vmov_n_s32(int32_t __p0) {
13809  int32x2_t __ret;
13810  __ret = (int32x2_t) {__p0, __p0};
13811  return __ret;
13812 }
13813 #else
13814 __ai int32x2_t vmov_n_s32(int32_t __p0) {
13815  int32x2_t __ret;
13816  __ret = (int32x2_t) {__p0, __p0};
13817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13818  return __ret;
13819 }
13820 #endif
13821 
13822 #ifdef __LITTLE_ENDIAN__
13823 __ai int64x1_t vmov_n_s64(int64_t __p0) {
13824  int64x1_t __ret;
13825  __ret = (int64x1_t) {__p0};
13826  return __ret;
13827 }
13828 #else
13829 __ai int64x1_t vmov_n_s64(int64_t __p0) {
13830  int64x1_t __ret;
13831  __ret = (int64x1_t) {__p0};
13832  return __ret;
13833 }
13834 #endif
13835 
13836 #ifdef __LITTLE_ENDIAN__
13837 __ai int16x4_t vmov_n_s16(int16_t __p0) {
13838  int16x4_t __ret;
13839  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13840  return __ret;
13841 }
13842 #else
13843 __ai int16x4_t vmov_n_s16(int16_t __p0) {
13844  int16x4_t __ret;
13845  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13846  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13847  return __ret;
13848 }
13849 #endif
13850 
13851 #ifdef __LITTLE_ENDIAN__
13852 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13853  uint16x8_t __ret;
13854  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13855  return __ret;
13856 }
13857 #else
13858 __ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13859  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13860  uint16x8_t __ret;
13861  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
13862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13863  return __ret;
13864 }
13865 __ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
13866  uint16x8_t __ret;
13867  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13868  return __ret;
13869 }
13870 #endif
13871 
13872 #ifdef __LITTLE_ENDIAN__
13873 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13874  uint64x2_t __ret;
13875  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13876  return __ret;
13877 }
13878 #else
13879 __ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13880  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13881  uint64x2_t __ret;
13882  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
13883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13884  return __ret;
13885 }
13886 __ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
13887  uint64x2_t __ret;
13888  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13889  return __ret;
13890 }
13891 #endif
13892 
13893 #ifdef __LITTLE_ENDIAN__
13894 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13895  uint32x4_t __ret;
13896  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13897  return __ret;
13898 }
13899 #else
13900 __ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13901  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13902  uint32x4_t __ret;
13903  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
13904  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13905  return __ret;
13906 }
13907 __ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
13908  uint32x4_t __ret;
13909  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13910  return __ret;
13911 }
13912 #endif
13913 
13914 #ifdef __LITTLE_ENDIAN__
13915 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
13916  int16x8_t __ret;
13917  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13918  return __ret;
13919 }
13920 #else
13921 __ai int16x8_t vmovl_s8(int8x8_t __p0) {
13922  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13923  int16x8_t __ret;
13924  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
13925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13926  return __ret;
13927 }
13928 __ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
13929  int16x8_t __ret;
13930  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13931  return __ret;
13932 }
13933 #endif
13934 
13935 #ifdef __LITTLE_ENDIAN__
13936 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
13937  int64x2_t __ret;
13938  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13939  return __ret;
13940 }
13941 #else
13942 __ai int64x2_t vmovl_s32(int32x2_t __p0) {
13943  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13944  int64x2_t __ret;
13945  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
13946  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13947  return __ret;
13948 }
13949 __ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
13950  int64x2_t __ret;
13951  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13952  return __ret;
13953 }
13954 #endif
13955 
13956 #ifdef __LITTLE_ENDIAN__
13957 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
13958  int32x4_t __ret;
13959  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13960  return __ret;
13961 }
13962 #else
13963 __ai int32x4_t vmovl_s16(int16x4_t __p0) {
13964  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13965  int32x4_t __ret;
13966  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
13967  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13968  return __ret;
13969 }
13970 __ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
13971  int32x4_t __ret;
13972  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13973  return __ret;
13974 }
13975 #endif
13976 
13977 #ifdef __LITTLE_ENDIAN__
13978 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13979  uint16x4_t __ret;
13980  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13981  return __ret;
13982 }
13983 #else
13984 __ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13985  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13986  uint16x4_t __ret;
13987  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
13988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13989  return __ret;
13990 }
13991 __ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
13992  uint16x4_t __ret;
13993  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13994  return __ret;
13995 }
13996 #endif
13997 
13998 #ifdef __LITTLE_ENDIAN__
13999 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14000  uint32x2_t __ret;
14001  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14002  return __ret;
14003 }
14004 #else
14005 __ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14006  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14007  uint32x2_t __ret;
14008  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
14009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14010  return __ret;
14011 }
14012 __ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
14013  uint32x2_t __ret;
14014  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14015  return __ret;
14016 }
14017 #endif
14018 
14019 #ifdef __LITTLE_ENDIAN__
14020 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14021  uint8x8_t __ret;
14022  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14023  return __ret;
14024 }
14025 #else
14026 __ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14027  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14028  uint8x8_t __ret;
14029  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
14030  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14031  return __ret;
14032 }
14033 __ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
14034  uint8x8_t __ret;
14035  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14036  return __ret;
14037 }
14038 #endif
14039 
14040 #ifdef __LITTLE_ENDIAN__
14041 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
14042  int16x4_t __ret;
14043  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14044  return __ret;
14045 }
14046 #else
14047 __ai int16x4_t vmovn_s32(int32x4_t __p0) {
14048  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14049  int16x4_t __ret;
14050  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
14051  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14052  return __ret;
14053 }
14054 __ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
14055  int16x4_t __ret;
14056  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14057  return __ret;
14058 }
14059 #endif
14060 
14061 #ifdef __LITTLE_ENDIAN__
14062 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
14063  int32x2_t __ret;
14064  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14065  return __ret;
14066 }
14067 #else
14068 __ai int32x2_t vmovn_s64(int64x2_t __p0) {
14069  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14070  int32x2_t __ret;
14071  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
14072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14073  return __ret;
14074 }
14075 __ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
14076  int32x2_t __ret;
14077  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14078  return __ret;
14079 }
14080 #endif
14081 
14082 #ifdef __LITTLE_ENDIAN__
14083 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
14084  int8x8_t __ret;
14085  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14086  return __ret;
14087 }
14088 #else
14089 __ai int8x8_t vmovn_s16(int16x8_t __p0) {
14090  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14091  int8x8_t __ret;
14092  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
14093  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14094  return __ret;
14095 }
14096 __ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
14097  int8x8_t __ret;
14098  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14099  return __ret;
14100 }
14101 #endif
14102 
14103 #ifdef __LITTLE_ENDIAN__
14104 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14105  uint8x16_t __ret;
14106  __ret = __p0 * __p1;
14107  return __ret;
14108 }
14109 #else
14110 __ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14111  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14112  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14113  uint8x16_t __ret;
14114  __ret = __rev0 * __rev1;
14115  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14116  return __ret;
14117 }
14118 #endif
14119 
14120 #ifdef __LITTLE_ENDIAN__
14121 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14122  uint32x4_t __ret;
14123  __ret = __p0 * __p1;
14124  return __ret;
14125 }
14126 #else
14127 __ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14128  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14129  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14130  uint32x4_t __ret;
14131  __ret = __rev0 * __rev1;
14132  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14133  return __ret;
14134 }
14135 #endif
14136 
14137 #ifdef __LITTLE_ENDIAN__
14138 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14139  uint16x8_t __ret;
14140  __ret = __p0 * __p1;
14141  return __ret;
14142 }
14143 #else
14144 __ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14145  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14146  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14147  uint16x8_t __ret;
14148  __ret = __rev0 * __rev1;
14149  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14150  return __ret;
14151 }
14152 #endif
14153 
14154 #ifdef __LITTLE_ENDIAN__
14155 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14156  int8x16_t __ret;
14157  __ret = __p0 * __p1;
14158  return __ret;
14159 }
14160 #else
14161 __ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14162  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14163  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14164  int8x16_t __ret;
14165  __ret = __rev0 * __rev1;
14166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14167  return __ret;
14168 }
14169 #endif
14170 
14171 #ifdef __LITTLE_ENDIAN__
14172 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14173  float32x4_t __ret;
14174  __ret = __p0 * __p1;
14175  return __ret;
14176 }
14177 #else
14178 __ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14179  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14180  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14181  float32x4_t __ret;
14182  __ret = __rev0 * __rev1;
14183  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14184  return __ret;
14185 }
14186 #endif
14187 
14188 #ifdef __LITTLE_ENDIAN__
14189 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14190  int32x4_t __ret;
14191  __ret = __p0 * __p1;
14192  return __ret;
14193 }
14194 #else
14195 __ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14196  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14197  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14198  int32x4_t __ret;
14199  __ret = __rev0 * __rev1;
14200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14201  return __ret;
14202 }
14203 #endif
14204 
14205 #ifdef __LITTLE_ENDIAN__
14206 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14207  int16x8_t __ret;
14208  __ret = __p0 * __p1;
14209  return __ret;
14210 }
14211 #else
14212 __ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14213  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14214  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14215  int16x8_t __ret;
14216  __ret = __rev0 * __rev1;
14217  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14218  return __ret;
14219 }
14220 #endif
14221 
14222 #ifdef __LITTLE_ENDIAN__
14223 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14224  uint8x8_t __ret;
14225  __ret = __p0 * __p1;
14226  return __ret;
14227 }
14228 #else
14229 __ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14230  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14231  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14232  uint8x8_t __ret;
14233  __ret = __rev0 * __rev1;
14234  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14235  return __ret;
14236 }
14237 #endif
14238 
14239 #ifdef __LITTLE_ENDIAN__
14240 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14241  uint32x2_t __ret;
14242  __ret = __p0 * __p1;
14243  return __ret;
14244 }
14245 #else
14246 __ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14247  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14248  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14249  uint32x2_t __ret;
14250  __ret = __rev0 * __rev1;
14251  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14252  return __ret;
14253 }
14254 #endif
14255 
14256 #ifdef __LITTLE_ENDIAN__
14257 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14258  uint16x4_t __ret;
14259  __ret = __p0 * __p1;
14260  return __ret;
14261 }
14262 #else
14263 __ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14264  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14265  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14266  uint16x4_t __ret;
14267  __ret = __rev0 * __rev1;
14268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14269  return __ret;
14270 }
14271 #endif
14272 
14273 #ifdef __LITTLE_ENDIAN__
14274 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14275  int8x8_t __ret;
14276  __ret = __p0 * __p1;
14277  return __ret;
14278 }
14279 #else
14280 __ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14281  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14282  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14283  int8x8_t __ret;
14284  __ret = __rev0 * __rev1;
14285  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14286  return __ret;
14287 }
14288 #endif
14289 
14290 #ifdef __LITTLE_ENDIAN__
14291 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14292  float32x2_t __ret;
14293  __ret = __p0 * __p1;
14294  return __ret;
14295 }
14296 #else
14297 __ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14298  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14299  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14300  float32x2_t __ret;
14301  __ret = __rev0 * __rev1;
14302  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14303  return __ret;
14304 }
14305 #endif
14306 
14307 #ifdef __LITTLE_ENDIAN__
14308 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14309  int32x2_t __ret;
14310  __ret = __p0 * __p1;
14311  return __ret;
14312 }
14313 #else
14314 __ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14315  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14316  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14317  int32x2_t __ret;
14318  __ret = __rev0 * __rev1;
14319  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14320  return __ret;
14321 }
14322 #endif
14323 
14324 #ifdef __LITTLE_ENDIAN__
14325 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14326  int16x4_t __ret;
14327  __ret = __p0 * __p1;
14328  return __ret;
14329 }
14330 #else
14331 __ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14332  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14333  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14334  int16x4_t __ret;
14335  __ret = __rev0 * __rev1;
14336  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14337  return __ret;
14338 }
14339 #endif
14340 
14341 #ifdef __LITTLE_ENDIAN__
14342 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14343  poly8x8_t __ret;
14344  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
14345  return __ret;
14346 }
14347 #else
14348 __ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14349  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14350  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14351  poly8x8_t __ret;
14352  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
14353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14354  return __ret;
14355 }
14356 #endif
14357 
14358 #ifdef __LITTLE_ENDIAN__
14359 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14360  poly8x16_t __ret;
14361  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
14362  return __ret;
14363 }
14364 #else
14365 __ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14366  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14367  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14368  poly8x16_t __ret;
14369  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
14370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14371  return __ret;
14372 }
14373 #endif
14374 
14375 #ifdef __LITTLE_ENDIAN__
14376 #define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14377  uint32x4_t __s0 = __p0; \
14378  uint32x2_t __s1 = __p1; \
14379  uint32x4_t __ret; \
14380  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14381  __ret; \
14382 })
14383 #else
14384 #define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14385  uint32x4_t __s0 = __p0; \
14386  uint32x2_t __s1 = __p1; \
14387  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14388  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14389  uint32x4_t __ret; \
14390  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14391  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14392  __ret; \
14393 })
14394 #endif
14395 
14396 #ifdef __LITTLE_ENDIAN__
14397 #define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14398  uint16x8_t __s0 = __p0; \
14399  uint16x4_t __s1 = __p1; \
14400  uint16x8_t __ret; \
14401  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14402  __ret; \
14403 })
14404 #else
14405 #define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14406  uint16x8_t __s0 = __p0; \
14407  uint16x4_t __s1 = __p1; \
14408  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14409  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14410  uint16x8_t __ret; \
14411  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14412  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14413  __ret; \
14414 })
14415 #endif
14416 
14417 #ifdef __LITTLE_ENDIAN__
14418 #define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14419  float32x4_t __s0 = __p0; \
14420  float32x2_t __s1 = __p1; \
14421  float32x4_t __ret; \
14422  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14423  __ret; \
14424 })
14425 #else
14426 #define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14427  float32x4_t __s0 = __p0; \
14428  float32x2_t __s1 = __p1; \
14429  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14430  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14431  float32x4_t __ret; \
14432  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14434  __ret; \
14435 })
14436 #endif
14437 
14438 #ifdef __LITTLE_ENDIAN__
14439 #define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14440  int32x4_t __s0 = __p0; \
14441  int32x2_t __s1 = __p1; \
14442  int32x4_t __ret; \
14443  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14444  __ret; \
14445 })
14446 #else
14447 #define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14448  int32x4_t __s0 = __p0; \
14449  int32x2_t __s1 = __p1; \
14450  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14451  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14452  int32x4_t __ret; \
14453  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14454  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14455  __ret; \
14456 })
14457 #endif
14458 
14459 #ifdef __LITTLE_ENDIAN__
14460 #define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14461  int16x8_t __s0 = __p0; \
14462  int16x4_t __s1 = __p1; \
14463  int16x8_t __ret; \
14464  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14465  __ret; \
14466 })
14467 #else
14468 #define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14469  int16x8_t __s0 = __p0; \
14470  int16x4_t __s1 = __p1; \
14471  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14472  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14473  int16x8_t __ret; \
14474  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14476  __ret; \
14477 })
14478 #endif
14479 
14480 #ifdef __LITTLE_ENDIAN__
14481 #define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14482  uint32x2_t __s0 = __p0; \
14483  uint32x2_t __s1 = __p1; \
14484  uint32x2_t __ret; \
14485  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14486  __ret; \
14487 })
14488 #else
14489 #define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14490  uint32x2_t __s0 = __p0; \
14491  uint32x2_t __s1 = __p1; \
14492  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14493  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14494  uint32x2_t __ret; \
14495  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14496  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14497  __ret; \
14498 })
14499 #endif
14500 
14501 #ifdef __LITTLE_ENDIAN__
14502 #define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14503  uint16x4_t __s0 = __p0; \
14504  uint16x4_t __s1 = __p1; \
14505  uint16x4_t __ret; \
14506  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14507  __ret; \
14508 })
14509 #else
14510 #define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14511  uint16x4_t __s0 = __p0; \
14512  uint16x4_t __s1 = __p1; \
14513  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14514  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14515  uint16x4_t __ret; \
14516  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14517  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14518  __ret; \
14519 })
14520 #endif
14521 
14522 #ifdef __LITTLE_ENDIAN__
14523 #define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14524  float32x2_t __s0 = __p0; \
14525  float32x2_t __s1 = __p1; \
14526  float32x2_t __ret; \
14527  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14528  __ret; \
14529 })
14530 #else
14531 #define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14532  float32x2_t __s0 = __p0; \
14533  float32x2_t __s1 = __p1; \
14534  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14535  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14536  float32x2_t __ret; \
14537  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14538  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14539  __ret; \
14540 })
14541 #endif
14542 
14543 #ifdef __LITTLE_ENDIAN__
14544 #define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14545  int32x2_t __s0 = __p0; \
14546  int32x2_t __s1 = __p1; \
14547  int32x2_t __ret; \
14548  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14549  __ret; \
14550 })
14551 #else
14552 #define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14553  int32x2_t __s0 = __p0; \
14554  int32x2_t __s1 = __p1; \
14555  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14556  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14557  int32x2_t __ret; \
14558  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14559  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14560  __ret; \
14561 })
14562 #endif
14563 
14564 #ifdef __LITTLE_ENDIAN__
14565 #define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14566  int16x4_t __s0 = __p0; \
14567  int16x4_t __s1 = __p1; \
14568  int16x4_t __ret; \
14569  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14570  __ret; \
14571 })
14572 #else
14573 #define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14574  int16x4_t __s0 = __p0; \
14575  int16x4_t __s1 = __p1; \
14576  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14577  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14578  int16x4_t __ret; \
14579  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14581  __ret; \
14582 })
14583 #endif
14584 
14585 #ifdef __LITTLE_ENDIAN__
14586 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14587  uint32x4_t __ret;
14588  __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14589  return __ret;
14590 }
14591 #else
14592 __ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14593  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14594  uint32x4_t __ret;
14595  __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14597  return __ret;
14598 }
14599 #endif
14600 
14601 #ifdef __LITTLE_ENDIAN__
14602 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14603  uint16x8_t __ret;
14604  __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14605  return __ret;
14606 }
14607 #else
14608 __ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14609  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14610  uint16x8_t __ret;
14611  __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14612  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14613  return __ret;
14614 }
14615 #endif
14616 
14617 #ifdef __LITTLE_ENDIAN__
14618 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14619  float32x4_t __ret;
14620  __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14621  return __ret;
14622 }
14623 #else
14624 __ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14625  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14626  float32x4_t __ret;
14627  __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14628  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14629  return __ret;
14630 }
14631 #endif
14632 
14633 #ifdef __LITTLE_ENDIAN__
14634 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14635  int32x4_t __ret;
14636  __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14637  return __ret;
14638 }
14639 #else
14640 __ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14641  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14642  int32x4_t __ret;
14643  __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14645  return __ret;
14646 }
14647 #endif
14648 
14649 #ifdef __LITTLE_ENDIAN__
14650 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14651  int16x8_t __ret;
14652  __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14653  return __ret;
14654 }
14655 #else
14656 __ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14657  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14658  int16x8_t __ret;
14659  __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14660  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14661  return __ret;
14662 }
14663 #endif
14664 
14665 #ifdef __LITTLE_ENDIAN__
14666 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14667  uint32x2_t __ret;
14668  __ret = __p0 * (uint32x2_t) {__p1, __p1};
14669  return __ret;
14670 }
14671 #else
14672 __ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14673  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14674  uint32x2_t __ret;
14675  __ret = __rev0 * (uint32x2_t) {__p1, __p1};
14676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14677  return __ret;
14678 }
14679 #endif
14680 
14681 #ifdef __LITTLE_ENDIAN__
14682 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14683  uint16x4_t __ret;
14684  __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14685  return __ret;
14686 }
14687 #else
14688 __ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14689  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14690  uint16x4_t __ret;
14691  __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14692  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14693  return __ret;
14694 }
14695 #endif
14696 
14697 #ifdef __LITTLE_ENDIAN__
14698 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14699  float32x2_t __ret;
14700  __ret = __p0 * (float32x2_t) {__p1, __p1};
14701  return __ret;
14702 }
14703 #else
14704 __ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14705  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14706  float32x2_t __ret;
14707  __ret = __rev0 * (float32x2_t) {__p1, __p1};
14708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14709  return __ret;
14710 }
14711 #endif
14712 
14713 #ifdef __LITTLE_ENDIAN__
14714 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14715  int32x2_t __ret;
14716  __ret = __p0 * (int32x2_t) {__p1, __p1};
14717  return __ret;
14718 }
14719 #else
14720 __ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14721  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14722  int32x2_t __ret;
14723  __ret = __rev0 * (int32x2_t) {__p1, __p1};
14724  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14725  return __ret;
14726 }
14727 #endif
14728 
14729 #ifdef __LITTLE_ENDIAN__
14730 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14731  int16x4_t __ret;
14732  __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14733  return __ret;
14734 }
14735 #else
14736 __ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14737  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14738  int16x4_t __ret;
14739  __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14740  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14741  return __ret;
14742 }
14743 #endif
14744 
14745 #ifdef __LITTLE_ENDIAN__
14746 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14747  poly16x8_t __ret;
14748  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14749  return __ret;
14750 }
14751 #else
14752 __ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14753  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14754  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14755  poly16x8_t __ret;
14756  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
14757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14758  return __ret;
14759 }
14760 __ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14761  poly16x8_t __ret;
14762  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14763  return __ret;
14764 }
14765 #endif
14766 
14767 #ifdef __LITTLE_ENDIAN__
14768 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14769  uint16x8_t __ret;
14770  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14771  return __ret;
14772 }
14773 #else
14774 __ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14775  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14776  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14777  uint16x8_t __ret;
14778  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
14779  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14780  return __ret;
14781 }
14782 __ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14783  uint16x8_t __ret;
14784  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14785  return __ret;
14786 }
14787 #endif
14788 
14789 #ifdef __LITTLE_ENDIAN__
14790 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14791  uint64x2_t __ret;
14792  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14793  return __ret;
14794 }
14795 #else
14796 __ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14797  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14798  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14799  uint64x2_t __ret;
14800  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
14801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14802  return __ret;
14803 }
14804 __ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14805  uint64x2_t __ret;
14806  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14807  return __ret;
14808 }
14809 #endif
14810 
14811 #ifdef __LITTLE_ENDIAN__
14812 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14813  uint32x4_t __ret;
14814  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14815  return __ret;
14816 }
14817 #else
14818 __ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14819  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14820  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14821  uint32x4_t __ret;
14822  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
14823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14824  return __ret;
14825 }
14826 __ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14827  uint32x4_t __ret;
14828  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14829  return __ret;
14830 }
14831 #endif
14832 
14833 #ifdef __LITTLE_ENDIAN__
14834 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14835  int16x8_t __ret;
14836  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14837  return __ret;
14838 }
14839 #else
14840 __ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14841  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14842  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14843  int16x8_t __ret;
14844  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
14845  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14846  return __ret;
14847 }
14848 __ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14849  int16x8_t __ret;
14850  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14851  return __ret;
14852 }
14853 #endif
14854 
14855 #ifdef __LITTLE_ENDIAN__
14856 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14857  int64x2_t __ret;
14858  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14859  return __ret;
14860 }
14861 #else
14862 __ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14863  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14864  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14865  int64x2_t __ret;
14866  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
14867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14868  return __ret;
14869 }
14870 __ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14871  int64x2_t __ret;
14872  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14873  return __ret;
14874 }
14875 #endif
14876 
14877 #ifdef __LITTLE_ENDIAN__
14878 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14879  int32x4_t __ret;
14880  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14881  return __ret;
14882 }
14883 #else
14884 __ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14885  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14886  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14887  int32x4_t __ret;
14888  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
14889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14890  return __ret;
14891 }
14892 __ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14893  int32x4_t __ret;
14894  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14895  return __ret;
14896 }
14897 #endif
14898 
14899 #ifdef __LITTLE_ENDIAN__
14900 #define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14901  uint32x2_t __s0 = __p0; \
14902  uint32x2_t __s1 = __p1; \
14903  uint64x2_t __ret; \
14904  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14905  __ret; \
14906 })
14907 #else
14908 #define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14909  uint32x2_t __s0 = __p0; \
14910  uint32x2_t __s1 = __p1; \
14911  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14912  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14913  uint64x2_t __ret; \
14914  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14916  __ret; \
14917 })
14918 #endif
14919 
14920 #ifdef __LITTLE_ENDIAN__
14921 #define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14922  uint16x4_t __s0 = __p0; \
14923  uint16x4_t __s1 = __p1; \
14924  uint32x4_t __ret; \
14925  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14926  __ret; \
14927 })
14928 #else
14929 #define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14930  uint16x4_t __s0 = __p0; \
14931  uint16x4_t __s1 = __p1; \
14932  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14933  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14934  uint32x4_t __ret; \
14935  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14937  __ret; \
14938 })
14939 #endif
14940 
14941 #ifdef __LITTLE_ENDIAN__
14942 #define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14943  int32x2_t __s0 = __p0; \
14944  int32x2_t __s1 = __p1; \
14945  int64x2_t __ret; \
14946  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14947  __ret; \
14948 })
14949 #else
14950 #define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14951  int32x2_t __s0 = __p0; \
14952  int32x2_t __s1 = __p1; \
14953  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14954  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14955  int64x2_t __ret; \
14956  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14957  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14958  __ret; \
14959 })
14960 #endif
14961 
14962 #ifdef __LITTLE_ENDIAN__
14963 #define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14964  int16x4_t __s0 = __p0; \
14965  int16x4_t __s1 = __p1; \
14966  int32x4_t __ret; \
14967  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14968  __ret; \
14969 })
14970 #else
14971 #define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14972  int16x4_t __s0 = __p0; \
14973  int16x4_t __s1 = __p1; \
14974  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14975  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14976  int32x4_t __ret; \
14977  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14978  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14979  __ret; \
14980 })
14981 #endif
14982 
14983 #ifdef __LITTLE_ENDIAN__
14984 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14985  uint64x2_t __ret;
14986  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14987  return __ret;
14988 }
14989 #else
14990 __ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14991  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14992  uint64x2_t __ret;
14993  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14995  return __ret;
14996 }
14997 __ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14998  uint64x2_t __ret;
14999  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
15000  return __ret;
15001 }
15002 #endif
15003 
15004 #ifdef __LITTLE_ENDIAN__
15005 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15006  uint32x4_t __ret;
15007  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15008  return __ret;
15009 }
15010 #else
15011 __ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15012  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15013  uint32x4_t __ret;
15014  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15016  return __ret;
15017 }
15018 __ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15019  uint32x4_t __ret;
15020  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15021  return __ret;
15022 }
15023 #endif
15024 
15025 #ifdef __LITTLE_ENDIAN__
15026 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15027  int64x2_t __ret;
15028  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15029  return __ret;
15030 }
15031 #else
15032 __ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15033  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15034  int64x2_t __ret;
15035  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15036  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15037  return __ret;
15038 }
15039 __ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15040  int64x2_t __ret;
15041  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15042  return __ret;
15043 }
15044 #endif
15045 
15046 #ifdef __LITTLE_ENDIAN__
15047 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15048  int32x4_t __ret;
15049  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15050  return __ret;
15051 }
15052 #else
15053 __ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15054  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15055  int32x4_t __ret;
15056  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15057  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15058  return __ret;
15059 }
15060 __ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15061  int32x4_t __ret;
15062  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15063  return __ret;
15064 }
15065 #endif
15066 
15067 #ifdef __LITTLE_ENDIAN__
15068 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15069  poly8x8_t __ret;
15070  __ret = ~__p0;
15071  return __ret;
15072 }
15073 #else
15074 __ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15075  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15076  poly8x8_t __ret;
15077  __ret = ~__rev0;
15078  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15079  return __ret;
15080 }
15081 #endif
15082 
15083 #ifdef __LITTLE_ENDIAN__
15084 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15085  poly8x16_t __ret;
15086  __ret = ~__p0;
15087  return __ret;
15088 }
15089 #else
15090 __ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15091  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15092  poly8x16_t __ret;
15093  __ret = ~__rev0;
15094  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15095  return __ret;
15096 }
15097 #endif
15098 
15099 #ifdef __LITTLE_ENDIAN__
15100 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15101  uint8x16_t __ret;
15102  __ret = ~__p0;
15103  return __ret;
15104 }
15105 #else
15106 __ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15107  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15108  uint8x16_t __ret;
15109  __ret = ~__rev0;
15110  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15111  return __ret;
15112 }
15113 #endif
15114 
15115 #ifdef __LITTLE_ENDIAN__
15116 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15117  uint32x4_t __ret;
15118  __ret = ~__p0;
15119  return __ret;
15120 }
15121 #else
15122 __ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15123  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15124  uint32x4_t __ret;
15125  __ret = ~__rev0;
15126  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15127  return __ret;
15128 }
15129 #endif
15130 
15131 #ifdef __LITTLE_ENDIAN__
15132 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15133  uint16x8_t __ret;
15134  __ret = ~__p0;
15135  return __ret;
15136 }
15137 #else
15138 __ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15139  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15140  uint16x8_t __ret;
15141  __ret = ~__rev0;
15142  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15143  return __ret;
15144 }
15145 #endif
15146 
15147 #ifdef __LITTLE_ENDIAN__
15148 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15149  int8x16_t __ret;
15150  __ret = ~__p0;
15151  return __ret;
15152 }
15153 #else
15154 __ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15155  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15156  int8x16_t __ret;
15157  __ret = ~__rev0;
15158  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15159  return __ret;
15160 }
15161 #endif
15162 
15163 #ifdef __LITTLE_ENDIAN__
15164 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15165  int32x4_t __ret;
15166  __ret = ~__p0;
15167  return __ret;
15168 }
15169 #else
15170 __ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15171  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15172  int32x4_t __ret;
15173  __ret = ~__rev0;
15174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15175  return __ret;
15176 }
15177 #endif
15178 
15179 #ifdef __LITTLE_ENDIAN__
15180 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15181  int16x8_t __ret;
15182  __ret = ~__p0;
15183  return __ret;
15184 }
15185 #else
15186 __ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15187  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15188  int16x8_t __ret;
15189  __ret = ~__rev0;
15190  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15191  return __ret;
15192 }
15193 #endif
15194 
15195 #ifdef __LITTLE_ENDIAN__
15196 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15197  uint8x8_t __ret;
15198  __ret = ~__p0;
15199  return __ret;
15200 }
15201 #else
15202 __ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15203  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15204  uint8x8_t __ret;
15205  __ret = ~__rev0;
15206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15207  return __ret;
15208 }
15209 #endif
15210 
15211 #ifdef __LITTLE_ENDIAN__
15212 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15213  uint32x2_t __ret;
15214  __ret = ~__p0;
15215  return __ret;
15216 }
15217 #else
15218 __ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15219  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15220  uint32x2_t __ret;
15221  __ret = ~__rev0;
15222  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15223  return __ret;
15224 }
15225 #endif
15226 
15227 #ifdef __LITTLE_ENDIAN__
15228 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15229  uint16x4_t __ret;
15230  __ret = ~__p0;
15231  return __ret;
15232 }
15233 #else
15234 __ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15235  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15236  uint16x4_t __ret;
15237  __ret = ~__rev0;
15238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15239  return __ret;
15240 }
15241 #endif
15242 
15243 #ifdef __LITTLE_ENDIAN__
15244 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
15245  int8x8_t __ret;
15246  __ret = ~__p0;
15247  return __ret;
15248 }
15249 #else
15250 __ai int8x8_t vmvn_s8(int8x8_t __p0) {
15251  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15252  int8x8_t __ret;
15253  __ret = ~__rev0;
15254  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15255  return __ret;
15256 }
15257 #endif
15258 
15259 #ifdef __LITTLE_ENDIAN__
15260 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
15261  int32x2_t __ret;
15262  __ret = ~__p0;
15263  return __ret;
15264 }
15265 #else
15266 __ai int32x2_t vmvn_s32(int32x2_t __p0) {
15267  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15268  int32x2_t __ret;
15269  __ret = ~__rev0;
15270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15271  return __ret;
15272 }
15273 #endif
15274 
15275 #ifdef __LITTLE_ENDIAN__
15276 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
15277  int16x4_t __ret;
15278  __ret = ~__p0;
15279  return __ret;
15280 }
15281 #else
15282 __ai int16x4_t vmvn_s16(int16x4_t __p0) {
15283  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15284  int16x4_t __ret;
15285  __ret = ~__rev0;
15286  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15287  return __ret;
15288 }
15289 #endif
15290 
15291 #ifdef __LITTLE_ENDIAN__
15292 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
15293  int8x16_t __ret;
15294  __ret = -__p0;
15295  return __ret;
15296 }
15297 #else
15298 __ai int8x16_t vnegq_s8(int8x16_t __p0) {
15299  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15300  int8x16_t __ret;
15301  __ret = -__rev0;
15302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15303  return __ret;
15304 }
15305 #endif
15306 
15307 #ifdef __LITTLE_ENDIAN__
15308 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
15309  float32x4_t __ret;
15310  __ret = -__p0;
15311  return __ret;
15312 }
15313 #else
15314 __ai float32x4_t vnegq_f32(float32x4_t __p0) {
15315  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15316  float32x4_t __ret;
15317  __ret = -__rev0;
15318  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15319  return __ret;
15320 }
15321 #endif
15322 
15323 #ifdef __LITTLE_ENDIAN__
15324 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
15325  int32x4_t __ret;
15326  __ret = -__p0;
15327  return __ret;
15328 }
15329 #else
15330 __ai int32x4_t vnegq_s32(int32x4_t __p0) {
15331  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15332  int32x4_t __ret;
15333  __ret = -__rev0;
15334  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15335  return __ret;
15336 }
15337 #endif
15338 
15339 #ifdef __LITTLE_ENDIAN__
15340 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
15341  int16x8_t __ret;
15342  __ret = -__p0;
15343  return __ret;
15344 }
15345 #else
15346 __ai int16x8_t vnegq_s16(int16x8_t __p0) {
15347  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15348  int16x8_t __ret;
15349  __ret = -__rev0;
15350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15351  return __ret;
15352 }
15353 #endif
15354 
15355 #ifdef __LITTLE_ENDIAN__
15356 __ai int8x8_t vneg_s8(int8x8_t __p0) {
15357  int8x8_t __ret;
15358  __ret = -__p0;
15359  return __ret;
15360 }
15361 #else
15362 __ai int8x8_t vneg_s8(int8x8_t __p0) {
15363  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15364  int8x8_t __ret;
15365  __ret = -__rev0;
15366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15367  return __ret;
15368 }
15369 #endif
15370 
15371 #ifdef __LITTLE_ENDIAN__
15372 __ai float32x2_t vneg_f32(float32x2_t __p0) {
15373  float32x2_t __ret;
15374  __ret = -__p0;
15375  return __ret;
15376 }
15377 #else
15378 __ai float32x2_t vneg_f32(float32x2_t __p0) {
15379  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15380  float32x2_t __ret;
15381  __ret = -__rev0;
15382  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15383  return __ret;
15384 }
15385 #endif
15386 
15387 #ifdef __LITTLE_ENDIAN__
15388 __ai int32x2_t vneg_s32(int32x2_t __p0) {
15389  int32x2_t __ret;
15390  __ret = -__p0;
15391  return __ret;
15392 }
15393 #else
15394 __ai int32x2_t vneg_s32(int32x2_t __p0) {
15395  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15396  int32x2_t __ret;
15397  __ret = -__rev0;
15398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15399  return __ret;
15400 }
15401 #endif
15402 
15403 #ifdef __LITTLE_ENDIAN__
15404 __ai int16x4_t vneg_s16(int16x4_t __p0) {
15405  int16x4_t __ret;
15406  __ret = -__p0;
15407  return __ret;
15408 }
15409 #else
15410 __ai int16x4_t vneg_s16(int16x4_t __p0) {
15411  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15412  int16x4_t __ret;
15413  __ret = -__rev0;
15414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15415  return __ret;
15416 }
15417 #endif
15418 
15419 #ifdef __LITTLE_ENDIAN__
15420 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15421  uint8x16_t __ret;
15422  __ret = __p0 | ~__p1;
15423  return __ret;
15424 }
15425 #else
15426 __ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15427  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15428  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15429  uint8x16_t __ret;
15430  __ret = __rev0 | ~__rev1;
15431  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15432  return __ret;
15433 }
15434 #endif
15435 
15436 #ifdef __LITTLE_ENDIAN__
15437 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15438  uint32x4_t __ret;
15439  __ret = __p0 | ~__p1;
15440  return __ret;
15441 }
15442 #else
15443 __ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15444  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15445  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15446  uint32x4_t __ret;
15447  __ret = __rev0 | ~__rev1;
15448  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15449  return __ret;
15450 }
15451 #endif
15452 
15453 #ifdef __LITTLE_ENDIAN__
15454 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15455  uint64x2_t __ret;
15456  __ret = __p0 | ~__p1;
15457  return __ret;
15458 }
15459 #else
15460 __ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15461  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15462  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15463  uint64x2_t __ret;
15464  __ret = __rev0 | ~__rev1;
15465  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15466  return __ret;
15467 }
15468 #endif
15469 
15470 #ifdef __LITTLE_ENDIAN__
15471 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15472  uint16x8_t __ret;
15473  __ret = __p0 | ~__p1;
15474  return __ret;
15475 }
15476 #else
15477 __ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15478  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15479  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15480  uint16x8_t __ret;
15481  __ret = __rev0 | ~__rev1;
15482  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15483  return __ret;
15484 }
15485 #endif
15486 
15487 #ifdef __LITTLE_ENDIAN__
15488 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15489  int8x16_t __ret;
15490  __ret = __p0 | ~__p1;
15491  return __ret;
15492 }
15493 #else
15494 __ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15495  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15496  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15497  int8x16_t __ret;
15498  __ret = __rev0 | ~__rev1;
15499  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15500  return __ret;
15501 }
15502 #endif
15503 
15504 #ifdef __LITTLE_ENDIAN__
15505 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15506  int32x4_t __ret;
15507  __ret = __p0 | ~__p1;
15508  return __ret;
15509 }
15510 #else
15511 __ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15512  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15513  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15514  int32x4_t __ret;
15515  __ret = __rev0 | ~__rev1;
15516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15517  return __ret;
15518 }
15519 #endif
15520 
15521 #ifdef __LITTLE_ENDIAN__
15522 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15523  int64x2_t __ret;
15524  __ret = __p0 | ~__p1;
15525  return __ret;
15526 }
15527 #else
15528 __ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15529  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15530  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15531  int64x2_t __ret;
15532  __ret = __rev0 | ~__rev1;
15533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15534  return __ret;
15535 }
15536 #endif
15537 
15538 #ifdef __LITTLE_ENDIAN__
15539 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15540  int16x8_t __ret;
15541  __ret = __p0 | ~__p1;
15542  return __ret;
15543 }
15544 #else
15545 __ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15546  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15547  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15548  int16x8_t __ret;
15549  __ret = __rev0 | ~__rev1;
15550  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15551  return __ret;
15552 }
15553 #endif
15554 
15555 #ifdef __LITTLE_ENDIAN__
15556 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15557  uint8x8_t __ret;
15558  __ret = __p0 | ~__p1;
15559  return __ret;
15560 }
15561 #else
15562 __ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15563  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15564  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15565  uint8x8_t __ret;
15566  __ret = __rev0 | ~__rev1;
15567  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15568  return __ret;
15569 }
15570 #endif
15571 
15572 #ifdef __LITTLE_ENDIAN__
15573 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15574  uint32x2_t __ret;
15575  __ret = __p0 | ~__p1;
15576  return __ret;
15577 }
15578 #else
15579 __ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15580  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15581  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15582  uint32x2_t __ret;
15583  __ret = __rev0 | ~__rev1;
15584  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15585  return __ret;
15586 }
15587 #endif
15588 
15589 #ifdef __LITTLE_ENDIAN__
15590 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15591  uint64x1_t __ret;
15592  __ret = __p0 | ~__p1;
15593  return __ret;
15594 }
15595 #else
15596 __ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15597  uint64x1_t __ret;
15598  __ret = __p0 | ~__p1;
15599  return __ret;
15600 }
15601 #endif
15602 
15603 #ifdef __LITTLE_ENDIAN__
15604 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15605  uint16x4_t __ret;
15606  __ret = __p0 | ~__p1;
15607  return __ret;
15608 }
15609 #else
15610 __ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15611  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15612  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15613  uint16x4_t __ret;
15614  __ret = __rev0 | ~__rev1;
15615  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15616  return __ret;
15617 }
15618 #endif
15619 
15620 #ifdef __LITTLE_ENDIAN__
15621 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15622  int8x8_t __ret;
15623  __ret = __p0 | ~__p1;
15624  return __ret;
15625 }
15626 #else
15627 __ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15628  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15629  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15630  int8x8_t __ret;
15631  __ret = __rev0 | ~__rev1;
15632  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15633  return __ret;
15634 }
15635 #endif
15636 
15637 #ifdef __LITTLE_ENDIAN__
15638 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15639  int32x2_t __ret;
15640  __ret = __p0 | ~__p1;
15641  return __ret;
15642 }
15643 #else
15644 __ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15645  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15646  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15647  int32x2_t __ret;
15648  __ret = __rev0 | ~__rev1;
15649  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15650  return __ret;
15651 }
15652 #endif
15653 
15654 #ifdef __LITTLE_ENDIAN__
15655 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15656  int64x1_t __ret;
15657  __ret = __p0 | ~__p1;
15658  return __ret;
15659 }
15660 #else
15661 __ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15662  int64x1_t __ret;
15663  __ret = __p0 | ~__p1;
15664  return __ret;
15665 }
15666 #endif
15667 
15668 #ifdef __LITTLE_ENDIAN__
15669 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15670  int16x4_t __ret;
15671  __ret = __p0 | ~__p1;
15672  return __ret;
15673 }
15674 #else
15675 __ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15676  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15677  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15678  int16x4_t __ret;
15679  __ret = __rev0 | ~__rev1;
15680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15681  return __ret;
15682 }
15683 #endif
15684 
15685 #ifdef __LITTLE_ENDIAN__
15686 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15687  uint8x16_t __ret;
15688  __ret = __p0 | __p1;
15689  return __ret;
15690 }
15691 #else
15692 __ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15693  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15694  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15695  uint8x16_t __ret;
15696  __ret = __rev0 | __rev1;
15697  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15698  return __ret;
15699 }
15700 #endif
15701 
15702 #ifdef __LITTLE_ENDIAN__
15703 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15704  uint32x4_t __ret;
15705  __ret = __p0 | __p1;
15706  return __ret;
15707 }
15708 #else
15709 __ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15710  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15711  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15712  uint32x4_t __ret;
15713  __ret = __rev0 | __rev1;
15714  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15715  return __ret;
15716 }
15717 #endif
15718 
15719 #ifdef __LITTLE_ENDIAN__
15720 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15721  uint64x2_t __ret;
15722  __ret = __p0 | __p1;
15723  return __ret;
15724 }
15725 #else
15726 __ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15727  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15728  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15729  uint64x2_t __ret;
15730  __ret = __rev0 | __rev1;
15731  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15732  return __ret;
15733 }
15734 #endif
15735 
15736 #ifdef __LITTLE_ENDIAN__
15737 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15738  uint16x8_t __ret;
15739  __ret = __p0 | __p1;
15740  return __ret;
15741 }
15742 #else
15743 __ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15744  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15745  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15746  uint16x8_t __ret;
15747  __ret = __rev0 | __rev1;
15748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15749  return __ret;
15750 }
15751 #endif
15752 
15753 #ifdef __LITTLE_ENDIAN__
15754 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15755  int8x16_t __ret;
15756  __ret = __p0 | __p1;
15757  return __ret;
15758 }
15759 #else
15760 __ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15761  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15762  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15763  int8x16_t __ret;
15764  __ret = __rev0 | __rev1;
15765  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15766  return __ret;
15767 }
15768 #endif
15769 
15770 #ifdef __LITTLE_ENDIAN__
15771 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15772  int32x4_t __ret;
15773  __ret = __p0 | __p1;
15774  return __ret;
15775 }
15776 #else
15777 __ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15778  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15779  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15780  int32x4_t __ret;
15781  __ret = __rev0 | __rev1;
15782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15783  return __ret;
15784 }
15785 #endif
15786 
15787 #ifdef __LITTLE_ENDIAN__
15788 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15789  int64x2_t __ret;
15790  __ret = __p0 | __p1;
15791  return __ret;
15792 }
15793 #else
15794 __ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15795  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15796  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15797  int64x2_t __ret;
15798  __ret = __rev0 | __rev1;
15799  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15800  return __ret;
15801 }
15802 #endif
15803 
15804 #ifdef __LITTLE_ENDIAN__
15805 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15806  int16x8_t __ret;
15807  __ret = __p0 | __p1;
15808  return __ret;
15809 }
15810 #else
15811 __ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15812  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15813  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15814  int16x8_t __ret;
15815  __ret = __rev0 | __rev1;
15816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15817  return __ret;
15818 }
15819 #endif
15820 
15821 #ifdef __LITTLE_ENDIAN__
15822 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15823  uint8x8_t __ret;
15824  __ret = __p0 | __p1;
15825  return __ret;
15826 }
15827 #else
15828 __ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15829  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15830  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15831  uint8x8_t __ret;
15832  __ret = __rev0 | __rev1;
15833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15834  return __ret;
15835 }
15836 #endif
15837 
15838 #ifdef __LITTLE_ENDIAN__
15839 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15840  uint32x2_t __ret;
15841  __ret = __p0 | __p1;
15842  return __ret;
15843 }
15844 #else
15845 __ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15846  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15847  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15848  uint32x2_t __ret;
15849  __ret = __rev0 | __rev1;
15850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15851  return __ret;
15852 }
15853 #endif
15854 
15855 #ifdef __LITTLE_ENDIAN__
15856 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15857  uint64x1_t __ret;
15858  __ret = __p0 | __p1;
15859  return __ret;
15860 }
15861 #else
15862 __ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15863  uint64x1_t __ret;
15864  __ret = __p0 | __p1;
15865  return __ret;
15866 }
15867 #endif
15868 
15869 #ifdef __LITTLE_ENDIAN__
15870 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15871  uint16x4_t __ret;
15872  __ret = __p0 | __p1;
15873  return __ret;
15874 }
15875 #else
15876 __ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15877  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15878  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15879  uint16x4_t __ret;
15880  __ret = __rev0 | __rev1;
15881  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15882  return __ret;
15883 }
15884 #endif
15885 
15886 #ifdef __LITTLE_ENDIAN__
15887 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15888  int8x8_t __ret;
15889  __ret = __p0 | __p1;
15890  return __ret;
15891 }
15892 #else
15893 __ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15894  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15895  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15896  int8x8_t __ret;
15897  __ret = __rev0 | __rev1;
15898  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15899  return __ret;
15900 }
15901 #endif
15902 
15903 #ifdef __LITTLE_ENDIAN__
15904 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15905  int32x2_t __ret;
15906  __ret = __p0 | __p1;
15907  return __ret;
15908 }
15909 #else
15910 __ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15911  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15912  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15913  int32x2_t __ret;
15914  __ret = __rev0 | __rev1;
15915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15916  return __ret;
15917 }
15918 #endif
15919 
15920 #ifdef __LITTLE_ENDIAN__
15921 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15922  int64x1_t __ret;
15923  __ret = __p0 | __p1;
15924  return __ret;
15925 }
15926 #else
15927 __ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15928  int64x1_t __ret;
15929  __ret = __p0 | __p1;
15930  return __ret;
15931 }
15932 #endif
15933 
15934 #ifdef __LITTLE_ENDIAN__
15935 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15936  int16x4_t __ret;
15937  __ret = __p0 | __p1;
15938  return __ret;
15939 }
15940 #else
15941 __ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15942  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15943  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15944  int16x4_t __ret;
15945  __ret = __rev0 | __rev1;
15946  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15947  return __ret;
15948 }
15949 #endif
15950 
15951 #ifdef __LITTLE_ENDIAN__
15952 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15953  uint16x8_t __ret;
15954  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
15955  return __ret;
15956 }
15957 #else
15958 __ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15959  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15960  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15961  uint16x8_t __ret;
15962  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
15963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15964  return __ret;
15965 }
15966 #endif
15967 
15968 #ifdef __LITTLE_ENDIAN__
15969 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15970  uint64x2_t __ret;
15971  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
15972  return __ret;
15973 }
15974 #else
15975 __ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15976  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15977  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15978  uint64x2_t __ret;
15979  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
15980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15981  return __ret;
15982 }
15983 #endif
15984 
15985 #ifdef __LITTLE_ENDIAN__
15986 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15987  uint32x4_t __ret;
15988  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
15989  return __ret;
15990 }
15991 #else
15992 __ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15993  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15994  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15995  uint32x4_t __ret;
15996  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
15997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15998  return __ret;
15999 }
16000 #endif
16001 
16002 #ifdef __LITTLE_ENDIAN__
16003 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16004  int16x8_t __ret;
16005  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16006  return __ret;
16007 }
16008 #else
16009 __ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16010  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16011  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16012  int16x8_t __ret;
16013  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16014  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16015  return __ret;
16016 }
16017 #endif
16018 
16019 #ifdef __LITTLE_ENDIAN__
16020 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16021  int64x2_t __ret;
16022  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16023  return __ret;
16024 }
16025 #else
16026 __ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16027  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16028  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16029  int64x2_t __ret;
16030  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16032  return __ret;
16033 }
16034 #endif
16035 
16036 #ifdef __LITTLE_ENDIAN__
16037 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16038  int32x4_t __ret;
16039  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16040  return __ret;
16041 }
16042 #else
16043 __ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16044  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16045  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16046  int32x4_t __ret;
16047  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16049  return __ret;
16050 }
16051 #endif
16052 
16053 #ifdef __LITTLE_ENDIAN__
16054 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16055  uint16x4_t __ret;
16056  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16057  return __ret;
16058 }
16059 #else
16060 __ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16061  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16062  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16063  uint16x4_t __ret;
16064  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16065  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16066  return __ret;
16067 }
16068 #endif
16069 
16070 #ifdef __LITTLE_ENDIAN__
16071 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16072  uint64x1_t __ret;
16073  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16074  return __ret;
16075 }
16076 #else
16077 __ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16078  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16079  uint64x1_t __ret;
16080  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
16081  return __ret;
16082 }
16083 #endif
16084 
16085 #ifdef __LITTLE_ENDIAN__
16086 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16087  uint32x2_t __ret;
16088  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16089  return __ret;
16090 }
16091 #else
16092 __ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16093  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16094  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16095  uint32x2_t __ret;
16096  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16097  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16098  return __ret;
16099 }
16100 #endif
16101 
16102 #ifdef __LITTLE_ENDIAN__
16103 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16104  int16x4_t __ret;
16105  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16106  return __ret;
16107 }
16108 #else
16109 __ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16110  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16111  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16112  int16x4_t __ret;
16113  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16114  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16115  return __ret;
16116 }
16117 #endif
16118 
16119 #ifdef __LITTLE_ENDIAN__
16120 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16121  int64x1_t __ret;
16122  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16123  return __ret;
16124 }
16125 #else
16126 __ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16127  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16128  int64x1_t __ret;
16129  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
16130  return __ret;
16131 }
16132 #endif
16133 
16134 #ifdef __LITTLE_ENDIAN__
16135 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16136  int32x2_t __ret;
16137  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16138  return __ret;
16139 }
16140 #else
16141 __ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16142  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16143  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16144  int32x2_t __ret;
16145  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16146  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16147  return __ret;
16148 }
16149 #endif
16150 
16151 #ifdef __LITTLE_ENDIAN__
16152 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16153  uint8x8_t __ret;
16154  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16155  return __ret;
16156 }
16157 #else
16158 __ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16159  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16160  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16161  uint8x8_t __ret;
16162  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16163  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16164  return __ret;
16165 }
16166 #endif
16167 
16168 #ifdef __LITTLE_ENDIAN__
16169 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16170  uint32x2_t __ret;
16171  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16172  return __ret;
16173 }
16174 #else
16175 __ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16176  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16177  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16178  uint32x2_t __ret;
16179  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16180  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16181  return __ret;
16182 }
16183 #endif
16184 
16185 #ifdef __LITTLE_ENDIAN__
16186 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16187  uint16x4_t __ret;
16188  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16189  return __ret;
16190 }
16191 #else
16192 __ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16193  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16194  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16195  uint16x4_t __ret;
16196  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16197  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16198  return __ret;
16199 }
16200 #endif
16201 
16202 #ifdef __LITTLE_ENDIAN__
16203 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16204  int8x8_t __ret;
16205  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16206  return __ret;
16207 }
16208 #else
16209 __ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16210  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16211  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16212  int8x8_t __ret;
16213  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16215  return __ret;
16216 }
16217 #endif
16218 
16219 #ifdef __LITTLE_ENDIAN__
16220 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16221  float32x2_t __ret;
16222  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16223  return __ret;
16224 }
16225 #else
16226 __ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16227  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16228  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16229  float32x2_t __ret;
16230  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16232  return __ret;
16233 }
16234 #endif
16235 
16236 #ifdef __LITTLE_ENDIAN__
16237 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16238  int32x2_t __ret;
16239  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16240  return __ret;
16241 }
16242 #else
16243 __ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16244  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16245  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16246  int32x2_t __ret;
16247  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16248  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16249  return __ret;
16250 }
16251 #endif
16252 
16253 #ifdef __LITTLE_ENDIAN__
16254 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16255  int16x4_t __ret;
16256  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16257  return __ret;
16258 }
16259 #else
16260 __ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16261  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16262  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16263  int16x4_t __ret;
16264  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16265  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16266  return __ret;
16267 }
16268 #endif
16269 
16270 #ifdef __LITTLE_ENDIAN__
16271 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16272  uint16x8_t __ret;
16273  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
16274  return __ret;
16275 }
16276 #else
16277 __ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16278  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16279  uint16x8_t __ret;
16280  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
16281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16282  return __ret;
16283 }
16284 #endif
16285 
16286 #ifdef __LITTLE_ENDIAN__
16287 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16288  uint64x2_t __ret;
16289  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
16290  return __ret;
16291 }
16292 #else
16293 __ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16294  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16295  uint64x2_t __ret;
16296  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
16297  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16298  return __ret;
16299 }
16300 #endif
16301 
16302 #ifdef __LITTLE_ENDIAN__
16303 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16304  uint32x4_t __ret;
16305  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
16306  return __ret;
16307 }
16308 #else
16309 __ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16310  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16311  uint32x4_t __ret;
16312  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
16313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16314  return __ret;
16315 }
16316 #endif
16317 
16318 #ifdef __LITTLE_ENDIAN__
16319 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16320  int16x8_t __ret;
16321  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
16322  return __ret;
16323 }
16324 #else
16325 __ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16326  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16327  int16x8_t __ret;
16328  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
16329  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16330  return __ret;
16331 }
16332 #endif
16333 
16334 #ifdef __LITTLE_ENDIAN__
16335 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16336  int64x2_t __ret;
16337  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
16338  return __ret;
16339 }
16340 #else
16341 __ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16342  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16343  int64x2_t __ret;
16344  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
16345  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16346  return __ret;
16347 }
16348 #endif
16349 
16350 #ifdef __LITTLE_ENDIAN__
16351 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16352  int32x4_t __ret;
16353  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
16354  return __ret;
16355 }
16356 #else
16357 __ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16358  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16359  int32x4_t __ret;
16360  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
16361  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16362  return __ret;
16363 }
16364 #endif
16365 
16366 #ifdef __LITTLE_ENDIAN__
16367 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16368  uint16x4_t __ret;
16369  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
16370  return __ret;
16371 }
16372 #else
16373 __ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16374  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16375  uint16x4_t __ret;
16376  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
16377  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16378  return __ret;
16379 }
16380 #endif
16381 
16382 #ifdef __LITTLE_ENDIAN__
16383 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16384  uint64x1_t __ret;
16385  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
16386  return __ret;
16387 }
16388 #else
16389 __ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16390  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16391  uint64x1_t __ret;
16392  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
16393  return __ret;
16394 }
16395 #endif
16396 
16397 #ifdef __LITTLE_ENDIAN__
16398 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16399  uint32x2_t __ret;
16400  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
16401  return __ret;
16402 }
16403 #else
16404 __ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16405  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16406  uint32x2_t __ret;
16407  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
16408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16409  return __ret;
16410 }
16411 #endif
16412 
16413 #ifdef __LITTLE_ENDIAN__
16414 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16415  int16x4_t __ret;
16416  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
16417  return __ret;
16418 }
16419 #else
16420 __ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16421  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16422  int16x4_t __ret;
16423  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
16424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16425  return __ret;
16426 }
16427 #endif
16428 
16429 #ifdef __LITTLE_ENDIAN__
16430 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16431  int64x1_t __ret;
16432  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
16433  return __ret;
16434 }
16435 #else
16436 __ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16437  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16438  int64x1_t __ret;
16439  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
16440  return __ret;
16441 }
16442 #endif
16443 
16444 #ifdef __LITTLE_ENDIAN__
16445 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16446  int32x2_t __ret;
16447  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
16448  return __ret;
16449 }
16450 #else
16451 __ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16452  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16453  int32x2_t __ret;
16454  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
16455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16456  return __ret;
16457 }
16458 #endif
16459 
16460 #ifdef __LITTLE_ENDIAN__
16461 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16462  uint8x8_t __ret;
16463  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16464  return __ret;
16465 }
16466 #else
16467 __ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16468  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16469  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16470  uint8x8_t __ret;
16471  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16472  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16473  return __ret;
16474 }
16475 #endif
16476 
16477 #ifdef __LITTLE_ENDIAN__
16478 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16479  uint32x2_t __ret;
16480  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16481  return __ret;
16482 }
16483 #else
16484 __ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16485  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16486  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16487  uint32x2_t __ret;
16488  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16490  return __ret;
16491 }
16492 #endif
16493 
16494 #ifdef __LITTLE_ENDIAN__
16495 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16496  uint16x4_t __ret;
16497  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16498  return __ret;
16499 }
16500 #else
16501 __ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16502  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16503  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16504  uint16x4_t __ret;
16505  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16506  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16507  return __ret;
16508 }
16509 #endif
16510 
16511 #ifdef __LITTLE_ENDIAN__
16512 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16513  int8x8_t __ret;
16514  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16515  return __ret;
16516 }
16517 #else
16518 __ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16519  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16520  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16521  int8x8_t __ret;
16522  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16523  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16524  return __ret;
16525 }
16526 #endif
16527 
16528 #ifdef __LITTLE_ENDIAN__
16529 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16530  float32x2_t __ret;
16531  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16532  return __ret;
16533 }
16534 #else
16535 __ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16536  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16537  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16538  float32x2_t __ret;
16539  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16540  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16541  return __ret;
16542 }
16543 #endif
16544 
16545 #ifdef __LITTLE_ENDIAN__
16546 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16547  int32x2_t __ret;
16548  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16549  return __ret;
16550 }
16551 #else
16552 __ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16553  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16554  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16555  int32x2_t __ret;
16556  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16558  return __ret;
16559 }
16560 #endif
16561 
16562 #ifdef __LITTLE_ENDIAN__
16563 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16564  int16x4_t __ret;
16565  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16566  return __ret;
16567 }
16568 #else
16569 __ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16570  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16571  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16572  int16x4_t __ret;
16573  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16575  return __ret;
16576 }
16577 #endif
16578 
16579 #ifdef __LITTLE_ENDIAN__
16580 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16581  uint8x8_t __ret;
16582  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16583  return __ret;
16584 }
16585 #else
16586 __ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16587  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16588  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16589  uint8x8_t __ret;
16590  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16591  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16592  return __ret;
16593 }
16594 #endif
16595 
16596 #ifdef __LITTLE_ENDIAN__
16597 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16598  uint32x2_t __ret;
16599  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16600  return __ret;
16601 }
16602 #else
16603 __ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16604  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16605  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16606  uint32x2_t __ret;
16607  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16608  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16609  return __ret;
16610 }
16611 #endif
16612 
16613 #ifdef __LITTLE_ENDIAN__
16614 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16615  uint16x4_t __ret;
16616  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16617  return __ret;
16618 }
16619 #else
16620 __ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16621  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16622  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16623  uint16x4_t __ret;
16624  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16625  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16626  return __ret;
16627 }
16628 #endif
16629 
16630 #ifdef __LITTLE_ENDIAN__
16631 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16632  int8x8_t __ret;
16633  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16634  return __ret;
16635 }
16636 #else
16637 __ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16638  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16639  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16640  int8x8_t __ret;
16641  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16642  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16643  return __ret;
16644 }
16645 #endif
16646 
16647 #ifdef __LITTLE_ENDIAN__
16648 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16649  float32x2_t __ret;
16650  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16651  return __ret;
16652 }
16653 #else
16654 __ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16655  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16656  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16657  float32x2_t __ret;
16658  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16659  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16660  return __ret;
16661 }
16662 #endif
16663 
16664 #ifdef __LITTLE_ENDIAN__
16665 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16666  int32x2_t __ret;
16667  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16668  return __ret;
16669 }
16670 #else
16671 __ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16672  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16673  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16674  int32x2_t __ret;
16675  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16677  return __ret;
16678 }
16679 #endif
16680 
16681 #ifdef __LITTLE_ENDIAN__
16682 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16683  int16x4_t __ret;
16684  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16685  return __ret;
16686 }
16687 #else
16688 __ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16689  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16690  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16691  int16x4_t __ret;
16692  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16694  return __ret;
16695 }
16696 #endif
16697 
16698 #ifdef __LITTLE_ENDIAN__
16699 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16700  int8x16_t __ret;
16701  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
16702  return __ret;
16703 }
16704 #else
16705 __ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16706  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16707  int8x16_t __ret;
16708  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
16709  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16710  return __ret;
16711 }
16712 #endif
16713 
16714 #ifdef __LITTLE_ENDIAN__
16715 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16716  int32x4_t __ret;
16717  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
16718  return __ret;
16719 }
16720 #else
16721 __ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16722  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16723  int32x4_t __ret;
16724  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
16725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16726  return __ret;
16727 }
16728 #endif
16729 
16730 #ifdef __LITTLE_ENDIAN__
16731 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16732  int16x8_t __ret;
16733  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
16734  return __ret;
16735 }
16736 #else
16737 __ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16738  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16739  int16x8_t __ret;
16740  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
16741  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16742  return __ret;
16743 }
16744 #endif
16745 
16746 #ifdef __LITTLE_ENDIAN__
16747 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
16748  int8x8_t __ret;
16749  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
16750  return __ret;
16751 }
16752 #else
16753 __ai int8x8_t vqabs_s8(int8x8_t __p0) {
16754  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16755  int8x8_t __ret;
16756  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
16757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16758  return __ret;
16759 }
16760 #endif
16761 
16762 #ifdef __LITTLE_ENDIAN__
16763 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
16764  int32x2_t __ret;
16765  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
16766  return __ret;
16767 }
16768 #else
16769 __ai int32x2_t vqabs_s32(int32x2_t __p0) {
16770  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16771  int32x2_t __ret;
16772  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
16773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16774  return __ret;
16775 }
16776 #endif
16777 
16778 #ifdef __LITTLE_ENDIAN__
16779 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
16780  int16x4_t __ret;
16781  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
16782  return __ret;
16783 }
16784 #else
16785 __ai int16x4_t vqabs_s16(int16x4_t __p0) {
16786  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16787  int16x4_t __ret;
16788  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
16789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16790  return __ret;
16791 }
16792 #endif
16793 
16794 #ifdef __LITTLE_ENDIAN__
16795 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16796  uint8x16_t __ret;
16797  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
16798  return __ret;
16799 }
16800 #else
16801 __ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16802  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16803  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16804  uint8x16_t __ret;
16805  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
16806  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16807  return __ret;
16808 }
16809 #endif
16810 
16811 #ifdef __LITTLE_ENDIAN__
16812 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16813  uint32x4_t __ret;
16814  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
16815  return __ret;
16816 }
16817 #else
16818 __ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16819  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16820  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16821  uint32x4_t __ret;
16822  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
16823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16824  return __ret;
16825 }
16826 #endif
16827 
16828 #ifdef __LITTLE_ENDIAN__
16829 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16830  uint64x2_t __ret;
16831  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
16832  return __ret;
16833 }
16834 #else
16835 __ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16836  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16837  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16838  uint64x2_t __ret;
16839  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
16840  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16841  return __ret;
16842 }
16843 #endif
16844 
16845 #ifdef __LITTLE_ENDIAN__
16846 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16847  uint16x8_t __ret;
16848  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
16849  return __ret;
16850 }
16851 #else
16852 __ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16853  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16854  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16855  uint16x8_t __ret;
16856  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
16857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16858  return __ret;
16859 }
16860 #endif
16861 
16862 #ifdef __LITTLE_ENDIAN__
16863 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16864  int8x16_t __ret;
16865  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
16866  return __ret;
16867 }
16868 #else
16869 __ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16870  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16871  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16872  int8x16_t __ret;
16873  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
16874  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16875  return __ret;
16876 }
16877 #endif
16878 
16879 #ifdef __LITTLE_ENDIAN__
16880 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16881  int32x4_t __ret;
16882  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16883  return __ret;
16884 }
16885 #else
16886 __ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16887  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16888  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16889  int32x4_t __ret;
16890  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16891  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16892  return __ret;
16893 }
16894 __ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16895  int32x4_t __ret;
16896  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16897  return __ret;
16898 }
16899 #endif
16900 
16901 #ifdef __LITTLE_ENDIAN__
16902 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16903  int64x2_t __ret;
16904  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16905  return __ret;
16906 }
16907 #else
16908 __ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16909  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16910  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16911  int64x2_t __ret;
16912  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16914  return __ret;
16915 }
16916 #endif
16917 
16918 #ifdef __LITTLE_ENDIAN__
16919 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16920  int16x8_t __ret;
16921  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16922  return __ret;
16923 }
16924 #else
16925 __ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16926  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16927  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16928  int16x8_t __ret;
16929  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16930  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16931  return __ret;
16932 }
16933 __ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16934  int16x8_t __ret;
16935  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16936  return __ret;
16937 }
16938 #endif
16939 
16940 #ifdef __LITTLE_ENDIAN__
16941 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16942  uint8x8_t __ret;
16943  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16944  return __ret;
16945 }
16946 #else
16947 __ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16948  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16949  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16950  uint8x8_t __ret;
16951  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16952  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16953  return __ret;
16954 }
16955 #endif
16956 
16957 #ifdef __LITTLE_ENDIAN__
16958 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16959  uint32x2_t __ret;
16960  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16961  return __ret;
16962 }
16963 #else
16964 __ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16965  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16966  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16967  uint32x2_t __ret;
16968  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16970  return __ret;
16971 }
16972 #endif
16973 
16974 #ifdef __LITTLE_ENDIAN__
16975 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16976  uint64x1_t __ret;
16977  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16978  return __ret;
16979 }
16980 #else
16981 __ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16982  uint64x1_t __ret;
16983  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16984  return __ret;
16985 }
16986 #endif
16987 
16988 #ifdef __LITTLE_ENDIAN__
16989 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16990  uint16x4_t __ret;
16991  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16992  return __ret;
16993 }
16994 #else
16995 __ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16996  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16997  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16998  uint16x4_t __ret;
16999  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
17000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17001  return __ret;
17002 }
17003 #endif
17004 
17005 #ifdef __LITTLE_ENDIAN__
17006 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
17007  int8x8_t __ret;
17008  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
17009  return __ret;
17010 }
17011 #else
17012 __ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
17013  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17014  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17015  int8x8_t __ret;
17016  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
17017  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17018  return __ret;
17019 }
17020 #endif
17021 
17022 #ifdef __LITTLE_ENDIAN__
17023 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17024  int32x2_t __ret;
17025  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17026  return __ret;
17027 }
17028 #else
17029 __ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17030  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17031  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17032  int32x2_t __ret;
17033  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17035  return __ret;
17036 }
17037 __ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17038  int32x2_t __ret;
17039  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17040  return __ret;
17041 }
17042 #endif
17043 
17044 #ifdef __LITTLE_ENDIAN__
17045 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17046  int64x1_t __ret;
17047  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17048  return __ret;
17049 }
17050 #else
17051 __ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17052  int64x1_t __ret;
17053  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17054  return __ret;
17055 }
17056 #endif
17057 
17058 #ifdef __LITTLE_ENDIAN__
17059 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17060  int16x4_t __ret;
17061  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17062  return __ret;
17063 }
17064 #else
17065 __ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17066  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17067  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17068  int16x4_t __ret;
17069  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17070  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17071  return __ret;
17072 }
17073 __ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17074  int16x4_t __ret;
17075  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17076  return __ret;
17077 }
17078 #endif
17079 
17080 #ifdef __LITTLE_ENDIAN__
17081 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17082  int64x2_t __ret;
17083  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17084  return __ret;
17085 }
17086 #else
17087 __ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17088  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17089  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17090  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17091  int64x2_t __ret;
17092  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17094  return __ret;
17095 }
17096 __ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17097  int64x2_t __ret;
17098  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17099  return __ret;
17100 }
17101 #endif
17102 
17103 #ifdef __LITTLE_ENDIAN__
17104 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17105  int32x4_t __ret;
17106  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17107  return __ret;
17108 }
17109 #else
17110 __ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17111  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17112  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17113  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17114  int32x4_t __ret;
17115  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17116  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17117  return __ret;
17118 }
17119 __ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17120  int32x4_t __ret;
17121  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17122  return __ret;
17123 }
17124 #endif
17125 
17126 #ifdef __LITTLE_ENDIAN__
17127 #define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17128  int64x2_t __s0 = __p0; \
17129  int32x2_t __s1 = __p1; \
17130  int32x2_t __s2 = __p2; \
17131  int64x2_t __ret; \
17132  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17133  __ret; \
17134 })
17135 #else
17136 #define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17137  int64x2_t __s0 = __p0; \
17138  int32x2_t __s1 = __p1; \
17139  int32x2_t __s2 = __p2; \
17140  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17141  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17142  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17143  int64x2_t __ret; \
17144  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17145  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17146  __ret; \
17147 })
17148 #endif
17149 
17150 #ifdef __LITTLE_ENDIAN__
17151 #define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17152  int32x4_t __s0 = __p0; \
17153  int16x4_t __s1 = __p1; \
17154  int16x4_t __s2 = __p2; \
17155  int32x4_t __ret; \
17156  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17157  __ret; \
17158 })
17159 #else
17160 #define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17161  int32x4_t __s0 = __p0; \
17162  int16x4_t __s1 = __p1; \
17163  int16x4_t __s2 = __p2; \
17164  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17165  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17166  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17167  int32x4_t __ret; \
17168  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17169  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17170  __ret; \
17171 })
17172 #endif
17173 
17174 #ifdef __LITTLE_ENDIAN__
17175 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17176  int64x2_t __ret;
17177  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17178  return __ret;
17179 }
17180 #else
17181 __ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17182  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17183  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17184  int64x2_t __ret;
17185  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17186  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17187  return __ret;
17188 }
17189 __ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17190  int64x2_t __ret;
17191  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17192  return __ret;
17193 }
17194 #endif
17195 
17196 #ifdef __LITTLE_ENDIAN__
17197 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17198  int32x4_t __ret;
17199  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17200  return __ret;
17201 }
17202 #else
17203 __ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17204  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17205  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17206  int32x4_t __ret;
17207  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17208  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17209  return __ret;
17210 }
17211 __ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17212  int32x4_t __ret;
17213  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17214  return __ret;
17215 }
17216 #endif
17217 
17218 #ifdef __LITTLE_ENDIAN__
17219 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17220  int64x2_t __ret;
17221  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17222  return __ret;
17223 }
17224 #else
17225 __ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17226  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17227  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17228  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17229  int64x2_t __ret;
17230  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17232  return __ret;
17233 }
17234 __ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17235  int64x2_t __ret;
17236  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17237  return __ret;
17238 }
17239 #endif
17240 
17241 #ifdef __LITTLE_ENDIAN__
17242 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17243  int32x4_t __ret;
17244  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17245  return __ret;
17246 }
17247 #else
17248 __ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17249  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17250  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17251  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17252  int32x4_t __ret;
17253  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17254  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17255  return __ret;
17256 }
17257 __ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17258  int32x4_t __ret;
17259  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17260  return __ret;
17261 }
17262 #endif
17263 
17264 #ifdef __LITTLE_ENDIAN__
17265 #define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17266  int64x2_t __s0 = __p0; \
17267  int32x2_t __s1 = __p1; \
17268  int32x2_t __s2 = __p2; \
17269  int64x2_t __ret; \
17270  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17271  __ret; \
17272 })
17273 #else
17274 #define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17275  int64x2_t __s0 = __p0; \
17276  int32x2_t __s1 = __p1; \
17277  int32x2_t __s2 = __p2; \
17278  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17279  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17280  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17281  int64x2_t __ret; \
17282  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17284  __ret; \
17285 })
17286 #endif
17287 
17288 #ifdef __LITTLE_ENDIAN__
17289 #define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17290  int32x4_t __s0 = __p0; \
17291  int16x4_t __s1 = __p1; \
17292  int16x4_t __s2 = __p2; \
17293  int32x4_t __ret; \
17294  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17295  __ret; \
17296 })
17297 #else
17298 #define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17299  int32x4_t __s0 = __p0; \
17300  int16x4_t __s1 = __p1; \
17301  int16x4_t __s2 = __p2; \
17302  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17303  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17304  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17305  int32x4_t __ret; \
17306  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17307  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17308  __ret; \
17309 })
17310 #endif
17311 
17312 #ifdef __LITTLE_ENDIAN__
17313 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17314  int64x2_t __ret;
17315  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17316  return __ret;
17317 }
17318 #else
17319 __ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17320  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17321  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17322  int64x2_t __ret;
17323  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17324  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17325  return __ret;
17326 }
17327 __ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17328  int64x2_t __ret;
17329  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17330  return __ret;
17331 }
17332 #endif
17333 
17334 #ifdef __LITTLE_ENDIAN__
17335 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17336  int32x4_t __ret;
17337  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17338  return __ret;
17339 }
17340 #else
17341 __ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17342  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17343  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17344  int32x4_t __ret;
17345  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17346  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17347  return __ret;
17348 }
17349 __ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17350  int32x4_t __ret;
17351  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17352  return __ret;
17353 }
17354 #endif
17355 
17356 #ifdef __LITTLE_ENDIAN__
17357 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17358  int32x4_t __ret;
17359  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17360  return __ret;
17361 }
17362 #else
17363 __ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17364  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17365  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17366  int32x4_t __ret;
17367  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17368  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17369  return __ret;
17370 }
17371 __ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17372  int32x4_t __ret;
17373  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17374  return __ret;
17375 }
17376 #endif
17377 
17378 #ifdef __LITTLE_ENDIAN__
17379 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17380  int16x8_t __ret;
17381  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17382  return __ret;
17383 }
17384 #else
17385 __ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17386  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17387  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17388  int16x8_t __ret;
17389  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17390  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17391  return __ret;
17392 }
17393 __ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17394  int16x8_t __ret;
17395  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17396  return __ret;
17397 }
17398 #endif
17399 
17400 #ifdef __LITTLE_ENDIAN__
17401 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17402  int32x2_t __ret;
17403  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17404  return __ret;
17405 }
17406 #else
17407 __ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17408  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17409  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17410  int32x2_t __ret;
17411  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17412  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17413  return __ret;
17414 }
17415 __ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17416  int32x2_t __ret;
17417  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17418  return __ret;
17419 }
17420 #endif
17421 
17422 #ifdef __LITTLE_ENDIAN__
17423 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17424  int16x4_t __ret;
17425  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17426  return __ret;
17427 }
17428 #else
17429 __ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17430  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17431  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17432  int16x4_t __ret;
17433  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17434  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17435  return __ret;
17436 }
17437 __ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17438  int16x4_t __ret;
17439  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17440  return __ret;
17441 }
17442 #endif
17443 
17444 #ifdef __LITTLE_ENDIAN__
17445 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17446  int32x4_t __s0 = __p0; \
17447  int32x2_t __s1 = __p1; \
17448  int32x4_t __ret; \
17449  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17450  __ret; \
17451 })
17452 #else
17453 #define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17454  int32x4_t __s0 = __p0; \
17455  int32x2_t __s1 = __p1; \
17456  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17457  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17458  int32x4_t __ret; \
17459  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17461  __ret; \
17462 })
17463 #endif
17464 
17465 #ifdef __LITTLE_ENDIAN__
17466 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17467  int16x8_t __s0 = __p0; \
17468  int16x4_t __s1 = __p1; \
17469  int16x8_t __ret; \
17470  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17471  __ret; \
17472 })
17473 #else
17474 #define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17475  int16x8_t __s0 = __p0; \
17476  int16x4_t __s1 = __p1; \
17477  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
17478  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17479  int16x8_t __ret; \
17480  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
17482  __ret; \
17483 })
17484 #endif
17485 
17486 #ifdef __LITTLE_ENDIAN__
17487 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17488  int32x2_t __s0 = __p0; \
17489  int32x2_t __s1 = __p1; \
17490  int32x2_t __ret; \
17491  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17492  __ret; \
17493 })
17494 #else
17495 #define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17496  int32x2_t __s0 = __p0; \
17497  int32x2_t __s1 = __p1; \
17498  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17499  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17500  int32x2_t __ret; \
17501  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17502  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17503  __ret; \
17504 })
17505 #endif
17506 
17507 #ifdef __LITTLE_ENDIAN__
17508 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17509  int16x4_t __s0 = __p0; \
17510  int16x4_t __s1 = __p1; \
17511  int16x4_t __ret; \
17512  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17513  __ret; \
17514 })
17515 #else
17516 #define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17517  int16x4_t __s0 = __p0; \
17518  int16x4_t __s1 = __p1; \
17519  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17520  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17521  int16x4_t __ret; \
17522  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17523  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17524  __ret; \
17525 })
17526 #endif
17527 
17528 #ifdef __LITTLE_ENDIAN__
17529 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17530  int32x4_t __ret;
17531  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17532  return __ret;
17533 }
17534 #else
17535 __ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17536  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17537  int32x4_t __ret;
17538  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17539  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17540  return __ret;
17541 }
17542 #endif
17543 
17544 #ifdef __LITTLE_ENDIAN__
17545 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17546  int16x8_t __ret;
17547  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17548  return __ret;
17549 }
17550 #else
17551 __ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17552  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17553  int16x8_t __ret;
17554  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17555  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17556  return __ret;
17557 }
17558 #endif
17559 
17560 #ifdef __LITTLE_ENDIAN__
17561 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17562  int32x2_t __ret;
17563  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17564  return __ret;
17565 }
17566 #else
17567 __ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17568  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17569  int32x2_t __ret;
17570  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17572  return __ret;
17573 }
17574 #endif
17575 
17576 #ifdef __LITTLE_ENDIAN__
17577 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17578  int16x4_t __ret;
17579  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17580  return __ret;
17581 }
17582 #else
17583 __ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17584  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17585  int16x4_t __ret;
17586  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17587  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17588  return __ret;
17589 }
17590 #endif
17591 
17592 #ifdef __LITTLE_ENDIAN__
17593 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17594  int64x2_t __ret;
17595  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17596  return __ret;
17597 }
17598 #else
17599 __ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17600  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17601  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17602  int64x2_t __ret;
17603  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
17604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17605  return __ret;
17606 }
17607 __ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17608  int64x2_t __ret;
17609  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17610  return __ret;
17611 }
17612 #endif
17613 
17614 #ifdef __LITTLE_ENDIAN__
17615 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17616  int32x4_t __ret;
17617  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17618  return __ret;
17619 }
17620 #else
17621 __ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17622  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17623  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17624  int32x4_t __ret;
17625  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
17626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17627  return __ret;
17628 }
17629 __ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17630  int32x4_t __ret;
17631  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17632  return __ret;
17633 }
17634 #endif
17635 
17636 #ifdef __LITTLE_ENDIAN__
17637 #define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17638  int32x2_t __s0 = __p0; \
17639  int32x2_t __s1 = __p1; \
17640  int64x2_t __ret; \
17641  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17642  __ret; \
17643 })
17644 #else
17645 #define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17646  int32x2_t __s0 = __p0; \
17647  int32x2_t __s1 = __p1; \
17648  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17649  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17650  int64x2_t __ret; \
17651  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17652  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17653  __ret; \
17654 })
17655 #endif
17656 
17657 #ifdef __LITTLE_ENDIAN__
17658 #define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17659  int16x4_t __s0 = __p0; \
17660  int16x4_t __s1 = __p1; \
17661  int32x4_t __ret; \
17662  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17663  __ret; \
17664 })
17665 #else
17666 #define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17667  int16x4_t __s0 = __p0; \
17668  int16x4_t __s1 = __p1; \
17669  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17670  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17671  int32x4_t __ret; \
17672  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17673  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17674  __ret; \
17675 })
17676 #endif
17677 
17678 #ifdef __LITTLE_ENDIAN__
17679 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17680  int64x2_t __ret;
17681  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17682  return __ret;
17683 }
17684 #else
17685 __ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17686  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17687  int64x2_t __ret;
17688  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17690  return __ret;
17691 }
17692 __ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17693  int64x2_t __ret;
17694  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17695  return __ret;
17696 }
17697 #endif
17698 
17699 #ifdef __LITTLE_ENDIAN__
17700 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17701  int32x4_t __ret;
17702  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17703  return __ret;
17704 }
17705 #else
17706 __ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17707  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17708  int32x4_t __ret;
17709  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17710  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17711  return __ret;
17712 }
17713 __ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17714  int32x4_t __ret;
17715  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17716  return __ret;
17717 }
17718 #endif
17719 
17720 #ifdef __LITTLE_ENDIAN__
17721 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17722  uint16x4_t __ret;
17723  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17724  return __ret;
17725 }
17726 #else
17727 __ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17728  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17729  uint16x4_t __ret;
17730  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
17731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17732  return __ret;
17733 }
17734 __ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
17735  uint16x4_t __ret;
17736  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17737  return __ret;
17738 }
17739 #endif
17740 
17741 #ifdef __LITTLE_ENDIAN__
17742 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17743  uint32x2_t __ret;
17744  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17745  return __ret;
17746 }
17747 #else
17748 __ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17749  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17750  uint32x2_t __ret;
17751  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
17752  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17753  return __ret;
17754 }
17755 __ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
17756  uint32x2_t __ret;
17757  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17758  return __ret;
17759 }
17760 #endif
17761 
17762 #ifdef __LITTLE_ENDIAN__
17763 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17764  uint8x8_t __ret;
17765  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17766  return __ret;
17767 }
17768 #else
17769 __ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17770  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17771  uint8x8_t __ret;
17772  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
17773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17774  return __ret;
17775 }
17776 __ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
17777  uint8x8_t __ret;
17778  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17779  return __ret;
17780 }
17781 #endif
17782 
17783 #ifdef __LITTLE_ENDIAN__
17784 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17785  int16x4_t __ret;
17786  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17787  return __ret;
17788 }
17789 #else
17790 __ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17791  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17792  int16x4_t __ret;
17793  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
17794  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17795  return __ret;
17796 }
17797 __ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
17798  int16x4_t __ret;
17799  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17800  return __ret;
17801 }
17802 #endif
17803 
17804 #ifdef __LITTLE_ENDIAN__
17805 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17806  int32x2_t __ret;
17807  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17808  return __ret;
17809 }
17810 #else
17811 __ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17812  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17813  int32x2_t __ret;
17814  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
17815  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17816  return __ret;
17817 }
17818 __ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
17819  int32x2_t __ret;
17820  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17821  return __ret;
17822 }
17823 #endif
17824 
17825 #ifdef __LITTLE_ENDIAN__
17826 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17827  int8x8_t __ret;
17828  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17829  return __ret;
17830 }
17831 #else
17832 __ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17833  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17834  int8x8_t __ret;
17835  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
17836  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17837  return __ret;
17838 }
17839 __ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
17840  int8x8_t __ret;
17841  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17842  return __ret;
17843 }
17844 #endif
17845 
17846 #ifdef __LITTLE_ENDIAN__
17847 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17848  uint16x4_t __ret;
17849  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17850  return __ret;
17851 }
17852 #else
17853 __ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17854  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17855  uint16x4_t __ret;
17856  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
17857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17858  return __ret;
17859 }
17860 __ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
17861  uint16x4_t __ret;
17862  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17863  return __ret;
17864 }
17865 #endif
17866 
17867 #ifdef __LITTLE_ENDIAN__
17868 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17869  uint32x2_t __ret;
17870  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17871  return __ret;
17872 }
17873 #else
17874 __ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17875  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17876  uint32x2_t __ret;
17877  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
17878  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17879  return __ret;
17880 }
17881 __ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
17882  uint32x2_t __ret;
17883  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17884  return __ret;
17885 }
17886 #endif
17887 
17888 #ifdef __LITTLE_ENDIAN__
17889 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17890  uint8x8_t __ret;
17891  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17892  return __ret;
17893 }
17894 #else
17895 __ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17896  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17897  uint8x8_t __ret;
17898  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
17899  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17900  return __ret;
17901 }
17902 __ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
17903  uint8x8_t __ret;
17904  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17905  return __ret;
17906 }
17907 #endif
17908 
17909 #ifdef __LITTLE_ENDIAN__
17910 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17911  int8x16_t __ret;
17912  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
17913  return __ret;
17914 }
17915 #else
17916 __ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17917  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17918  int8x16_t __ret;
17919  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
17920  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17921  return __ret;
17922 }
17923 #endif
17924 
17925 #ifdef __LITTLE_ENDIAN__
17926 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17927  int32x4_t __ret;
17928  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
17929  return __ret;
17930 }
17931 #else
17932 __ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17933  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17934  int32x4_t __ret;
17935  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
17936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17937  return __ret;
17938 }
17939 #endif
17940 
17941 #ifdef __LITTLE_ENDIAN__
17942 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17943  int16x8_t __ret;
17944  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
17945  return __ret;
17946 }
17947 #else
17948 __ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17949  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17950  int16x8_t __ret;
17951  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
17952  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17953  return __ret;
17954 }
17955 #endif
17956 
17957 #ifdef __LITTLE_ENDIAN__
17958 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
17959  int8x8_t __ret;
17960  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
17961  return __ret;
17962 }
17963 #else
17964 __ai int8x8_t vqneg_s8(int8x8_t __p0) {
17965  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17966  int8x8_t __ret;
17967  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
17968  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17969  return __ret;
17970 }
17971 #endif
17972 
17973 #ifdef __LITTLE_ENDIAN__
17974 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
17975  int32x2_t __ret;
17976  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
17977  return __ret;
17978 }
17979 #else
17980 __ai int32x2_t vqneg_s32(int32x2_t __p0) {
17981  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17982  int32x2_t __ret;
17983  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
17984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17985  return __ret;
17986 }
17987 #endif
17988 
17989 #ifdef __LITTLE_ENDIAN__
17990 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
17991  int16x4_t __ret;
17992  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
17993  return __ret;
17994 }
17995 #else
17996 __ai int16x4_t vqneg_s16(int16x4_t __p0) {
17997  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17998  int16x4_t __ret;
17999  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
18000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18001  return __ret;
18002 }
18003 #endif
18004 
18005 #ifdef __LITTLE_ENDIAN__
18006 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18007  int32x4_t __ret;
18008  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18009  return __ret;
18010 }
18011 #else
18012 __ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18013  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18014  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18015  int32x4_t __ret;
18016  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18018  return __ret;
18019 }
18020 __ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18021  int32x4_t __ret;
18022  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18023  return __ret;
18024 }
18025 #endif
18026 
18027 #ifdef __LITTLE_ENDIAN__
18028 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18029  int16x8_t __ret;
18030  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18031  return __ret;
18032 }
18033 #else
18034 __ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18035  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18036  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18037  int16x8_t __ret;
18038  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18039  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18040  return __ret;
18041 }
18042 __ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18043  int16x8_t __ret;
18044  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18045  return __ret;
18046 }
18047 #endif
18048 
18049 #ifdef __LITTLE_ENDIAN__
18050 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18051  int32x2_t __ret;
18052  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18053  return __ret;
18054 }
18055 #else
18056 __ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18057  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18058  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18059  int32x2_t __ret;
18060  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18061  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18062  return __ret;
18063 }
18064 __ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18065  int32x2_t __ret;
18066  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18067  return __ret;
18068 }
18069 #endif
18070 
18071 #ifdef __LITTLE_ENDIAN__
18072 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18073  int16x4_t __ret;
18074  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18075  return __ret;
18076 }
18077 #else
18078 __ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18079  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18080  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18081  int16x4_t __ret;
18082  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18084  return __ret;
18085 }
18086 __ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18087  int16x4_t __ret;
18088  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18089  return __ret;
18090 }
18091 #endif
18092 
18093 #ifdef __LITTLE_ENDIAN__
18094 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18095  int32x4_t __s0 = __p0; \
18096  int32x2_t __s1 = __p1; \
18097  int32x4_t __ret; \
18098  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18099  __ret; \
18100 })
18101 #else
18102 #define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18103  int32x4_t __s0 = __p0; \
18104  int32x2_t __s1 = __p1; \
18105  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18106  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18107  int32x4_t __ret; \
18108  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18109  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18110  __ret; \
18111 })
18112 #endif
18113 
18114 #ifdef __LITTLE_ENDIAN__
18115 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18116  int16x8_t __s0 = __p0; \
18117  int16x4_t __s1 = __p1; \
18118  int16x8_t __ret; \
18119  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18120  __ret; \
18121 })
18122 #else
18123 #define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18124  int16x8_t __s0 = __p0; \
18125  int16x4_t __s1 = __p1; \
18126  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18127  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18128  int16x8_t __ret; \
18129  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18131  __ret; \
18132 })
18133 #endif
18134 
18135 #ifdef __LITTLE_ENDIAN__
18136 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18137  int32x2_t __s0 = __p0; \
18138  int32x2_t __s1 = __p1; \
18139  int32x2_t __ret; \
18140  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
18141  __ret; \
18142 })
18143 #else
18144 #define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18145  int32x2_t __s0 = __p0; \
18146  int32x2_t __s1 = __p1; \
18147  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18148  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18149  int32x2_t __ret; \
18150  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
18151  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18152  __ret; \
18153 })
18154 #endif
18155 
18156 #ifdef __LITTLE_ENDIAN__
18157 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18158  int16x4_t __s0 = __p0; \
18159  int16x4_t __s1 = __p1; \
18160  int16x4_t __ret; \
18161  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18162  __ret; \
18163 })
18164 #else
18165 #define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18166  int16x4_t __s0 = __p0; \
18167  int16x4_t __s1 = __p1; \
18168  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18169  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18170  int16x4_t __ret; \
18171  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18172  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18173  __ret; \
18174 })
18175 #endif
18176 
18177 #ifdef __LITTLE_ENDIAN__
18178 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18179  int32x4_t __ret;
18180  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18181  return __ret;
18182 }
18183 #else
18184 __ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18185  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18186  int32x4_t __ret;
18187  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18188  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18189  return __ret;
18190 }
18191 #endif
18192 
18193 #ifdef __LITTLE_ENDIAN__
18194 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18195  int16x8_t __ret;
18196  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18197  return __ret;
18198 }
18199 #else
18200 __ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18201  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18202  int16x8_t __ret;
18203  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18204  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18205  return __ret;
18206 }
18207 #endif
18208 
18209 #ifdef __LITTLE_ENDIAN__
18210 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18211  int32x2_t __ret;
18212  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18213  return __ret;
18214 }
18215 #else
18216 __ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18217  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18218  int32x2_t __ret;
18219  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18220  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18221  return __ret;
18222 }
18223 #endif
18224 
18225 #ifdef __LITTLE_ENDIAN__
18226 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18227  int16x4_t __ret;
18228  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18229  return __ret;
18230 }
18231 #else
18232 __ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18233  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18234  int16x4_t __ret;
18235  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18237  return __ret;
18238 }
18239 #endif
18240 
18241 #ifdef __LITTLE_ENDIAN__
18242 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18243  uint8x16_t __ret;
18244  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18245  return __ret;
18246 }
18247 #else
18248 __ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18249  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18250  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18251  uint8x16_t __ret;
18252  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18253  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18254  return __ret;
18255 }
18256 #endif
18257 
18258 #ifdef __LITTLE_ENDIAN__
18259 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18260  uint32x4_t __ret;
18261  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18262  return __ret;
18263 }
18264 #else
18265 __ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18266  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18267  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18268  uint32x4_t __ret;
18269  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18271  return __ret;
18272 }
18273 #endif
18274 
18275 #ifdef __LITTLE_ENDIAN__
18276 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18277  uint64x2_t __ret;
18278  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18279  return __ret;
18280 }
18281 #else
18282 __ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18283  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18284  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18285  uint64x2_t __ret;
18286  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18287  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18288  return __ret;
18289 }
18290 #endif
18291 
18292 #ifdef __LITTLE_ENDIAN__
18293 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18294  uint16x8_t __ret;
18295  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18296  return __ret;
18297 }
18298 #else
18299 __ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18300  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18301  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18302  uint16x8_t __ret;
18303  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18305  return __ret;
18306 }
18307 #endif
18308 
18309 #ifdef __LITTLE_ENDIAN__
18310 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18311  int8x16_t __ret;
18312  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18313  return __ret;
18314 }
18315 #else
18316 __ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18317  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18318  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18319  int8x16_t __ret;
18320  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18321  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18322  return __ret;
18323 }
18324 #endif
18325 
18326 #ifdef __LITTLE_ENDIAN__
18327 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18328  int32x4_t __ret;
18329  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18330  return __ret;
18331 }
18332 #else
18333 __ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18334  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18335  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18336  int32x4_t __ret;
18337  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18338  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18339  return __ret;
18340 }
18341 #endif
18342 
18343 #ifdef __LITTLE_ENDIAN__
18344 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18345  int64x2_t __ret;
18346  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18347  return __ret;
18348 }
18349 #else
18350 __ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18351  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18352  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18353  int64x2_t __ret;
18354  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18356  return __ret;
18357 }
18358 #endif
18359 
18360 #ifdef __LITTLE_ENDIAN__
18361 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18362  int16x8_t __ret;
18363  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18364  return __ret;
18365 }
18366 #else
18367 __ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18368  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18369  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18370  int16x8_t __ret;
18371  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18372  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18373  return __ret;
18374 }
18375 #endif
18376 
18377 #ifdef __LITTLE_ENDIAN__
18378 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18379  uint8x8_t __ret;
18380  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18381  return __ret;
18382 }
18383 #else
18384 __ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18385  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18386  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18387  uint8x8_t __ret;
18388  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18390  return __ret;
18391 }
18392 #endif
18393 
18394 #ifdef __LITTLE_ENDIAN__
18395 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18396  uint32x2_t __ret;
18397  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18398  return __ret;
18399 }
18400 #else
18401 __ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18402  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18403  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18404  uint32x2_t __ret;
18405  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18407  return __ret;
18408 }
18409 #endif
18410 
18411 #ifdef __LITTLE_ENDIAN__
18412 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18413  uint64x1_t __ret;
18414  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18415  return __ret;
18416 }
18417 #else
18418 __ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18419  uint64x1_t __ret;
18420  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18421  return __ret;
18422 }
18423 #endif
18424 
18425 #ifdef __LITTLE_ENDIAN__
18426 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18427  uint16x4_t __ret;
18428  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18429  return __ret;
18430 }
18431 #else
18432 __ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18433  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18434  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18435  uint16x4_t __ret;
18436  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18437  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18438  return __ret;
18439 }
18440 #endif
18441 
18442 #ifdef __LITTLE_ENDIAN__
18443 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18444  int8x8_t __ret;
18445  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18446  return __ret;
18447 }
18448 #else
18449 __ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18450  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18451  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18452  int8x8_t __ret;
18453  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18454  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18455  return __ret;
18456 }
18457 #endif
18458 
18459 #ifdef __LITTLE_ENDIAN__
18460 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18461  int32x2_t __ret;
18462  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18463  return __ret;
18464 }
18465 #else
18466 __ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18467  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18468  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18469  int32x2_t __ret;
18470  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18471  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18472  return __ret;
18473 }
18474 #endif
18475 
18476 #ifdef __LITTLE_ENDIAN__
18477 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18478  int64x1_t __ret;
18479  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18480  return __ret;
18481 }
18482 #else
18483 __ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18484  int64x1_t __ret;
18485  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18486  return __ret;
18487 }
18488 #endif
18489 
18490 #ifdef __LITTLE_ENDIAN__
18491 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18492  int16x4_t __ret;
18493  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18494  return __ret;
18495 }
18496 #else
18497 __ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18498  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18499  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18500  int16x4_t __ret;
18501  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18502  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18503  return __ret;
18504 }
18505 #endif
18506 
18507 #ifdef __LITTLE_ENDIAN__
18508 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18509  uint32x4_t __s0 = __p0; \
18510  uint16x4_t __ret; \
18511  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18512  __ret; \
18513 })
18514 #else
18515 #define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18516  uint32x4_t __s0 = __p0; \
18517  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18518  uint16x4_t __ret; \
18519  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
18520  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18521  __ret; \
18522 })
18523 #define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18524  uint32x4_t __s0 = __p0; \
18525  uint16x4_t __ret; \
18526  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18527  __ret; \
18528 })
18529 #endif
18530 
18531 #ifdef __LITTLE_ENDIAN__
18532 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18533  uint64x2_t __s0 = __p0; \
18534  uint32x2_t __ret; \
18535  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18536  __ret; \
18537 })
18538 #else
18539 #define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18540  uint64x2_t __s0 = __p0; \
18541  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18542  uint32x2_t __ret; \
18543  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
18544  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18545  __ret; \
18546 })
18547 #define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18548  uint64x2_t __s0 = __p0; \
18549  uint32x2_t __ret; \
18550  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18551  __ret; \
18552 })
18553 #endif
18554 
18555 #ifdef __LITTLE_ENDIAN__
18556 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18557  uint16x8_t __s0 = __p0; \
18558  uint8x8_t __ret; \
18559  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18560  __ret; \
18561 })
18562 #else
18563 #define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18564  uint16x8_t __s0 = __p0; \
18565  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18566  uint8x8_t __ret; \
18567  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
18568  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18569  __ret; \
18570 })
18571 #define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18572  uint16x8_t __s0 = __p0; \
18573  uint8x8_t __ret; \
18574  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18575  __ret; \
18576 })
18577 #endif
18578 
18579 #ifdef __LITTLE_ENDIAN__
18580 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18581  int32x4_t __s0 = __p0; \
18582  int16x4_t __ret; \
18583  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18584  __ret; \
18585 })
18586 #else
18587 #define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18588  int32x4_t __s0 = __p0; \
18589  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18590  int16x4_t __ret; \
18591  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
18592  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18593  __ret; \
18594 })
18595 #define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18596  int32x4_t __s0 = __p0; \
18597  int16x4_t __ret; \
18598  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18599  __ret; \
18600 })
18601 #endif
18602 
18603 #ifdef __LITTLE_ENDIAN__
18604 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18605  int64x2_t __s0 = __p0; \
18606  int32x2_t __ret; \
18607  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18608  __ret; \
18609 })
18610 #else
18611 #define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18612  int64x2_t __s0 = __p0; \
18613  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18614  int32x2_t __ret; \
18615  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
18616  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18617  __ret; \
18618 })
18619 #define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18620  int64x2_t __s0 = __p0; \
18621  int32x2_t __ret; \
18622  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18623  __ret; \
18624 })
18625 #endif
18626 
18627 #ifdef __LITTLE_ENDIAN__
18628 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18629  int16x8_t __s0 = __p0; \
18630  int8x8_t __ret; \
18631  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18632  __ret; \
18633 })
18634 #else
18635 #define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18636  int16x8_t __s0 = __p0; \
18637  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18638  int8x8_t __ret; \
18639  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
18640  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18641  __ret; \
18642 })
18643 #define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18644  int16x8_t __s0 = __p0; \
18645  int8x8_t __ret; \
18646  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18647  __ret; \
18648 })
18649 #endif
18650 
18651 #ifdef __LITTLE_ENDIAN__
18652 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18653  int32x4_t __s0 = __p0; \
18654  uint16x4_t __ret; \
18655  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18656  __ret; \
18657 })
18658 #else
18659 #define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18660  int32x4_t __s0 = __p0; \
18661  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18662  uint16x4_t __ret; \
18663  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
18664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18665  __ret; \
18666 })
18667 #define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18668  int32x4_t __s0 = __p0; \
18669  uint16x4_t __ret; \
18670  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18671  __ret; \
18672 })
18673 #endif
18674 
18675 #ifdef __LITTLE_ENDIAN__
18676 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18677  int64x2_t __s0 = __p0; \
18678  uint32x2_t __ret; \
18679  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18680  __ret; \
18681 })
18682 #else
18683 #define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18684  int64x2_t __s0 = __p0; \
18685  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18686  uint32x2_t __ret; \
18687  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
18688  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18689  __ret; \
18690 })
18691 #define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18692  int64x2_t __s0 = __p0; \
18693  uint32x2_t __ret; \
18694  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18695  __ret; \
18696 })
18697 #endif
18698 
18699 #ifdef __LITTLE_ENDIAN__
18700 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18701  int16x8_t __s0 = __p0; \
18702  uint8x8_t __ret; \
18703  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18704  __ret; \
18705 })
18706 #else
18707 #define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18708  int16x8_t __s0 = __p0; \
18709  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18710  uint8x8_t __ret; \
18711  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
18712  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18713  __ret; \
18714 })
18715 #define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18716  int16x8_t __s0 = __p0; \
18717  uint8x8_t __ret; \
18718  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18719  __ret; \
18720 })
18721 #endif
18722 
18723 #ifdef __LITTLE_ENDIAN__
18724 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18725  uint8x16_t __ret;
18726  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18727  return __ret;
18728 }
18729 #else
18730 __ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18731  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18732  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18733  uint8x16_t __ret;
18734  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18736  return __ret;
18737 }
18738 #endif
18739 
18740 #ifdef __LITTLE_ENDIAN__
18741 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18742  uint32x4_t __ret;
18743  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18744  return __ret;
18745 }
18746 #else
18747 __ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18748  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18749  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18750  uint32x4_t __ret;
18751  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18753  return __ret;
18754 }
18755 #endif
18756 
18757 #ifdef __LITTLE_ENDIAN__
18758 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18759  uint64x2_t __ret;
18760  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18761  return __ret;
18762 }
18763 #else
18764 __ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18765  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18766  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18767  uint64x2_t __ret;
18768  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18770  return __ret;
18771 }
18772 #endif
18773 
18774 #ifdef __LITTLE_ENDIAN__
18775 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18776  uint16x8_t __ret;
18777  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18778  return __ret;
18779 }
18780 #else
18781 __ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18782  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18783  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18784  uint16x8_t __ret;
18785  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18786  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18787  return __ret;
18788 }
18789 #endif
18790 
18791 #ifdef __LITTLE_ENDIAN__
18792 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18793  int8x16_t __ret;
18794  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18795  return __ret;
18796 }
18797 #else
18798 __ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18799  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18800  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18801  int8x16_t __ret;
18802  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18803  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18804  return __ret;
18805 }
18806 #endif
18807 
18808 #ifdef __LITTLE_ENDIAN__
18809 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18810  int32x4_t __ret;
18811  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18812  return __ret;
18813 }
18814 #else
18815 __ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18816  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18817  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18818  int32x4_t __ret;
18819  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18820  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18821  return __ret;
18822 }
18823 #endif
18824 
18825 #ifdef __LITTLE_ENDIAN__
18826 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18827  int64x2_t __ret;
18828  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18829  return __ret;
18830 }
18831 #else
18832 __ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18833  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18834  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18835  int64x2_t __ret;
18836  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18838  return __ret;
18839 }
18840 #endif
18841 
18842 #ifdef __LITTLE_ENDIAN__
18843 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18844  int16x8_t __ret;
18845  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18846  return __ret;
18847 }
18848 #else
18849 __ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18850  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18851  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18852  int16x8_t __ret;
18853  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18854  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18855  return __ret;
18856 }
18857 #endif
18858 
18859 #ifdef __LITTLE_ENDIAN__
18860 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18861  uint8x8_t __ret;
18862  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18863  return __ret;
18864 }
18865 #else
18866 __ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18867  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18868  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18869  uint8x8_t __ret;
18870  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18872  return __ret;
18873 }
18874 #endif
18875 
18876 #ifdef __LITTLE_ENDIAN__
18877 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18878  uint32x2_t __ret;
18879  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18880  return __ret;
18881 }
18882 #else
18883 __ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18884  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18885  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18886  uint32x2_t __ret;
18887  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18888  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18889  return __ret;
18890 }
18891 #endif
18892 
18893 #ifdef __LITTLE_ENDIAN__
18894 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18895  uint64x1_t __ret;
18896  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18897  return __ret;
18898 }
18899 #else
18900 __ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18901  uint64x1_t __ret;
18902  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18903  return __ret;
18904 }
18905 #endif
18906 
18907 #ifdef __LITTLE_ENDIAN__
18908 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18909  uint16x4_t __ret;
18910  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18911  return __ret;
18912 }
18913 #else
18914 __ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18915  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18916  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18917  uint16x4_t __ret;
18918  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18919  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18920  return __ret;
18921 }
18922 #endif
18923 
18924 #ifdef __LITTLE_ENDIAN__
18925 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18926  int8x8_t __ret;
18927  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18928  return __ret;
18929 }
18930 #else
18931 __ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18932  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18933  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18934  int8x8_t __ret;
18935  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18936  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18937  return __ret;
18938 }
18939 #endif
18940 
18941 #ifdef __LITTLE_ENDIAN__
18942 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18943  int32x2_t __ret;
18944  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18945  return __ret;
18946 }
18947 #else
18948 __ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18949  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18950  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18951  int32x2_t __ret;
18952  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18953  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18954  return __ret;
18955 }
18956 #endif
18957 
18958 #ifdef __LITTLE_ENDIAN__
18959 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18960  int64x1_t __ret;
18961  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18962  return __ret;
18963 }
18964 #else
18965 __ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18966  int64x1_t __ret;
18967  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18968  return __ret;
18969 }
18970 #endif
18971 
18972 #ifdef __LITTLE_ENDIAN__
18973 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18974  int16x4_t __ret;
18975  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18976  return __ret;
18977 }
18978 #else
18979 __ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18980  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18981  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18982  int16x4_t __ret;
18983  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18984  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18985  return __ret;
18986 }
18987 #endif
18988 
18989 #ifdef __LITTLE_ENDIAN__
18990 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18991  uint8x16_t __s0 = __p0; \
18992  uint8x16_t __ret; \
18993  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
18994  __ret; \
18995 })
18996 #else
18997 #define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18998  uint8x16_t __s0 = __p0; \
18999  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19000  uint8x16_t __ret; \
19001  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
19002  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19003  __ret; \
19004 })
19005 #endif
19006 
19007 #ifdef __LITTLE_ENDIAN__
19008 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
19009  uint32x4_t __s0 = __p0; \
19010  uint32x4_t __ret; \
19011  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
19012  __ret; \
19013 })
19014 #else
19015 #define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
19016  uint32x4_t __s0 = __p0; \
19017  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19018  uint32x4_t __ret; \
19019  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
19020  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19021  __ret; \
19022 })
19023 #endif
19024 
19025 #ifdef __LITTLE_ENDIAN__
19026 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19027  uint64x2_t __s0 = __p0; \
19028  uint64x2_t __ret; \
19029  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
19030  __ret; \
19031 })
19032 #else
19033 #define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19034  uint64x2_t __s0 = __p0; \
19035  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19036  uint64x2_t __ret; \
19037  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
19038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19039  __ret; \
19040 })
19041 #endif
19042 
19043 #ifdef __LITTLE_ENDIAN__
19044 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19045  uint16x8_t __s0 = __p0; \
19046  uint16x8_t __ret; \
19047  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
19048  __ret; \
19049 })
19050 #else
19051 #define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19052  uint16x8_t __s0 = __p0; \
19053  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19054  uint16x8_t __ret; \
19055  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
19056  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19057  __ret; \
19058 })
19059 #endif
19060 
19061 #ifdef __LITTLE_ENDIAN__
19062 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19063  int8x16_t __s0 = __p0; \
19064  int8x16_t __ret; \
19065  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
19066  __ret; \
19067 })
19068 #else
19069 #define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19070  int8x16_t __s0 = __p0; \
19071  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19072  int8x16_t __ret; \
19073  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
19074  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19075  __ret; \
19076 })
19077 #endif
19078 
19079 #ifdef __LITTLE_ENDIAN__
19080 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19081  int32x4_t __s0 = __p0; \
19082  int32x4_t __ret; \
19083  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
19084  __ret; \
19085 })
19086 #else
19087 #define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19088  int32x4_t __s0 = __p0; \
19089  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19090  int32x4_t __ret; \
19091  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
19092  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19093  __ret; \
19094 })
19095 #endif
19096 
19097 #ifdef __LITTLE_ENDIAN__
19098 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19099  int64x2_t __s0 = __p0; \
19100  int64x2_t __ret; \
19101  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
19102  __ret; \
19103 })
19104 #else
19105 #define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19106  int64x2_t __s0 = __p0; \
19107  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19108  int64x2_t __ret; \
19109  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
19110  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19111  __ret; \
19112 })
19113 #endif
19114 
19115 #ifdef __LITTLE_ENDIAN__
19116 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19117  int16x8_t __s0 = __p0; \
19118  int16x8_t __ret; \
19119  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
19120  __ret; \
19121 })
19122 #else
19123 #define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19124  int16x8_t __s0 = __p0; \
19125  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19126  int16x8_t __ret; \
19127  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
19128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19129  __ret; \
19130 })
19131 #endif
19132 
19133 #ifdef __LITTLE_ENDIAN__
19134 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19135  uint8x8_t __s0 = __p0; \
19136  uint8x8_t __ret; \
19137  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
19138  __ret; \
19139 })
19140 #else
19141 #define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19142  uint8x8_t __s0 = __p0; \
19143  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19144  uint8x8_t __ret; \
19145  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
19146  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19147  __ret; \
19148 })
19149 #endif
19150 
19151 #ifdef __LITTLE_ENDIAN__
19152 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19153  uint32x2_t __s0 = __p0; \
19154  uint32x2_t __ret; \
19155  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
19156  __ret; \
19157 })
19158 #else
19159 #define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19160  uint32x2_t __s0 = __p0; \
19161  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19162  uint32x2_t __ret; \
19163  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
19164  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19165  __ret; \
19166 })
19167 #endif
19168 
19169 #ifdef __LITTLE_ENDIAN__
19170 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19171  uint64x1_t __s0 = __p0; \
19172  uint64x1_t __ret; \
19173  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19174  __ret; \
19175 })
19176 #else
19177 #define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19178  uint64x1_t __s0 = __p0; \
19179  uint64x1_t __ret; \
19180  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19181  __ret; \
19182 })
19183 #endif
19184 
19185 #ifdef __LITTLE_ENDIAN__
19186 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19187  uint16x4_t __s0 = __p0; \
19188  uint16x4_t __ret; \
19189  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
19190  __ret; \
19191 })
19192 #else
19193 #define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19194  uint16x4_t __s0 = __p0; \
19195  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19196  uint16x4_t __ret; \
19197  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
19198  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19199  __ret; \
19200 })
19201 #endif
19202 
19203 #ifdef __LITTLE_ENDIAN__
19204 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19205  int8x8_t __s0 = __p0; \
19206  int8x8_t __ret; \
19207  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
19208  __ret; \
19209 })
19210 #else
19211 #define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19212  int8x8_t __s0 = __p0; \
19213  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19214  int8x8_t __ret; \
19215  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
19216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19217  __ret; \
19218 })
19219 #endif
19220 
19221 #ifdef __LITTLE_ENDIAN__
19222 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19223  int32x2_t __s0 = __p0; \
19224  int32x2_t __ret; \
19225  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
19226  __ret; \
19227 })
19228 #else
19229 #define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19230  int32x2_t __s0 = __p0; \
19231  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19232  int32x2_t __ret; \
19233  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
19234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19235  __ret; \
19236 })
19237 #endif
19238 
19239 #ifdef __LITTLE_ENDIAN__
19240 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19241  int64x1_t __s0 = __p0; \
19242  int64x1_t __ret; \
19243  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19244  __ret; \
19245 })
19246 #else
19247 #define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19248  int64x1_t __s0 = __p0; \
19249  int64x1_t __ret; \
19250  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19251  __ret; \
19252 })
19253 #endif
19254 
19255 #ifdef __LITTLE_ENDIAN__
19256 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19257  int16x4_t __s0 = __p0; \
19258  int16x4_t __ret; \
19259  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
19260  __ret; \
19261 })
19262 #else
19263 #define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19264  int16x4_t __s0 = __p0; \
19265  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19266  int16x4_t __ret; \
19267  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
19268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19269  __ret; \
19270 })
19271 #endif
19272 
19273 #ifdef __LITTLE_ENDIAN__
19274 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19275  int8x16_t __s0 = __p0; \
19276  uint8x16_t __ret; \
19277  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
19278  __ret; \
19279 })
19280 #else
19281 #define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19282  int8x16_t __s0 = __p0; \
19283  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19284  uint8x16_t __ret; \
19285  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
19286  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19287  __ret; \
19288 })
19289 #endif
19290 
19291 #ifdef __LITTLE_ENDIAN__
19292 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19293  int32x4_t __s0 = __p0; \
19294  uint32x4_t __ret; \
19295  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
19296  __ret; \
19297 })
19298 #else
19299 #define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19300  int32x4_t __s0 = __p0; \
19301  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19302  uint32x4_t __ret; \
19303  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
19304  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19305  __ret; \
19306 })
19307 #endif
19308 
19309 #ifdef __LITTLE_ENDIAN__
19310 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19311  int64x2_t __s0 = __p0; \
19312  uint64x2_t __ret; \
19313  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
19314  __ret; \
19315 })
19316 #else
19317 #define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19318  int64x2_t __s0 = __p0; \
19319  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19320  uint64x2_t __ret; \
19321  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
19322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19323  __ret; \
19324 })
19325 #endif
19326 
19327 #ifdef __LITTLE_ENDIAN__
19328 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19329  int16x8_t __s0 = __p0; \
19330  uint16x8_t __ret; \
19331  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
19332  __ret; \
19333 })
19334 #else
19335 #define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19336  int16x8_t __s0 = __p0; \
19337  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19338  uint16x8_t __ret; \
19339  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
19340  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19341  __ret; \
19342 })
19343 #endif
19344 
19345 #ifdef __LITTLE_ENDIAN__
19346 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19347  int8x8_t __s0 = __p0; \
19348  uint8x8_t __ret; \
19349  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
19350  __ret; \
19351 })
19352 #else
19353 #define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19354  int8x8_t __s0 = __p0; \
19355  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19356  uint8x8_t __ret; \
19357  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
19358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19359  __ret; \
19360 })
19361 #endif
19362 
19363 #ifdef __LITTLE_ENDIAN__
19364 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19365  int32x2_t __s0 = __p0; \
19366  uint32x2_t __ret; \
19367  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
19368  __ret; \
19369 })
19370 #else
19371 #define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19372  int32x2_t __s0 = __p0; \
19373  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19374  uint32x2_t __ret; \
19375  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
19376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19377  __ret; \
19378 })
19379 #endif
19380 
19381 #ifdef __LITTLE_ENDIAN__
19382 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19383  int64x1_t __s0 = __p0; \
19384  uint64x1_t __ret; \
19385  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19386  __ret; \
19387 })
19388 #else
19389 #define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19390  int64x1_t __s0 = __p0; \
19391  uint64x1_t __ret; \
19392  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19393  __ret; \
19394 })
19395 #endif
19396 
19397 #ifdef __LITTLE_ENDIAN__
19398 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19399  int16x4_t __s0 = __p0; \
19400  uint16x4_t __ret; \
19401  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
19402  __ret; \
19403 })
19404 #else
19405 #define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19406  int16x4_t __s0 = __p0; \
19407  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19408  uint16x4_t __ret; \
19409  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
19410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19411  __ret; \
19412 })
19413 #endif
19414 
19415 #ifdef __LITTLE_ENDIAN__
19416 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19417  uint32x4_t __s0 = __p0; \
19418  uint16x4_t __ret; \
19419  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19420  __ret; \
19421 })
19422 #else
19423 #define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19424  uint32x4_t __s0 = __p0; \
19425  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19426  uint16x4_t __ret; \
19427  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
19428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19429  __ret; \
19430 })
19431 #define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19432  uint32x4_t __s0 = __p0; \
19433  uint16x4_t __ret; \
19434  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19435  __ret; \
19436 })
19437 #endif
19438 
19439 #ifdef __LITTLE_ENDIAN__
19440 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19441  uint64x2_t __s0 = __p0; \
19442  uint32x2_t __ret; \
19443  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19444  __ret; \
19445 })
19446 #else
19447 #define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19448  uint64x2_t __s0 = __p0; \
19449  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19450  uint32x2_t __ret; \
19451  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
19452  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19453  __ret; \
19454 })
19455 #define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19456  uint64x2_t __s0 = __p0; \
19457  uint32x2_t __ret; \
19458  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19459  __ret; \
19460 })
19461 #endif
19462 
19463 #ifdef __LITTLE_ENDIAN__
19464 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19465  uint16x8_t __s0 = __p0; \
19466  uint8x8_t __ret; \
19467  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19468  __ret; \
19469 })
19470 #else
19471 #define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19472  uint16x8_t __s0 = __p0; \
19473  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19474  uint8x8_t __ret; \
19475  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
19476  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19477  __ret; \
19478 })
19479 #define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19480  uint16x8_t __s0 = __p0; \
19481  uint8x8_t __ret; \
19482  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19483  __ret; \
19484 })
19485 #endif
19486 
19487 #ifdef __LITTLE_ENDIAN__
19488 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19489  int32x4_t __s0 = __p0; \
19490  int16x4_t __ret; \
19491  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19492  __ret; \
19493 })
19494 #else
19495 #define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19496  int32x4_t __s0 = __p0; \
19497  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19498  int16x4_t __ret; \
19499  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
19500  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19501  __ret; \
19502 })
19503 #define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19504  int32x4_t __s0 = __p0; \
19505  int16x4_t __ret; \
19506  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19507  __ret; \
19508 })
19509 #endif
19510 
19511 #ifdef __LITTLE_ENDIAN__
19512 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19513  int64x2_t __s0 = __p0; \
19514  int32x2_t __ret; \
19515  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19516  __ret; \
19517 })
19518 #else
19519 #define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19520  int64x2_t __s0 = __p0; \
19521  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19522  int32x2_t __ret; \
19523  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
19524  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19525  __ret; \
19526 })
19527 #define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19528  int64x2_t __s0 = __p0; \
19529  int32x2_t __ret; \
19530  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19531  __ret; \
19532 })
19533 #endif
19534 
19535 #ifdef __LITTLE_ENDIAN__
19536 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19537  int16x8_t __s0 = __p0; \
19538  int8x8_t __ret; \
19539  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19540  __ret; \
19541 })
19542 #else
19543 #define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19544  int16x8_t __s0 = __p0; \
19545  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19546  int8x8_t __ret; \
19547  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
19548  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19549  __ret; \
19550 })
19551 #define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19552  int16x8_t __s0 = __p0; \
19553  int8x8_t __ret; \
19554  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19555  __ret; \
19556 })
19557 #endif
19558 
19559 #ifdef __LITTLE_ENDIAN__
19560 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19561  int32x4_t __s0 = __p0; \
19562  uint16x4_t __ret; \
19563  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19564  __ret; \
19565 })
19566 #else
19567 #define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19568  int32x4_t __s0 = __p0; \
19569  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19570  uint16x4_t __ret; \
19571  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
19572  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19573  __ret; \
19574 })
19575 #define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19576  int32x4_t __s0 = __p0; \
19577  uint16x4_t __ret; \
19578  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19579  __ret; \
19580 })
19581 #endif
19582 
19583 #ifdef __LITTLE_ENDIAN__
19584 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19585  int64x2_t __s0 = __p0; \
19586  uint32x2_t __ret; \
19587  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19588  __ret; \
19589 })
19590 #else
19591 #define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19592  int64x2_t __s0 = __p0; \
19593  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19594  uint32x2_t __ret; \
19595  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
19596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19597  __ret; \
19598 })
19599 #define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19600  int64x2_t __s0 = __p0; \
19601  uint32x2_t __ret; \
19602  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19603  __ret; \
19604 })
19605 #endif
19606 
19607 #ifdef __LITTLE_ENDIAN__
19608 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19609  int16x8_t __s0 = __p0; \
19610  uint8x8_t __ret; \
19611  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19612  __ret; \
19613 })
19614 #else
19615 #define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19616  int16x8_t __s0 = __p0; \
19617  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19618  uint8x8_t __ret; \
19619  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
19620  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19621  __ret; \
19622 })
19623 #define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19624  int16x8_t __s0 = __p0; \
19625  uint8x8_t __ret; \
19626  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19627  __ret; \
19628 })
19629 #endif
19630 
19631 #ifdef __LITTLE_ENDIAN__
19632 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19633  uint8x16_t __ret;
19634  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
19635  return __ret;
19636 }
19637 #else
19638 __ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19639  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19640  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19641  uint8x16_t __ret;
19642  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
19643  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19644  return __ret;
19645 }
19646 #endif
19647 
19648 #ifdef __LITTLE_ENDIAN__
19649 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19650  uint32x4_t __ret;
19651  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
19652  return __ret;
19653 }
19654 #else
19655 __ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19656  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19657  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19658  uint32x4_t __ret;
19659  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
19660  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19661  return __ret;
19662 }
19663 #endif
19664 
19665 #ifdef __LITTLE_ENDIAN__
19666 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19667  uint64x2_t __ret;
19668  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
19669  return __ret;
19670 }
19671 #else
19672 __ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19673  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19674  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19675  uint64x2_t __ret;
19676  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
19677  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19678  return __ret;
19679 }
19680 #endif
19681 
19682 #ifdef __LITTLE_ENDIAN__
19683 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19684  uint16x8_t __ret;
19685  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
19686  return __ret;
19687 }
19688 #else
19689 __ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19690  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19691  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19692  uint16x8_t __ret;
19693  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
19694  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19695  return __ret;
19696 }
19697 #endif
19698 
19699 #ifdef __LITTLE_ENDIAN__
19700 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19701  int8x16_t __ret;
19702  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
19703  return __ret;
19704 }
19705 #else
19706 __ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19707  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19708  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19709  int8x16_t __ret;
19710  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
19711  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19712  return __ret;
19713 }
19714 #endif
19715 
19716 #ifdef __LITTLE_ENDIAN__
19717 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19718  int32x4_t __ret;
19719  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19720  return __ret;
19721 }
19722 #else
19723 __ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19724  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19725  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19726  int32x4_t __ret;
19727  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19728  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19729  return __ret;
19730 }
19731 __ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19732  int32x4_t __ret;
19733  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19734  return __ret;
19735 }
19736 #endif
19737 
19738 #ifdef __LITTLE_ENDIAN__
19739 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19740  int64x2_t __ret;
19741  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
19742  return __ret;
19743 }
19744 #else
19745 __ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19746  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19747  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19748  int64x2_t __ret;
19749  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
19750  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19751  return __ret;
19752 }
19753 #endif
19754 
19755 #ifdef __LITTLE_ENDIAN__
19756 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19757  int16x8_t __ret;
19758  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19759  return __ret;
19760 }
19761 #else
19762 __ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19763  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19764  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19765  int16x8_t __ret;
19766  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19767  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19768  return __ret;
19769 }
19770 __ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19771  int16x8_t __ret;
19772  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19773  return __ret;
19774 }
19775 #endif
19776 
19777 #ifdef __LITTLE_ENDIAN__
19778 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19779  uint8x8_t __ret;
19780  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
19781  return __ret;
19782 }
19783 #else
19784 __ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19785  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19786  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19787  uint8x8_t __ret;
19788  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
19789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19790  return __ret;
19791 }
19792 #endif
19793 
19794 #ifdef __LITTLE_ENDIAN__
19795 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19796  uint32x2_t __ret;
19797  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
19798  return __ret;
19799 }
19800 #else
19801 __ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19802  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19803  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19804  uint32x2_t __ret;
19805  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
19806  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19807  return __ret;
19808 }
19809 #endif
19810 
19811 #ifdef __LITTLE_ENDIAN__
19812 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19813  uint64x1_t __ret;
19814  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19815  return __ret;
19816 }
19817 #else
19818 __ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19819  uint64x1_t __ret;
19820  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19821  return __ret;
19822 }
19823 #endif
19824 
19825 #ifdef __LITTLE_ENDIAN__
19826 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19827  uint16x4_t __ret;
19828  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
19829  return __ret;
19830 }
19831 #else
19832 __ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19833  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19834  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19835  uint16x4_t __ret;
19836  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
19837  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19838  return __ret;
19839 }
19840 #endif
19841 
19842 #ifdef __LITTLE_ENDIAN__
19843 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19844  int8x8_t __ret;
19845  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
19846  return __ret;
19847 }
19848 #else
19849 __ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19850  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19851  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19852  int8x8_t __ret;
19853  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
19854  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19855  return __ret;
19856 }
19857 #endif
19858 
19859 #ifdef __LITTLE_ENDIAN__
19860 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19861  int32x2_t __ret;
19862  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19863  return __ret;
19864 }
19865 #else
19866 __ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19867  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19868  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19869  int32x2_t __ret;
19870  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19871  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19872  return __ret;
19873 }
19874 __ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19875  int32x2_t __ret;
19876  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19877  return __ret;
19878 }
19879 #endif
19880 
19881 #ifdef __LITTLE_ENDIAN__
19882 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19883  int64x1_t __ret;
19884  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19885  return __ret;
19886 }
19887 #else
19888 __ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19889  int64x1_t __ret;
19890  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19891  return __ret;
19892 }
19893 #endif
19894 
19895 #ifdef __LITTLE_ENDIAN__
19896 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19897  int16x4_t __ret;
19898  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19899  return __ret;
19900 }
19901 #else
19902 __ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19903  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19904  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19905  int16x4_t __ret;
19906  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19907  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19908  return __ret;
19909 }
19910 __ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19911  int16x4_t __ret;
19912  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19913  return __ret;
19914 }
19915 #endif
19916 
19917 #ifdef __LITTLE_ENDIAN__
19918 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19919  uint16x4_t __ret;
19920  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19921  return __ret;
19922 }
19923 #else
19924 __ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19925  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19926  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19927  uint16x4_t __ret;
19928  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
19929  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19930  return __ret;
19931 }
19932 __ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19933  uint16x4_t __ret;
19934  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19935  return __ret;
19936 }
19937 #endif
19938 
19939 #ifdef __LITTLE_ENDIAN__
19940 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19941  uint32x2_t __ret;
19942  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19943  return __ret;
19944 }
19945 #else
19946 __ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19947  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19948  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19949  uint32x2_t __ret;
19950  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
19951  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19952  return __ret;
19953 }
19954 __ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19955  uint32x2_t __ret;
19956  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19957  return __ret;
19958 }
19959 #endif
19960 
19961 #ifdef __LITTLE_ENDIAN__
19962 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19963  uint8x8_t __ret;
19964  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19965  return __ret;
19966 }
19967 #else
19968 __ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19969  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19970  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19971  uint8x8_t __ret;
19972  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
19973  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19974  return __ret;
19975 }
19976 __ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19977  uint8x8_t __ret;
19978  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19979  return __ret;
19980 }
19981 #endif
19982 
19983 #ifdef __LITTLE_ENDIAN__
19984 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19985  int16x4_t __ret;
19986  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19987  return __ret;
19988 }
19989 #else
19990 __ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19991  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19992  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19993  int16x4_t __ret;
19994  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
19995  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19996  return __ret;
19997 }
19998 __ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19999  int16x4_t __ret;
20000  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
20001  return __ret;
20002 }
20003 #endif
20004 
20005 #ifdef __LITTLE_ENDIAN__
20006 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20007  int32x2_t __ret;
20008  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
20009  return __ret;
20010 }
20011 #else
20012 __ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20013  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20014  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20015  int32x2_t __ret;
20016  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
20017  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20018  return __ret;
20019 }
20020 __ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20021  int32x2_t __ret;
20022  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
20023  return __ret;
20024 }
20025 #endif
20026 
20027 #ifdef __LITTLE_ENDIAN__
20028 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20029  int8x8_t __ret;
20030  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
20031  return __ret;
20032 }
20033 #else
20034 __ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20035  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20036  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20037  int8x8_t __ret;
20038  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
20039  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20040  return __ret;
20041 }
20042 __ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20043  int8x8_t __ret;
20044  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
20045  return __ret;
20046 }
20047 #endif
20048 
20049 #ifdef __LITTLE_ENDIAN__
20050 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20051  uint32x4_t __ret;
20052  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
20053  return __ret;
20054 }
20055 #else
20056 __ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20057  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20058  uint32x4_t __ret;
20059  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
20060  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20061  return __ret;
20062 }
20063 #endif
20064 
20065 #ifdef __LITTLE_ENDIAN__
20066 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20067  float32x4_t __ret;
20068  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
20069  return __ret;
20070 }
20071 #else
20072 __ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20073  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20074  float32x4_t __ret;
20075  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
20076  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20077  return __ret;
20078 }
20079 #endif
20080 
20081 #ifdef __LITTLE_ENDIAN__
20082 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20083  uint32x2_t __ret;
20084  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
20085  return __ret;
20086 }
20087 #else
20088 __ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20089  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20090  uint32x2_t __ret;
20091  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
20092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20093  return __ret;
20094 }
20095 #endif
20096 
20097 #ifdef __LITTLE_ENDIAN__
20098 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20099  float32x2_t __ret;
20100  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
20101  return __ret;
20102 }
20103 #else
20104 __ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20105  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20106  float32x2_t __ret;
20107  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
20108  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20109  return __ret;
20110 }
20111 #endif
20112 
20113 #ifdef __LITTLE_ENDIAN__
20114 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20115  float32x4_t __ret;
20116  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
20117  return __ret;
20118 }
20119 #else
20120 __ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20121  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20122  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20123  float32x4_t __ret;
20124  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
20125  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20126  return __ret;
20127 }
20128 #endif
20129 
20130 #ifdef __LITTLE_ENDIAN__
20131 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20132  float32x2_t __ret;
20133  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
20134  return __ret;
20135 }
20136 #else
20137 __ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20138  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20139  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20140  float32x2_t __ret;
20141  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
20142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20143  return __ret;
20144 }
20145 #endif
20146 
20147 #ifdef __LITTLE_ENDIAN__
20148 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20149  poly8x8_t __ret;
20150  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20151  return __ret;
20152 }
20153 #else
20154 __ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20155  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20156  poly8x8_t __ret;
20157  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20158  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20159  return __ret;
20160 }
20161 #endif
20162 
20163 #ifdef __LITTLE_ENDIAN__
20164 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20165  poly8x16_t __ret;
20166  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20167  return __ret;
20168 }
20169 #else
20170 __ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20171  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20172  poly8x16_t __ret;
20173  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20174  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20175  return __ret;
20176 }
20177 #endif
20178 
20179 #ifdef __LITTLE_ENDIAN__
20180 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20181  uint8x16_t __ret;
20182  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20183  return __ret;
20184 }
20185 #else
20186 __ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20187  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20188  uint8x16_t __ret;
20189  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20190  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20191  return __ret;
20192 }
20193 #endif
20194 
20195 #ifdef __LITTLE_ENDIAN__
20196 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20197  int8x16_t __ret;
20198  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20199  return __ret;
20200 }
20201 #else
20202 __ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20203  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20204  int8x16_t __ret;
20205  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20206  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20207  return __ret;
20208 }
20209 #endif
20210 
20211 #ifdef __LITTLE_ENDIAN__
20212 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20213  uint8x8_t __ret;
20214  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20215  return __ret;
20216 }
20217 #else
20218 __ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20219  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20220  uint8x8_t __ret;
20221  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20222  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20223  return __ret;
20224 }
20225 #endif
20226 
20227 #ifdef __LITTLE_ENDIAN__
20228 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
20229  int8x8_t __ret;
20230  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20231  return __ret;
20232 }
20233 #else
20234 __ai int8x8_t vrev16_s8(int8x8_t __p0) {
20235  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20236  int8x8_t __ret;
20237  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20238  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20239  return __ret;
20240 }
20241 #endif
20242 
20243 #ifdef __LITTLE_ENDIAN__
20244 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20245  poly8x8_t __ret;
20246  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20247  return __ret;
20248 }
20249 #else
20250 __ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20251  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20252  poly8x8_t __ret;
20253  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20254  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20255  return __ret;
20256 }
20257 #endif
20258 
20259 #ifdef __LITTLE_ENDIAN__
20260 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20261  poly16x4_t __ret;
20262  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20263  return __ret;
20264 }
20265 #else
20266 __ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20267  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20268  poly16x4_t __ret;
20269  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20271  return __ret;
20272 }
20273 #endif
20274 
20275 #ifdef __LITTLE_ENDIAN__
20276 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20277  poly8x16_t __ret;
20278  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20279  return __ret;
20280 }
20281 #else
20282 __ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20283  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20284  poly8x16_t __ret;
20285  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20286  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20287  return __ret;
20288 }
20289 #endif
20290 
20291 #ifdef __LITTLE_ENDIAN__
20292 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20293  poly16x8_t __ret;
20294  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20295  return __ret;
20296 }
20297 #else
20298 __ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20299  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20300  poly16x8_t __ret;
20301  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20302  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20303  return __ret;
20304 }
20305 #endif
20306 
20307 #ifdef __LITTLE_ENDIAN__
20308 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20309  uint8x16_t __ret;
20310  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20311  return __ret;
20312 }
20313 #else
20314 __ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20315  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20316  uint8x16_t __ret;
20317  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20318  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20319  return __ret;
20320 }
20321 #endif
20322 
20323 #ifdef __LITTLE_ENDIAN__
20324 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20325  uint16x8_t __ret;
20326  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20327  return __ret;
20328 }
20329 #else
20330 __ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20331  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20332  uint16x8_t __ret;
20333  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20334  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20335  return __ret;
20336 }
20337 #endif
20338 
20339 #ifdef __LITTLE_ENDIAN__
20340 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20341  int8x16_t __ret;
20342  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20343  return __ret;
20344 }
20345 #else
20346 __ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20347  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20348  int8x16_t __ret;
20349  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20350  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20351  return __ret;
20352 }
20353 #endif
20354 
20355 #ifdef __LITTLE_ENDIAN__
20356 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20357  int16x8_t __ret;
20358  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20359  return __ret;
20360 }
20361 #else
20362 __ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20363  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20364  int16x8_t __ret;
20365  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20367  return __ret;
20368 }
20369 #endif
20370 
20371 #ifdef __LITTLE_ENDIAN__
20372 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20373  uint8x8_t __ret;
20374  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20375  return __ret;
20376 }
20377 #else
20378 __ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20379  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20380  uint8x8_t __ret;
20381  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20382  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20383  return __ret;
20384 }
20385 #endif
20386 
20387 #ifdef __LITTLE_ENDIAN__
20388 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20389  uint16x4_t __ret;
20390  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20391  return __ret;
20392 }
20393 #else
20394 __ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20395  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20396  uint16x4_t __ret;
20397  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20398  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20399  return __ret;
20400 }
20401 #endif
20402 
20403 #ifdef __LITTLE_ENDIAN__
20404 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
20405  int8x8_t __ret;
20406  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20407  return __ret;
20408 }
20409 #else
20410 __ai int8x8_t vrev32_s8(int8x8_t __p0) {
20411  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20412  int8x8_t __ret;
20413  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20414  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20415  return __ret;
20416 }
20417 #endif
20418 
20419 #ifdef __LITTLE_ENDIAN__
20420 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
20421  int16x4_t __ret;
20422  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20423  return __ret;
20424 }
20425 #else
20426 __ai int16x4_t vrev32_s16(int16x4_t __p0) {
20427  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20428  int16x4_t __ret;
20429  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20430  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20431  return __ret;
20432 }
20433 #endif
20434 
20435 #ifdef __LITTLE_ENDIAN__
20436 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20437  poly8x8_t __ret;
20438  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20439  return __ret;
20440 }
20441 #else
20442 __ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20443  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20444  poly8x8_t __ret;
20445  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20446  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20447  return __ret;
20448 }
20449 #endif
20450 
20451 #ifdef __LITTLE_ENDIAN__
20452 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20453  poly16x4_t __ret;
20454  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20455  return __ret;
20456 }
20457 #else
20458 __ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20459  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20460  poly16x4_t __ret;
20461  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20462  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20463  return __ret;
20464 }
20465 #endif
20466 
20467 #ifdef __LITTLE_ENDIAN__
20468 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20469  poly8x16_t __ret;
20470  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20471  return __ret;
20472 }
20473 #else
20474 __ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20475  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20476  poly8x16_t __ret;
20477  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20478  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20479  return __ret;
20480 }
20481 #endif
20482 
20483 #ifdef __LITTLE_ENDIAN__
20484 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20485  poly16x8_t __ret;
20486  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20487  return __ret;
20488 }
20489 #else
20490 __ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20491  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20492  poly16x8_t __ret;
20493  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20495  return __ret;
20496 }
20497 #endif
20498 
20499 #ifdef __LITTLE_ENDIAN__
20500 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20501  uint8x16_t __ret;
20502  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20503  return __ret;
20504 }
20505 #else
20506 __ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20507  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20508  uint8x16_t __ret;
20509  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20511  return __ret;
20512 }
20513 #endif
20514 
20515 #ifdef __LITTLE_ENDIAN__
20516 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20517  uint32x4_t __ret;
20518  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20519  return __ret;
20520 }
20521 #else
20522 __ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20523  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20524  uint32x4_t __ret;
20525  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20526  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20527  return __ret;
20528 }
20529 #endif
20530 
20531 #ifdef __LITTLE_ENDIAN__
20532 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20533  uint16x8_t __ret;
20534  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20535  return __ret;
20536 }
20537 #else
20538 __ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20539  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20540  uint16x8_t __ret;
20541  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20543  return __ret;
20544 }
20545 #endif
20546 
20547 #ifdef __LITTLE_ENDIAN__
20548 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20549  int8x16_t __ret;
20550  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20551  return __ret;
20552 }
20553 #else
20554 __ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20555  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20556  int8x16_t __ret;
20557  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20558  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20559  return __ret;
20560 }
20561 #endif
20562 
20563 #ifdef __LITTLE_ENDIAN__
20564 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20565  float32x4_t __ret;
20566  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20567  return __ret;
20568 }
20569 #else
20570 __ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20571  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20572  float32x4_t __ret;
20573  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20575  return __ret;
20576 }
20577 #endif
20578 
20579 #ifdef __LITTLE_ENDIAN__
20580 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20581  int32x4_t __ret;
20582  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20583  return __ret;
20584 }
20585 #else
20586 __ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20587  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20588  int32x4_t __ret;
20589  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20590  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20591  return __ret;
20592 }
20593 #endif
20594 
20595 #ifdef __LITTLE_ENDIAN__
20596 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20597  int16x8_t __ret;
20598  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20599  return __ret;
20600 }
20601 #else
20602 __ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20603  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20604  int16x8_t __ret;
20605  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20606  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20607  return __ret;
20608 }
20609 #endif
20610 
20611 #ifdef __LITTLE_ENDIAN__
20612 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20613  uint8x8_t __ret;
20614  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20615  return __ret;
20616 }
20617 #else
20618 __ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20619  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20620  uint8x8_t __ret;
20621  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20623  return __ret;
20624 }
20625 #endif
20626 
20627 #ifdef __LITTLE_ENDIAN__
20628 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20629  uint32x2_t __ret;
20630  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20631  return __ret;
20632 }
20633 #else
20634 __ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20635  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20636  uint32x2_t __ret;
20637  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20639  return __ret;
20640 }
20641 #endif
20642 
20643 #ifdef __LITTLE_ENDIAN__
20644 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20645  uint16x4_t __ret;
20646  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20647  return __ret;
20648 }
20649 #else
20650 __ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20651  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20652  uint16x4_t __ret;
20653  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20654  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20655  return __ret;
20656 }
20657 #endif
20658 
20659 #ifdef __LITTLE_ENDIAN__
20660 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
20661  int8x8_t __ret;
20662  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20663  return __ret;
20664 }
20665 #else
20666 __ai int8x8_t vrev64_s8(int8x8_t __p0) {
20667  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20668  int8x8_t __ret;
20669  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20670  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20671  return __ret;
20672 }
20673 #endif
20674 
20675 #ifdef __LITTLE_ENDIAN__
20676 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
20677  float32x2_t __ret;
20678  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20679  return __ret;
20680 }
20681 #else
20682 __ai float32x2_t vrev64_f32(float32x2_t __p0) {
20683  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20684  float32x2_t __ret;
20685  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20686  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20687  return __ret;
20688 }
20689 #endif
20690 
20691 #ifdef __LITTLE_ENDIAN__
20692 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
20693  int32x2_t __ret;
20694  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20695  return __ret;
20696 }
20697 #else
20698 __ai int32x2_t vrev64_s32(int32x2_t __p0) {
20699  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20700  int32x2_t __ret;
20701  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20702  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20703  return __ret;
20704 }
20705 #endif
20706 
20707 #ifdef __LITTLE_ENDIAN__
20708 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
20709  int16x4_t __ret;
20710  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20711  return __ret;
20712 }
20713 #else
20714 __ai int16x4_t vrev64_s16(int16x4_t __p0) {
20715  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20716  int16x4_t __ret;
20717  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20719  return __ret;
20720 }
20721 #endif
20722 
20723 #ifdef __LITTLE_ENDIAN__
20724 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20725  uint8x16_t __ret;
20726  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20727  return __ret;
20728 }
20729 #else
20730 __ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20731  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20732  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20733  uint8x16_t __ret;
20734  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20736  return __ret;
20737 }
20738 #endif
20739 
20740 #ifdef __LITTLE_ENDIAN__
20741 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20742  uint32x4_t __ret;
20743  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20744  return __ret;
20745 }
20746 #else
20747 __ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20748  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20749  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20750  uint32x4_t __ret;
20751  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20753  return __ret;
20754 }
20755 #endif
20756 
20757 #ifdef __LITTLE_ENDIAN__
20758 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20759  uint16x8_t __ret;
20760  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20761  return __ret;
20762 }
20763 #else
20764 __ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20765  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20766  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20767  uint16x8_t __ret;
20768  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20770  return __ret;
20771 }
20772 #endif
20773 
20774 #ifdef __LITTLE_ENDIAN__
20775 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20776  int8x16_t __ret;
20777  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20778  return __ret;
20779 }
20780 #else
20781 __ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20782  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20783  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20784  int8x16_t __ret;
20785  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20786  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20787  return __ret;
20788 }
20789 #endif
20790 
20791 #ifdef __LITTLE_ENDIAN__
20792 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20793  int32x4_t __ret;
20794  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20795  return __ret;
20796 }
20797 #else
20798 __ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20799  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20800  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20801  int32x4_t __ret;
20802  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20804  return __ret;
20805 }
20806 #endif
20807 
20808 #ifdef __LITTLE_ENDIAN__
20809 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20810  int16x8_t __ret;
20811  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20812  return __ret;
20813 }
20814 #else
20815 __ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20816  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20817  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20818  int16x8_t __ret;
20819  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20821  return __ret;
20822 }
20823 #endif
20824 
20825 #ifdef __LITTLE_ENDIAN__
20826 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20827  uint8x8_t __ret;
20828  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20829  return __ret;
20830 }
20831 #else
20832 __ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20833  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20834  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20835  uint8x8_t __ret;
20836  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20837  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20838  return __ret;
20839 }
20840 #endif
20841 
20842 #ifdef __LITTLE_ENDIAN__
20843 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20844  uint32x2_t __ret;
20845  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20846  return __ret;
20847 }
20848 #else
20849 __ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20850  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20851  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20852  uint32x2_t __ret;
20853  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20854  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20855  return __ret;
20856 }
20857 #endif
20858 
20859 #ifdef __LITTLE_ENDIAN__
20860 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20861  uint16x4_t __ret;
20862  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20863  return __ret;
20864 }
20865 #else
20866 __ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20867  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20868  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20869  uint16x4_t __ret;
20870  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20871  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20872  return __ret;
20873 }
20874 #endif
20875 
20876 #ifdef __LITTLE_ENDIAN__
20877 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20878  int8x8_t __ret;
20879  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20880  return __ret;
20881 }
20882 #else
20883 __ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20884  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20885  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20886  int8x8_t __ret;
20887  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20888  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20889  return __ret;
20890 }
20891 #endif
20892 
20893 #ifdef __LITTLE_ENDIAN__
20894 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20895  int32x2_t __ret;
20896  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20897  return __ret;
20898 }
20899 #else
20900 __ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20901  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20902  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20903  int32x2_t __ret;
20904  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20906  return __ret;
20907 }
20908 #endif
20909 
20910 #ifdef __LITTLE_ENDIAN__
20911 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20912  int16x4_t __ret;
20913  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20914  return __ret;
20915 }
20916 #else
20917 __ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20918  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20919  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20920  int16x4_t __ret;
20921  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20922  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20923  return __ret;
20924 }
20925 #endif
20926 
20927 #ifdef __LITTLE_ENDIAN__
20928 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20929  uint8x16_t __ret;
20930  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20931  return __ret;
20932 }
20933 #else
20934 __ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20935  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20936  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20937  uint8x16_t __ret;
20938  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20940  return __ret;
20941 }
20942 #endif
20943 
20944 #ifdef __LITTLE_ENDIAN__
20945 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20946  uint32x4_t __ret;
20947  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20948  return __ret;
20949 }
20950 #else
20951 __ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20952  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20953  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20954  uint32x4_t __ret;
20955  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20957  return __ret;
20958 }
20959 #endif
20960 
20961 #ifdef __LITTLE_ENDIAN__
20962 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20963  uint64x2_t __ret;
20964  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20965  return __ret;
20966 }
20967 #else
20968 __ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20969  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20970  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20971  uint64x2_t __ret;
20972  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20974  return __ret;
20975 }
20976 #endif
20977 
20978 #ifdef __LITTLE_ENDIAN__
20979 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20980  uint16x8_t __ret;
20981  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20982  return __ret;
20983 }
20984 #else
20985 __ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20986  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20987  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20988  uint16x8_t __ret;
20989  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20990  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20991  return __ret;
20992 }
20993 #endif
20994 
20995 #ifdef __LITTLE_ENDIAN__
20996 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20997  int8x16_t __ret;
20998  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20999  return __ret;
21000 }
21001 #else
21002 __ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
21003  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21004  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21005  int8x16_t __ret;
21006  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
21007  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21008  return __ret;
21009 }
21010 #endif
21011 
21012 #ifdef __LITTLE_ENDIAN__
21013 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
21014  int32x4_t __ret;
21015  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
21016  return __ret;
21017 }
21018 #else
21019 __ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
21020  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21021  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21022  int32x4_t __ret;
21023  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
21024  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21025  return __ret;
21026 }
21027 #endif
21028 
21029 #ifdef __LITTLE_ENDIAN__
21030 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
21031  int64x2_t __ret;
21032  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
21033  return __ret;
21034 }
21035 #else
21036 __ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
21037  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21038  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21039  int64x2_t __ret;
21040  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
21041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21042  return __ret;
21043 }
21044 #endif
21045 
21046 #ifdef __LITTLE_ENDIAN__
21047 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21048  int16x8_t __ret;
21049  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21050  return __ret;
21051 }
21052 #else
21053 __ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21054  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21055  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21056  int16x8_t __ret;
21057  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
21058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21059  return __ret;
21060 }
21061 #endif
21062 
21063 #ifdef __LITTLE_ENDIAN__
21064 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21065  uint8x8_t __ret;
21066  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
21067  return __ret;
21068 }
21069 #else
21070 __ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21071  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21072  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21073  uint8x8_t __ret;
21074  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
21075  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21076  return __ret;
21077 }
21078 #endif
21079 
21080 #ifdef __LITTLE_ENDIAN__
21081 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21082  uint32x2_t __ret;
21083  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
21084  return __ret;
21085 }
21086 #else
21087 __ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21088  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21089  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21090  uint32x2_t __ret;
21091  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21093  return __ret;
21094 }
21095 #endif
21096 
21097 #ifdef __LITTLE_ENDIAN__
21098 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21099  uint64x1_t __ret;
21100  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21101  return __ret;
21102 }
21103 #else
21104 __ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21105  uint64x1_t __ret;
21106  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21107  return __ret;
21108 }
21109 #endif
21110 
21111 #ifdef __LITTLE_ENDIAN__
21112 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21113  uint16x4_t __ret;
21114  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21115  return __ret;
21116 }
21117 #else
21118 __ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21119  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21120  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21121  uint16x4_t __ret;
21122  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21124  return __ret;
21125 }
21126 #endif
21127 
21128 #ifdef __LITTLE_ENDIAN__
21129 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21130  int8x8_t __ret;
21131  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21132  return __ret;
21133 }
21134 #else
21135 __ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21136  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21137  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21138  int8x8_t __ret;
21139  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21141  return __ret;
21142 }
21143 #endif
21144 
21145 #ifdef __LITTLE_ENDIAN__
21146 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21147  int32x2_t __ret;
21148  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21149  return __ret;
21150 }
21151 #else
21152 __ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21153  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21154  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21155  int32x2_t __ret;
21156  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21157  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21158  return __ret;
21159 }
21160 #endif
21161 
21162 #ifdef __LITTLE_ENDIAN__
21163 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21164  int64x1_t __ret;
21165  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21166  return __ret;
21167 }
21168 #else
21169 __ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21170  int64x1_t __ret;
21171  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21172  return __ret;
21173 }
21174 #endif
21175 
21176 #ifdef __LITTLE_ENDIAN__
21177 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21178  int16x4_t __ret;
21179  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21180  return __ret;
21181 }
21182 #else
21183 __ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21184  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21185  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21186  int16x4_t __ret;
21187  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21188  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21189  return __ret;
21190 }
21191 #endif
21192 
21193 #ifdef __LITTLE_ENDIAN__
21194 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21195  uint8x16_t __s0 = __p0; \
21196  uint8x16_t __ret; \
21197  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
21198  __ret; \
21199 })
21200 #else
21201 #define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21202  uint8x16_t __s0 = __p0; \
21203  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21204  uint8x16_t __ret; \
21205  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
21206  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21207  __ret; \
21208 })
21209 #endif
21210 
21211 #ifdef __LITTLE_ENDIAN__
21212 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21213  uint32x4_t __s0 = __p0; \
21214  uint32x4_t __ret; \
21215  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
21216  __ret; \
21217 })
21218 #else
21219 #define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21220  uint32x4_t __s0 = __p0; \
21221  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21222  uint32x4_t __ret; \
21223  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
21224  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21225  __ret; \
21226 })
21227 #endif
21228 
21229 #ifdef __LITTLE_ENDIAN__
21230 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21231  uint64x2_t __s0 = __p0; \
21232  uint64x2_t __ret; \
21233  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
21234  __ret; \
21235 })
21236 #else
21237 #define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21238  uint64x2_t __s0 = __p0; \
21239  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21240  uint64x2_t __ret; \
21241  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
21242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21243  __ret; \
21244 })
21245 #endif
21246 
21247 #ifdef __LITTLE_ENDIAN__
21248 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21249  uint16x8_t __s0 = __p0; \
21250  uint16x8_t __ret; \
21251  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
21252  __ret; \
21253 })
21254 #else
21255 #define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21256  uint16x8_t __s0 = __p0; \
21257  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21258  uint16x8_t __ret; \
21259  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
21260  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21261  __ret; \
21262 })
21263 #endif
21264 
21265 #ifdef __LITTLE_ENDIAN__
21266 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21267  int8x16_t __s0 = __p0; \
21268  int8x16_t __ret; \
21269  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
21270  __ret; \
21271 })
21272 #else
21273 #define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21274  int8x16_t __s0 = __p0; \
21275  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21276  int8x16_t __ret; \
21277  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
21278  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21279  __ret; \
21280 })
21281 #endif
21282 
21283 #ifdef __LITTLE_ENDIAN__
21284 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21285  int32x4_t __s0 = __p0; \
21286  int32x4_t __ret; \
21287  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
21288  __ret; \
21289 })
21290 #else
21291 #define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21292  int32x4_t __s0 = __p0; \
21293  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21294  int32x4_t __ret; \
21295  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
21296  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21297  __ret; \
21298 })
21299 #endif
21300 
21301 #ifdef __LITTLE_ENDIAN__
21302 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21303  int64x2_t __s0 = __p0; \
21304  int64x2_t __ret; \
21305  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
21306  __ret; \
21307 })
21308 #else
21309 #define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21310  int64x2_t __s0 = __p0; \
21311  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21312  int64x2_t __ret; \
21313  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
21314  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21315  __ret; \
21316 })
21317 #endif
21318 
21319 #ifdef __LITTLE_ENDIAN__
21320 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21321  int16x8_t __s0 = __p0; \
21322  int16x8_t __ret; \
21323  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
21324  __ret; \
21325 })
21326 #else
21327 #define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21328  int16x8_t __s0 = __p0; \
21329  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21330  int16x8_t __ret; \
21331  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
21332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21333  __ret; \
21334 })
21335 #endif
21336 
21337 #ifdef __LITTLE_ENDIAN__
21338 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21339  uint8x8_t __s0 = __p0; \
21340  uint8x8_t __ret; \
21341  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
21342  __ret; \
21343 })
21344 #else
21345 #define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21346  uint8x8_t __s0 = __p0; \
21347  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21348  uint8x8_t __ret; \
21349  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
21350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21351  __ret; \
21352 })
21353 #endif
21354 
21355 #ifdef __LITTLE_ENDIAN__
21356 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21357  uint32x2_t __s0 = __p0; \
21358  uint32x2_t __ret; \
21359  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
21360  __ret; \
21361 })
21362 #else
21363 #define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21364  uint32x2_t __s0 = __p0; \
21365  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21366  uint32x2_t __ret; \
21367  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
21368  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21369  __ret; \
21370 })
21371 #endif
21372 
21373 #ifdef __LITTLE_ENDIAN__
21374 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21375  uint64x1_t __s0 = __p0; \
21376  uint64x1_t __ret; \
21377  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21378  __ret; \
21379 })
21380 #else
21381 #define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21382  uint64x1_t __s0 = __p0; \
21383  uint64x1_t __ret; \
21384  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21385  __ret; \
21386 })
21387 #endif
21388 
21389 #ifdef __LITTLE_ENDIAN__
21390 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21391  uint16x4_t __s0 = __p0; \
21392  uint16x4_t __ret; \
21393  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
21394  __ret; \
21395 })
21396 #else
21397 #define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21398  uint16x4_t __s0 = __p0; \
21399  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21400  uint16x4_t __ret; \
21401  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
21402  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21403  __ret; \
21404 })
21405 #endif
21406 
21407 #ifdef __LITTLE_ENDIAN__
21408 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21409  int8x8_t __s0 = __p0; \
21410  int8x8_t __ret; \
21411  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
21412  __ret; \
21413 })
21414 #else
21415 #define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21416  int8x8_t __s0 = __p0; \
21417  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21418  int8x8_t __ret; \
21419  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
21420  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21421  __ret; \
21422 })
21423 #endif
21424 
21425 #ifdef __LITTLE_ENDIAN__
21426 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21427  int32x2_t __s0 = __p0; \
21428  int32x2_t __ret; \
21429  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
21430  __ret; \
21431 })
21432 #else
21433 #define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21434  int32x2_t __s0 = __p0; \
21435  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21436  int32x2_t __ret; \
21437  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
21438  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21439  __ret; \
21440 })
21441 #endif
21442 
21443 #ifdef __LITTLE_ENDIAN__
21444 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21445  int64x1_t __s0 = __p0; \
21446  int64x1_t __ret; \
21447  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21448  __ret; \
21449 })
21450 #else
21451 #define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21452  int64x1_t __s0 = __p0; \
21453  int64x1_t __ret; \
21454  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21455  __ret; \
21456 })
21457 #endif
21458 
21459 #ifdef __LITTLE_ENDIAN__
21460 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21461  int16x4_t __s0 = __p0; \
21462  int16x4_t __ret; \
21463  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
21464  __ret; \
21465 })
21466 #else
21467 #define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21468  int16x4_t __s0 = __p0; \
21469  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21470  int16x4_t __ret; \
21471  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
21472  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21473  __ret; \
21474 })
21475 #endif
21476 
21477 #ifdef __LITTLE_ENDIAN__
21478 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21479  uint32x4_t __s0 = __p0; \
21480  uint16x4_t __ret; \
21481  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21482  __ret; \
21483 })
21484 #else
21485 #define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21486  uint32x4_t __s0 = __p0; \
21487  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21488  uint16x4_t __ret; \
21489  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21490  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21491  __ret; \
21492 })
21493 #define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21494  uint32x4_t __s0 = __p0; \
21495  uint16x4_t __ret; \
21496  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21497  __ret; \
21498 })
21499 #endif
21500 
21501 #ifdef __LITTLE_ENDIAN__
21502 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21503  uint64x2_t __s0 = __p0; \
21504  uint32x2_t __ret; \
21505  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21506  __ret; \
21507 })
21508 #else
21509 #define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21510  uint64x2_t __s0 = __p0; \
21511  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21512  uint32x2_t __ret; \
21513  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21515  __ret; \
21516 })
21517 #define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21518  uint64x2_t __s0 = __p0; \
21519  uint32x2_t __ret; \
21520  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21521  __ret; \
21522 })
21523 #endif
21524 
21525 #ifdef __LITTLE_ENDIAN__
21526 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21527  uint16x8_t __s0 = __p0; \
21528  uint8x8_t __ret; \
21529  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21530  __ret; \
21531 })
21532 #else
21533 #define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21534  uint16x8_t __s0 = __p0; \
21535  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21536  uint8x8_t __ret; \
21537  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21538  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21539  __ret; \
21540 })
21541 #define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21542  uint16x8_t __s0 = __p0; \
21543  uint8x8_t __ret; \
21544  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21545  __ret; \
21546 })
21547 #endif
21548 
21549 #ifdef __LITTLE_ENDIAN__
21550 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21551  int32x4_t __s0 = __p0; \
21552  int16x4_t __ret; \
21553  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21554  __ret; \
21555 })
21556 #else
21557 #define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21558  int32x4_t __s0 = __p0; \
21559  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21560  int16x4_t __ret; \
21561  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21562  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21563  __ret; \
21564 })
21565 #define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21566  int32x4_t __s0 = __p0; \
21567  int16x4_t __ret; \
21568  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21569  __ret; \
21570 })
21571 #endif
21572 
21573 #ifdef __LITTLE_ENDIAN__
21574 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21575  int64x2_t __s0 = __p0; \
21576  int32x2_t __ret; \
21577  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21578  __ret; \
21579 })
21580 #else
21581 #define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21582  int64x2_t __s0 = __p0; \
21583  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21584  int32x2_t __ret; \
21585  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21586  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21587  __ret; \
21588 })
21589 #define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21590  int64x2_t __s0 = __p0; \
21591  int32x2_t __ret; \
21592  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21593  __ret; \
21594 })
21595 #endif
21596 
21597 #ifdef __LITTLE_ENDIAN__
21598 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21599  int16x8_t __s0 = __p0; \
21600  int8x8_t __ret; \
21601  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21602  __ret; \
21603 })
21604 #else
21605 #define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21606  int16x8_t __s0 = __p0; \
21607  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21608  int8x8_t __ret; \
21609  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21611  __ret; \
21612 })
21613 #define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21614  int16x8_t __s0 = __p0; \
21615  int8x8_t __ret; \
21616  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21617  __ret; \
21618 })
21619 #endif
21620 
21621 #ifdef __LITTLE_ENDIAN__
21622 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21623  uint32x4_t __ret;
21624  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
21625  return __ret;
21626 }
21627 #else
21628 __ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21629  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21630  uint32x4_t __ret;
21631  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
21632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21633  return __ret;
21634 }
21635 #endif
21636 
21637 #ifdef __LITTLE_ENDIAN__
21638 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21639  float32x4_t __ret;
21640  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
21641  return __ret;
21642 }
21643 #else
21644 __ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21645  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21646  float32x4_t __ret;
21647  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
21648  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21649  return __ret;
21650 }
21651 #endif
21652 
21653 #ifdef __LITTLE_ENDIAN__
21654 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21655  uint32x2_t __ret;
21656  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
21657  return __ret;
21658 }
21659 #else
21660 __ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21661  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21662  uint32x2_t __ret;
21663  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
21664  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21665  return __ret;
21666 }
21667 #endif
21668 
21669 #ifdef __LITTLE_ENDIAN__
21670 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21671  float32x2_t __ret;
21672  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
21673  return __ret;
21674 }
21675 #else
21676 __ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21677  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21678  float32x2_t __ret;
21679  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
21680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21681  return __ret;
21682 }
21683 #endif
21684 
21685 #ifdef __LITTLE_ENDIAN__
21686 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21687  float32x4_t __ret;
21688  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21689  return __ret;
21690 }
21691 #else
21692 __ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21693  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21694  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21695  float32x4_t __ret;
21696  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21698  return __ret;
21699 }
21700 #endif
21701 
21702 #ifdef __LITTLE_ENDIAN__
21703 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21704  float32x2_t __ret;
21705  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21706  return __ret;
21707 }
21708 #else
21709 __ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21710  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21711  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21712  float32x2_t __ret;
21713  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21714  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21715  return __ret;
21716 }
21717 #endif
21718 
21719 #ifdef __LITTLE_ENDIAN__
21720 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21721  uint8x16_t __s0 = __p0; \
21722  uint8x16_t __s1 = __p1; \
21723  uint8x16_t __ret; \
21724  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
21725  __ret; \
21726 })
21727 #else
21728 #define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21729  uint8x16_t __s0 = __p0; \
21730  uint8x16_t __s1 = __p1; \
21731  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21732  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21733  uint8x16_t __ret; \
21734  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
21735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21736  __ret; \
21737 })
21738 #endif
21739 
21740 #ifdef __LITTLE_ENDIAN__
21741 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21742  uint32x4_t __s0 = __p0; \
21743  uint32x4_t __s1 = __p1; \
21744  uint32x4_t __ret; \
21745  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
21746  __ret; \
21747 })
21748 #else
21749 #define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21750  uint32x4_t __s0 = __p0; \
21751  uint32x4_t __s1 = __p1; \
21752  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21753  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21754  uint32x4_t __ret; \
21755  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
21756  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21757  __ret; \
21758 })
21759 #endif
21760 
21761 #ifdef __LITTLE_ENDIAN__
21762 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21763  uint64x2_t __s0 = __p0; \
21764  uint64x2_t __s1 = __p1; \
21765  uint64x2_t __ret; \
21766  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
21767  __ret; \
21768 })
21769 #else
21770 #define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21771  uint64x2_t __s0 = __p0; \
21772  uint64x2_t __s1 = __p1; \
21773  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21774  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21775  uint64x2_t __ret; \
21776  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
21777  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21778  __ret; \
21779 })
21780 #endif
21781 
21782 #ifdef __LITTLE_ENDIAN__
21783 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21784  uint16x8_t __s0 = __p0; \
21785  uint16x8_t __s1 = __p1; \
21786  uint16x8_t __ret; \
21787  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
21788  __ret; \
21789 })
21790 #else
21791 #define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21792  uint16x8_t __s0 = __p0; \
21793  uint16x8_t __s1 = __p1; \
21794  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21795  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21796  uint16x8_t __ret; \
21797  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
21798  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21799  __ret; \
21800 })
21801 #endif
21802 
21803 #ifdef __LITTLE_ENDIAN__
21804 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21805  int8x16_t __s0 = __p0; \
21806  int8x16_t __s1 = __p1; \
21807  int8x16_t __ret; \
21808  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
21809  __ret; \
21810 })
21811 #else
21812 #define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21813  int8x16_t __s0 = __p0; \
21814  int8x16_t __s1 = __p1; \
21815  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21816  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21817  int8x16_t __ret; \
21818  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
21819  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21820  __ret; \
21821 })
21822 #endif
21823 
21824 #ifdef __LITTLE_ENDIAN__
21825 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21826  int32x4_t __s0 = __p0; \
21827  int32x4_t __s1 = __p1; \
21828  int32x4_t __ret; \
21829  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
21830  __ret; \
21831 })
21832 #else
21833 #define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21834  int32x4_t __s0 = __p0; \
21835  int32x4_t __s1 = __p1; \
21836  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21837  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21838  int32x4_t __ret; \
21839  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
21840  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21841  __ret; \
21842 })
21843 #endif
21844 
21845 #ifdef __LITTLE_ENDIAN__
21846 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21847  int64x2_t __s0 = __p0; \
21848  int64x2_t __s1 = __p1; \
21849  int64x2_t __ret; \
21850  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
21851  __ret; \
21852 })
21853 #else
21854 #define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21855  int64x2_t __s0 = __p0; \
21856  int64x2_t __s1 = __p1; \
21857  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21858  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21859  int64x2_t __ret; \
21860  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
21861  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21862  __ret; \
21863 })
21864 #endif
21865 
21866 #ifdef __LITTLE_ENDIAN__
21867 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21868  int16x8_t __s0 = __p0; \
21869  int16x8_t __s1 = __p1; \
21870  int16x8_t __ret; \
21871  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
21872  __ret; \
21873 })
21874 #else
21875 #define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21876  int16x8_t __s0 = __p0; \
21877  int16x8_t __s1 = __p1; \
21878  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21879  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21880  int16x8_t __ret; \
21881  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
21882  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21883  __ret; \
21884 })
21885 #endif
21886 
21887 #ifdef __LITTLE_ENDIAN__
21888 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21889  uint8x8_t __s0 = __p0; \
21890  uint8x8_t __s1 = __p1; \
21891  uint8x8_t __ret; \
21892  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
21893  __ret; \
21894 })
21895 #else
21896 #define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21897  uint8x8_t __s0 = __p0; \
21898  uint8x8_t __s1 = __p1; \
21899  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21900  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21901  uint8x8_t __ret; \
21902  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
21903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21904  __ret; \
21905 })
21906 #endif
21907 
21908 #ifdef __LITTLE_ENDIAN__
21909 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21910  uint32x2_t __s0 = __p0; \
21911  uint32x2_t __s1 = __p1; \
21912  uint32x2_t __ret; \
21913  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
21914  __ret; \
21915 })
21916 #else
21917 #define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21918  uint32x2_t __s0 = __p0; \
21919  uint32x2_t __s1 = __p1; \
21920  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21921  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21922  uint32x2_t __ret; \
21923  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
21924  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21925  __ret; \
21926 })
21927 #endif
21928 
21929 #ifdef __LITTLE_ENDIAN__
21930 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21931  uint64x1_t __s0 = __p0; \
21932  uint64x1_t __s1 = __p1; \
21933  uint64x1_t __ret; \
21934  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21935  __ret; \
21936 })
21937 #else
21938 #define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21939  uint64x1_t __s0 = __p0; \
21940  uint64x1_t __s1 = __p1; \
21941  uint64x1_t __ret; \
21942  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21943  __ret; \
21944 })
21945 #endif
21946 
21947 #ifdef __LITTLE_ENDIAN__
21948 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21949  uint16x4_t __s0 = __p0; \
21950  uint16x4_t __s1 = __p1; \
21951  uint16x4_t __ret; \
21952  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
21953  __ret; \
21954 })
21955 #else
21956 #define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21957  uint16x4_t __s0 = __p0; \
21958  uint16x4_t __s1 = __p1; \
21959  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21960  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21961  uint16x4_t __ret; \
21962  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
21963  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21964  __ret; \
21965 })
21966 #endif
21967 
21968 #ifdef __LITTLE_ENDIAN__
21969 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21970  int8x8_t __s0 = __p0; \
21971  int8x8_t __s1 = __p1; \
21972  int8x8_t __ret; \
21973  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
21974  __ret; \
21975 })
21976 #else
21977 #define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21978  int8x8_t __s0 = __p0; \
21979  int8x8_t __s1 = __p1; \
21980  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21981  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21982  int8x8_t __ret; \
21983  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
21984  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21985  __ret; \
21986 })
21987 #endif
21988 
21989 #ifdef __LITTLE_ENDIAN__
21990 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21991  int32x2_t __s0 = __p0; \
21992  int32x2_t __s1 = __p1; \
21993  int32x2_t __ret; \
21994  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
21995  __ret; \
21996 })
21997 #else
21998 #define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21999  int32x2_t __s0 = __p0; \
22000  int32x2_t __s1 = __p1; \
22001  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22002  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22003  int32x2_t __ret; \
22004  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
22005  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22006  __ret; \
22007 })
22008 #endif
22009 
22010 #ifdef __LITTLE_ENDIAN__
22011 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
22012  int64x1_t __s0 = __p0; \
22013  int64x1_t __s1 = __p1; \
22014  int64x1_t __ret; \
22015  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
22016  __ret; \
22017 })
22018 #else
22019 #define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
22020  int64x1_t __s0 = __p0; \
22021  int64x1_t __s1 = __p1; \
22022  int64x1_t __ret; \
22023  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
22024  __ret; \
22025 })
22026 #endif
22027 
22028 #ifdef __LITTLE_ENDIAN__
22029 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
22030  int16x4_t __s0 = __p0; \
22031  int16x4_t __s1 = __p1; \
22032  int16x4_t __ret; \
22033  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
22034  __ret; \
22035 })
22036 #else
22037 #define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
22038  int16x4_t __s0 = __p0; \
22039  int16x4_t __s1 = __p1; \
22040  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22041  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22042  int16x4_t __ret; \
22043  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
22044  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22045  __ret; \
22046 })
22047 #endif
22048 
22049 #ifdef __LITTLE_ENDIAN__
22050 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22051  uint16x4_t __ret;
22052  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22053  return __ret;
22054 }
22055 #else
22056 __ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22057  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22058  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22059  uint16x4_t __ret;
22060  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
22061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22062  return __ret;
22063 }
22064 __ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22065  uint16x4_t __ret;
22066  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22067  return __ret;
22068 }
22069 #endif
22070 
22071 #ifdef __LITTLE_ENDIAN__
22072 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22073  uint32x2_t __ret;
22074  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22075  return __ret;
22076 }
22077 #else
22078 __ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22079  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22080  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22081  uint32x2_t __ret;
22082  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
22083  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22084  return __ret;
22085 }
22086 __ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22087  uint32x2_t __ret;
22088  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22089  return __ret;
22090 }
22091 #endif
22092 
22093 #ifdef __LITTLE_ENDIAN__
22094 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22095  uint8x8_t __ret;
22096  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22097  return __ret;
22098 }
22099 #else
22100 __ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22101  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22102  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22103  uint8x8_t __ret;
22104  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
22105  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22106  return __ret;
22107 }
22108 __ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22109  uint8x8_t __ret;
22110  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22111  return __ret;
22112 }
22113 #endif
22114 
22115 #ifdef __LITTLE_ENDIAN__
22116 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22117  int16x4_t __ret;
22118  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22119  return __ret;
22120 }
22121 #else
22122 __ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22123  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22124  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22125  int16x4_t __ret;
22126  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
22127  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22128  return __ret;
22129 }
22130 __ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22131  int16x4_t __ret;
22132  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22133  return __ret;
22134 }
22135 #endif
22136 
22137 #ifdef __LITTLE_ENDIAN__
22138 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22139  int32x2_t __ret;
22140  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22141  return __ret;
22142 }
22143 #else
22144 __ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22145  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22146  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22147  int32x2_t __ret;
22148  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
22149  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22150  return __ret;
22151 }
22152 __ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22153  int32x2_t __ret;
22154  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22155  return __ret;
22156 }
22157 #endif
22158 
22159 #ifdef __LITTLE_ENDIAN__
22160 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22161  int8x8_t __ret;
22162  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22163  return __ret;
22164 }
22165 #else
22166 __ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22167  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22168  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22169  int8x8_t __ret;
22170  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
22171  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22172  return __ret;
22173 }
22174 __ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22175  int8x8_t __ret;
22176  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22177  return __ret;
22178 }
22179 #endif
22180 
22181 #ifdef __LITTLE_ENDIAN__
22182 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22183  poly8_t __s0 = __p0; \
22184  poly8x8_t __s1 = __p1; \
22185  poly8x8_t __ret; \
22186  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22187  __ret; \
22188 })
22189 #else
22190 #define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22191  poly8_t __s0 = __p0; \
22192  poly8x8_t __s1 = __p1; \
22193  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22194  poly8x8_t __ret; \
22195  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22196  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22197  __ret; \
22198 })
22199 #define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22200  poly8_t __s0 = __p0; \
22201  poly8x8_t __s1 = __p1; \
22202  poly8x8_t __ret; \
22203  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22204  __ret; \
22205 })
22206 #endif
22207 
22208 #ifdef __LITTLE_ENDIAN__
22209 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22210  poly16_t __s0 = __p0; \
22211  poly16x4_t __s1 = __p1; \
22212  poly16x4_t __ret; \
22213  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22214  __ret; \
22215 })
22216 #else
22217 #define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22218  poly16_t __s0 = __p0; \
22219  poly16x4_t __s1 = __p1; \
22220  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22221  poly16x4_t __ret; \
22222  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22223  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22224  __ret; \
22225 })
22226 #define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22227  poly16_t __s0 = __p0; \
22228  poly16x4_t __s1 = __p1; \
22229  poly16x4_t __ret; \
22230  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22231  __ret; \
22232 })
22233 #endif
22234 
22235 #ifdef __LITTLE_ENDIAN__
22236 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22237  poly8_t __s0 = __p0; \
22238  poly8x16_t __s1 = __p1; \
22239  poly8x16_t __ret; \
22240  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22241  __ret; \
22242 })
22243 #else
22244 #define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22245  poly8_t __s0 = __p0; \
22246  poly8x16_t __s1 = __p1; \
22247  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22248  poly8x16_t __ret; \
22249  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22250  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22251  __ret; \
22252 })
22253 #define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22254  poly8_t __s0 = __p0; \
22255  poly8x16_t __s1 = __p1; \
22256  poly8x16_t __ret; \
22257  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22258  __ret; \
22259 })
22260 #endif
22261 
22262 #ifdef __LITTLE_ENDIAN__
22263 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22264  poly16_t __s0 = __p0; \
22265  poly16x8_t __s1 = __p1; \
22266  poly16x8_t __ret; \
22267  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22268  __ret; \
22269 })
22270 #else
22271 #define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22272  poly16_t __s0 = __p0; \
22273  poly16x8_t __s1 = __p1; \
22274  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22275  poly16x8_t __ret; \
22276  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22277  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22278  __ret; \
22279 })
22280 #define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22281  poly16_t __s0 = __p0; \
22282  poly16x8_t __s1 = __p1; \
22283  poly16x8_t __ret; \
22284  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22285  __ret; \
22286 })
22287 #endif
22288 
22289 #ifdef __LITTLE_ENDIAN__
22290 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22291  uint8_t __s0 = __p0; \
22292  uint8x16_t __s1 = __p1; \
22293  uint8x16_t __ret; \
22294  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22295  __ret; \
22296 })
22297 #else
22298 #define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22299  uint8_t __s0 = __p0; \
22300  uint8x16_t __s1 = __p1; \
22301  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22302  uint8x16_t __ret; \
22303  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22304  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22305  __ret; \
22306 })
22307 #define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22308  uint8_t __s0 = __p0; \
22309  uint8x16_t __s1 = __p1; \
22310  uint8x16_t __ret; \
22311  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22312  __ret; \
22313 })
22314 #endif
22315 
22316 #ifdef __LITTLE_ENDIAN__
22317 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22318  uint32_t __s0 = __p0; \
22319  uint32x4_t __s1 = __p1; \
22320  uint32x4_t __ret; \
22321  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22322  __ret; \
22323 })
22324 #else
22325 #define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22326  uint32_t __s0 = __p0; \
22327  uint32x4_t __s1 = __p1; \
22328  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22329  uint32x4_t __ret; \
22330  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22332  __ret; \
22333 })
22334 #define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22335  uint32_t __s0 = __p0; \
22336  uint32x4_t __s1 = __p1; \
22337  uint32x4_t __ret; \
22338  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22339  __ret; \
22340 })
22341 #endif
22342 
22343 #ifdef __LITTLE_ENDIAN__
22344 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22345  uint64_t __s0 = __p0; \
22346  uint64x2_t __s1 = __p1; \
22347  uint64x2_t __ret; \
22348  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22349  __ret; \
22350 })
22351 #else
22352 #define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22353  uint64_t __s0 = __p0; \
22354  uint64x2_t __s1 = __p1; \
22355  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22356  uint64x2_t __ret; \
22357  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22359  __ret; \
22360 })
22361 #define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22362  uint64_t __s0 = __p0; \
22363  uint64x2_t __s1 = __p1; \
22364  uint64x2_t __ret; \
22365  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22366  __ret; \
22367 })
22368 #endif
22369 
22370 #ifdef __LITTLE_ENDIAN__
22371 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22372  uint16_t __s0 = __p0; \
22373  uint16x8_t __s1 = __p1; \
22374  uint16x8_t __ret; \
22375  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22376  __ret; \
22377 })
22378 #else
22379 #define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22380  uint16_t __s0 = __p0; \
22381  uint16x8_t __s1 = __p1; \
22382  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22383  uint16x8_t __ret; \
22384  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22385  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22386  __ret; \
22387 })
22388 #define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22389  uint16_t __s0 = __p0; \
22390  uint16x8_t __s1 = __p1; \
22391  uint16x8_t __ret; \
22392  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22393  __ret; \
22394 })
22395 #endif
22396 
22397 #ifdef __LITTLE_ENDIAN__
22398 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22399  int8_t __s0 = __p0; \
22400  int8x16_t __s1 = __p1; \
22401  int8x16_t __ret; \
22402  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22403  __ret; \
22404 })
22405 #else
22406 #define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22407  int8_t __s0 = __p0; \
22408  int8x16_t __s1 = __p1; \
22409  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22410  int8x16_t __ret; \
22411  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22412  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22413  __ret; \
22414 })
22415 #define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22416  int8_t __s0 = __p0; \
22417  int8x16_t __s1 = __p1; \
22418  int8x16_t __ret; \
22419  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22420  __ret; \
22421 })
22422 #endif
22423 
22424 #ifdef __LITTLE_ENDIAN__
22425 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22426  float32_t __s0 = __p0; \
22427  float32x4_t __s1 = __p1; \
22428  float32x4_t __ret; \
22429  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22430  __ret; \
22431 })
22432 #else
22433 #define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22434  float32_t __s0 = __p0; \
22435  float32x4_t __s1 = __p1; \
22436  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22437  float32x4_t __ret; \
22438  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__rev1, __p2); \
22439  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22440  __ret; \
22441 })
22442 #define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22443  float32_t __s0 = __p0; \
22444  float32x4_t __s1 = __p1; \
22445  float32x4_t __ret; \
22446  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22447  __ret; \
22448 })
22449 #endif
22450 
22451 #ifdef __LITTLE_ENDIAN__
22452 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22453  int32_t __s0 = __p0; \
22454  int32x4_t __s1 = __p1; \
22455  int32x4_t __ret; \
22456  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22457  __ret; \
22458 })
22459 #else
22460 #define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22461  int32_t __s0 = __p0; \
22462  int32x4_t __s1 = __p1; \
22463  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22464  int32x4_t __ret; \
22465  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22467  __ret; \
22468 })
22469 #define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22470  int32_t __s0 = __p0; \
22471  int32x4_t __s1 = __p1; \
22472  int32x4_t __ret; \
22473  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22474  __ret; \
22475 })
22476 #endif
22477 
22478 #ifdef __LITTLE_ENDIAN__
22479 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22480  int64_t __s0 = __p0; \
22481  int64x2_t __s1 = __p1; \
22482  int64x2_t __ret; \
22483  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22484  __ret; \
22485 })
22486 #else
22487 #define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22488  int64_t __s0 = __p0; \
22489  int64x2_t __s1 = __p1; \
22490  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22491  int64x2_t __ret; \
22492  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22493  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22494  __ret; \
22495 })
22496 #define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22497  int64_t __s0 = __p0; \
22498  int64x2_t __s1 = __p1; \
22499  int64x2_t __ret; \
22500  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22501  __ret; \
22502 })
22503 #endif
22504 
22505 #ifdef __LITTLE_ENDIAN__
22506 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22507  int16_t __s0 = __p0; \
22508  int16x8_t __s1 = __p1; \
22509  int16x8_t __ret; \
22510  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22511  __ret; \
22512 })
22513 #else
22514 #define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22515  int16_t __s0 = __p0; \
22516  int16x8_t __s1 = __p1; \
22517  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22518  int16x8_t __ret; \
22519  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22520  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22521  __ret; \
22522 })
22523 #define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22524  int16_t __s0 = __p0; \
22525  int16x8_t __s1 = __p1; \
22526  int16x8_t __ret; \
22527  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22528  __ret; \
22529 })
22530 #endif
22531 
22532 #ifdef __LITTLE_ENDIAN__
22533 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22534  uint8_t __s0 = __p0; \
22535  uint8x8_t __s1 = __p1; \
22536  uint8x8_t __ret; \
22537  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22538  __ret; \
22539 })
22540 #else
22541 #define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22542  uint8_t __s0 = __p0; \
22543  uint8x8_t __s1 = __p1; \
22544  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22545  uint8x8_t __ret; \
22546  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22547  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22548  __ret; \
22549 })
22550 #define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22551  uint8_t __s0 = __p0; \
22552  uint8x8_t __s1 = __p1; \
22553  uint8x8_t __ret; \
22554  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22555  __ret; \
22556 })
22557 #endif
22558 
22559 #ifdef __LITTLE_ENDIAN__
22560 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22561  uint32_t __s0 = __p0; \
22562  uint32x2_t __s1 = __p1; \
22563  uint32x2_t __ret; \
22564  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22565  __ret; \
22566 })
22567 #else
22568 #define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22569  uint32_t __s0 = __p0; \
22570  uint32x2_t __s1 = __p1; \
22571  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22572  uint32x2_t __ret; \
22573  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22574  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22575  __ret; \
22576 })
22577 #define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22578  uint32_t __s0 = __p0; \
22579  uint32x2_t __s1 = __p1; \
22580  uint32x2_t __ret; \
22581  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22582  __ret; \
22583 })
22584 #endif
22585 
22586 #ifdef __LITTLE_ENDIAN__
22587 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22588  uint64_t __s0 = __p0; \
22589  uint64x1_t __s1 = __p1; \
22590  uint64x1_t __ret; \
22591  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22592  __ret; \
22593 })
22594 #else
22595 #define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22596  uint64_t __s0 = __p0; \
22597  uint64x1_t __s1 = __p1; \
22598  uint64x1_t __ret; \
22599  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22600  __ret; \
22601 })
22602 #define __noswap_vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22603  uint64_t __s0 = __p0; \
22604  uint64x1_t __s1 = __p1; \
22605  uint64x1_t __ret; \
22606  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22607  __ret; \
22608 })
22609 #endif
22610 
22611 #ifdef __LITTLE_ENDIAN__
22612 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22613  uint16_t __s0 = __p0; \
22614  uint16x4_t __s1 = __p1; \
22615  uint16x4_t __ret; \
22616  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22617  __ret; \
22618 })
22619 #else
22620 #define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22621  uint16_t __s0 = __p0; \
22622  uint16x4_t __s1 = __p1; \
22623  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22624  uint16x4_t __ret; \
22625  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22627  __ret; \
22628 })
22629 #define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22630  uint16_t __s0 = __p0; \
22631  uint16x4_t __s1 = __p1; \
22632  uint16x4_t __ret; \
22633  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22634  __ret; \
22635 })
22636 #endif
22637 
22638 #ifdef __LITTLE_ENDIAN__
22639 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22640  int8_t __s0 = __p0; \
22641  int8x8_t __s1 = __p1; \
22642  int8x8_t __ret; \
22643  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22644  __ret; \
22645 })
22646 #else
22647 #define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22648  int8_t __s0 = __p0; \
22649  int8x8_t __s1 = __p1; \
22650  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22651  int8x8_t __ret; \
22652  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22654  __ret; \
22655 })
22656 #define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22657  int8_t __s0 = __p0; \
22658  int8x8_t __s1 = __p1; \
22659  int8x8_t __ret; \
22660  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22661  __ret; \
22662 })
22663 #endif
22664 
22665 #ifdef __LITTLE_ENDIAN__
22666 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22667  float32_t __s0 = __p0; \
22668  float32x2_t __s1 = __p1; \
22669  float32x2_t __ret; \
22670  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22671  __ret; \
22672 })
22673 #else
22674 #define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22675  float32_t __s0 = __p0; \
22676  float32x2_t __s1 = __p1; \
22677  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22678  float32x2_t __ret; \
22679  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__rev1, __p2); \
22680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22681  __ret; \
22682 })
22683 #define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22684  float32_t __s0 = __p0; \
22685  float32x2_t __s1 = __p1; \
22686  float32x2_t __ret; \
22687  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22688  __ret; \
22689 })
22690 #endif
22691 
22692 #ifdef __LITTLE_ENDIAN__
22693 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22694  int32_t __s0 = __p0; \
22695  int32x2_t __s1 = __p1; \
22696  int32x2_t __ret; \
22697  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22698  __ret; \
22699 })
22700 #else
22701 #define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22702  int32_t __s0 = __p0; \
22703  int32x2_t __s1 = __p1; \
22704  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22705  int32x2_t __ret; \
22706  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22707  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22708  __ret; \
22709 })
22710 #define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22711  int32_t __s0 = __p0; \
22712  int32x2_t __s1 = __p1; \
22713  int32x2_t __ret; \
22714  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22715  __ret; \
22716 })
22717 #endif
22718 
22719 #ifdef __LITTLE_ENDIAN__
22720 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22721  int64_t __s0 = __p0; \
22722  int64x1_t __s1 = __p1; \
22723  int64x1_t __ret; \
22724  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22725  __ret; \
22726 })
22727 #else
22728 #define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22729  int64_t __s0 = __p0; \
22730  int64x1_t __s1 = __p1; \
22731  int64x1_t __ret; \
22732  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22733  __ret; \
22734 })
22735 #define __noswap_vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22736  int64_t __s0 = __p0; \
22737  int64x1_t __s1 = __p1; \
22738  int64x1_t __ret; \
22739  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22740  __ret; \
22741 })
22742 #endif
22743 
22744 #ifdef __LITTLE_ENDIAN__
22745 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22746  int16_t __s0 = __p0; \
22747  int16x4_t __s1 = __p1; \
22748  int16x4_t __ret; \
22749  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22750  __ret; \
22751 })
22752 #else
22753 #define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22754  int16_t __s0 = __p0; \
22755  int16x4_t __s1 = __p1; \
22756  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22757  int16x4_t __ret; \
22758  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22759  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22760  __ret; \
22761 })
22762 #define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22763  int16_t __s0 = __p0; \
22764  int16x4_t __s1 = __p1; \
22765  int16x4_t __ret; \
22766  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22767  __ret; \
22768 })
22769 #endif
22770 
22771 #ifdef __LITTLE_ENDIAN__
22772 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22773  uint8x16_t __ret;
22774  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22775  return __ret;
22776 }
22777 #else
22778 __ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22779  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22780  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22781  uint8x16_t __ret;
22782  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22783  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22784  return __ret;
22785 }
22786 #endif
22787 
22788 #ifdef __LITTLE_ENDIAN__
22789 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22790  uint32x4_t __ret;
22791  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22792  return __ret;
22793 }
22794 #else
22795 __ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22796  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22797  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22798  uint32x4_t __ret;
22799  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22800  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22801  return __ret;
22802 }
22803 #endif
22804 
22805 #ifdef __LITTLE_ENDIAN__
22806 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22807  uint64x2_t __ret;
22808  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22809  return __ret;
22810 }
22811 #else
22812 __ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22813  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22814  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22815  uint64x2_t __ret;
22816  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22818  return __ret;
22819 }
22820 #endif
22821 
22822 #ifdef __LITTLE_ENDIAN__
22823 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22824  uint16x8_t __ret;
22825  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22826  return __ret;
22827 }
22828 #else
22829 __ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22830  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22831  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22832  uint16x8_t __ret;
22833  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22834  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22835  return __ret;
22836 }
22837 #endif
22838 
22839 #ifdef __LITTLE_ENDIAN__
22840 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22841  int8x16_t __ret;
22842  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22843  return __ret;
22844 }
22845 #else
22846 __ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22847  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22848  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22849  int8x16_t __ret;
22850  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22851  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22852  return __ret;
22853 }
22854 #endif
22855 
22856 #ifdef __LITTLE_ENDIAN__
22857 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22858  int32x4_t __ret;
22859  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22860  return __ret;
22861 }
22862 #else
22863 __ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22864  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22865  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22866  int32x4_t __ret;
22867  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22869  return __ret;
22870 }
22871 #endif
22872 
22873 #ifdef __LITTLE_ENDIAN__
22874 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22875  int64x2_t __ret;
22876  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22877  return __ret;
22878 }
22879 #else
22880 __ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22881  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22882  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22883  int64x2_t __ret;
22884  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22886  return __ret;
22887 }
22888 #endif
22889 
22890 #ifdef __LITTLE_ENDIAN__
22891 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22892  int16x8_t __ret;
22893  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22894  return __ret;
22895 }
22896 #else
22897 __ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22898  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22899  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22900  int16x8_t __ret;
22901  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22903  return __ret;
22904 }
22905 #endif
22906 
22907 #ifdef __LITTLE_ENDIAN__
22908 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22909  uint8x8_t __ret;
22910  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22911  return __ret;
22912 }
22913 #else
22914 __ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22915  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22916  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22917  uint8x8_t __ret;
22918  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22919  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22920  return __ret;
22921 }
22922 #endif
22923 
22924 #ifdef __LITTLE_ENDIAN__
22925 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22926  uint32x2_t __ret;
22927  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22928  return __ret;
22929 }
22930 #else
22931 __ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22932  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22933  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22934  uint32x2_t __ret;
22935  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22936  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22937  return __ret;
22938 }
22939 #endif
22940 
22941 #ifdef __LITTLE_ENDIAN__
22942 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22943  uint64x1_t __ret;
22944  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22945  return __ret;
22946 }
22947 #else
22948 __ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22949  uint64x1_t __ret;
22950  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22951  return __ret;
22952 }
22953 #endif
22954 
22955 #ifdef __LITTLE_ENDIAN__
22956 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22957  uint16x4_t __ret;
22958  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22959  return __ret;
22960 }
22961 #else
22962 __ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22963  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22964  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22965  uint16x4_t __ret;
22966  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22967  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22968  return __ret;
22969 }
22970 #endif
22971 
22972 #ifdef __LITTLE_ENDIAN__
22973 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22974  int8x8_t __ret;
22975  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22976  return __ret;
22977 }
22978 #else
22979 __ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22980  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22981  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22982  int8x8_t __ret;
22983  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22984  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22985  return __ret;
22986 }
22987 #endif
22988 
22989 #ifdef __LITTLE_ENDIAN__
22990 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22991  int32x2_t __ret;
22992  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22993  return __ret;
22994 }
22995 #else
22996 __ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22997  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22998  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22999  int32x2_t __ret;
23000  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
23001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23002  return __ret;
23003 }
23004 #endif
23005 
23006 #ifdef __LITTLE_ENDIAN__
23007 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
23008  int64x1_t __ret;
23009  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
23010  return __ret;
23011 }
23012 #else
23013 __ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
23014  int64x1_t __ret;
23015  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
23016  return __ret;
23017 }
23018 #endif
23019 
23020 #ifdef __LITTLE_ENDIAN__
23021 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
23022  int16x4_t __ret;
23023  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
23024  return __ret;
23025 }
23026 #else
23027 __ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
23028  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23029  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23030  int16x4_t __ret;
23031  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
23032  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23033  return __ret;
23034 }
23035 #endif
23036 
23037 #ifdef __LITTLE_ENDIAN__
23038 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
23039  uint8x16_t __s0 = __p0; \
23040  uint8x16_t __ret; \
23041  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
23042  __ret; \
23043 })
23044 #else
23045 #define vshlq_n_u8(__p0, __p1) __extension__ ({ \
23046  uint8x16_t __s0 = __p0; \
23047  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23048  uint8x16_t __ret; \
23049  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
23050  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23051  __ret; \
23052 })
23053 #endif
23054 
23055 #ifdef __LITTLE_ENDIAN__
23056 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23057  uint32x4_t __s0 = __p0; \
23058  uint32x4_t __ret; \
23059  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
23060  __ret; \
23061 })
23062 #else
23063 #define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23064  uint32x4_t __s0 = __p0; \
23065  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23066  uint32x4_t __ret; \
23067  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
23068  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23069  __ret; \
23070 })
23071 #endif
23072 
23073 #ifdef __LITTLE_ENDIAN__
23074 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23075  uint64x2_t __s0 = __p0; \
23076  uint64x2_t __ret; \
23077  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
23078  __ret; \
23079 })
23080 #else
23081 #define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23082  uint64x2_t __s0 = __p0; \
23083  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23084  uint64x2_t __ret; \
23085  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
23086  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23087  __ret; \
23088 })
23089 #endif
23090 
23091 #ifdef __LITTLE_ENDIAN__
23092 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23093  uint16x8_t __s0 = __p0; \
23094  uint16x8_t __ret; \
23095  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
23096  __ret; \
23097 })
23098 #else
23099 #define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23100  uint16x8_t __s0 = __p0; \
23101  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23102  uint16x8_t __ret; \
23103  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
23104  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23105  __ret; \
23106 })
23107 #endif
23108 
23109 #ifdef __LITTLE_ENDIAN__
23110 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23111  int8x16_t __s0 = __p0; \
23112  int8x16_t __ret; \
23113  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
23114  __ret; \
23115 })
23116 #else
23117 #define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23118  int8x16_t __s0 = __p0; \
23119  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23120  int8x16_t __ret; \
23121  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
23122  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23123  __ret; \
23124 })
23125 #endif
23126 
23127 #ifdef __LITTLE_ENDIAN__
23128 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23129  int32x4_t __s0 = __p0; \
23130  int32x4_t __ret; \
23131  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
23132  __ret; \
23133 })
23134 #else
23135 #define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23136  int32x4_t __s0 = __p0; \
23137  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23138  int32x4_t __ret; \
23139  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
23140  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23141  __ret; \
23142 })
23143 #endif
23144 
23145 #ifdef __LITTLE_ENDIAN__
23146 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23147  int64x2_t __s0 = __p0; \
23148  int64x2_t __ret; \
23149  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
23150  __ret; \
23151 })
23152 #else
23153 #define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23154  int64x2_t __s0 = __p0; \
23155  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23156  int64x2_t __ret; \
23157  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
23158  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23159  __ret; \
23160 })
23161 #endif
23162 
23163 #ifdef __LITTLE_ENDIAN__
23164 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23165  int16x8_t __s0 = __p0; \
23166  int16x8_t __ret; \
23167  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
23168  __ret; \
23169 })
23170 #else
23171 #define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23172  int16x8_t __s0 = __p0; \
23173  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23174  int16x8_t __ret; \
23175  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
23176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23177  __ret; \
23178 })
23179 #endif
23180 
23181 #ifdef __LITTLE_ENDIAN__
23182 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
23183  uint8x8_t __s0 = __p0; \
23184  uint8x8_t __ret; \
23185  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
23186  __ret; \
23187 })
23188 #else
23189 #define vshl_n_u8(__p0, __p1) __extension__ ({ \
23190  uint8x8_t __s0 = __p0; \
23191  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23192  uint8x8_t __ret; \
23193  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
23194  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23195  __ret; \
23196 })
23197 #endif
23198 
23199 #ifdef __LITTLE_ENDIAN__
23200 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
23201  uint32x2_t __s0 = __p0; \
23202  uint32x2_t __ret; \
23203  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
23204  __ret; \
23205 })
23206 #else
23207 #define vshl_n_u32(__p0, __p1) __extension__ ({ \
23208  uint32x2_t __s0 = __p0; \
23209  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23210  uint32x2_t __ret; \
23211  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
23212  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23213  __ret; \
23214 })
23215 #endif
23216 
23217 #ifdef __LITTLE_ENDIAN__
23218 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
23219  uint64x1_t __s0 = __p0; \
23220  uint64x1_t __ret; \
23221  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23222  __ret; \
23223 })
23224 #else
23225 #define vshl_n_u64(__p0, __p1) __extension__ ({ \
23226  uint64x1_t __s0 = __p0; \
23227  uint64x1_t __ret; \
23228  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23229  __ret; \
23230 })
23231 #endif
23232 
23233 #ifdef __LITTLE_ENDIAN__
23234 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
23235  uint16x4_t __s0 = __p0; \
23236  uint16x4_t __ret; \
23237  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
23238  __ret; \
23239 })
23240 #else
23241 #define vshl_n_u16(__p0, __p1) __extension__ ({ \
23242  uint16x4_t __s0 = __p0; \
23243  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23244  uint16x4_t __ret; \
23245  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
23246  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23247  __ret; \
23248 })
23249 #endif
23250 
23251 #ifdef __LITTLE_ENDIAN__
23252 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
23253  int8x8_t __s0 = __p0; \
23254  int8x8_t __ret; \
23255  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
23256  __ret; \
23257 })
23258 #else
23259 #define vshl_n_s8(__p0, __p1) __extension__ ({ \
23260  int8x8_t __s0 = __p0; \
23261  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23262  int8x8_t __ret; \
23263  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
23264  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23265  __ret; \
23266 })
23267 #endif
23268 
23269 #ifdef __LITTLE_ENDIAN__
23270 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
23271  int32x2_t __s0 = __p0; \
23272  int32x2_t __ret; \
23273  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
23274  __ret; \
23275 })
23276 #else
23277 #define vshl_n_s32(__p0, __p1) __extension__ ({ \
23278  int32x2_t __s0 = __p0; \
23279  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23280  int32x2_t __ret; \
23281  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
23282  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23283  __ret; \
23284 })
23285 #endif
23286 
23287 #ifdef __LITTLE_ENDIAN__
23288 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
23289  int64x1_t __s0 = __p0; \
23290  int64x1_t __ret; \
23291  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23292  __ret; \
23293 })
23294 #else
23295 #define vshl_n_s64(__p0, __p1) __extension__ ({ \
23296  int64x1_t __s0 = __p0; \
23297  int64x1_t __ret; \
23298  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23299  __ret; \
23300 })
23301 #endif
23302 
23303 #ifdef __LITTLE_ENDIAN__
23304 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
23305  int16x4_t __s0 = __p0; \
23306  int16x4_t __ret; \
23307  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
23308  __ret; \
23309 })
23310 #else
23311 #define vshl_n_s16(__p0, __p1) __extension__ ({ \
23312  int16x4_t __s0 = __p0; \
23313  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23314  int16x4_t __ret; \
23315  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
23316  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23317  __ret; \
23318 })
23319 #endif
23320 
23321 #ifdef __LITTLE_ENDIAN__
23322 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
23323  uint8x8_t __s0 = __p0; \
23324  uint16x8_t __ret; \
23325  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23326  __ret; \
23327 })
23328 #else
23329 #define vshll_n_u8(__p0, __p1) __extension__ ({ \
23330  uint8x8_t __s0 = __p0; \
23331  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23332  uint16x8_t __ret; \
23333  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
23334  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23335  __ret; \
23336 })
23337 #define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
23338  uint8x8_t __s0 = __p0; \
23339  uint16x8_t __ret; \
23340  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23341  __ret; \
23342 })
23343 #endif
23344 
23345 #ifdef __LITTLE_ENDIAN__
23346 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
23347  uint32x2_t __s0 = __p0; \
23348  uint64x2_t __ret; \
23349  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23350  __ret; \
23351 })
23352 #else
23353 #define vshll_n_u32(__p0, __p1) __extension__ ({ \
23354  uint32x2_t __s0 = __p0; \
23355  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23356  uint64x2_t __ret; \
23357  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
23358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23359  __ret; \
23360 })
23361 #define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
23362  uint32x2_t __s0 = __p0; \
23363  uint64x2_t __ret; \
23364  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23365  __ret; \
23366 })
23367 #endif
23368 
23369 #ifdef __LITTLE_ENDIAN__
23370 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
23371  uint16x4_t __s0 = __p0; \
23372  uint32x4_t __ret; \
23373  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23374  __ret; \
23375 })
23376 #else
23377 #define vshll_n_u16(__p0, __p1) __extension__ ({ \
23378  uint16x4_t __s0 = __p0; \
23379  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23380  uint32x4_t __ret; \
23381  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
23382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23383  __ret; \
23384 })
23385 #define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
23386  uint16x4_t __s0 = __p0; \
23387  uint32x4_t __ret; \
23388  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23389  __ret; \
23390 })
23391 #endif
23392 
23393 #ifdef __LITTLE_ENDIAN__
23394 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
23395  int8x8_t __s0 = __p0; \
23396  int16x8_t __ret; \
23397  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23398  __ret; \
23399 })
23400 #else
23401 #define vshll_n_s8(__p0, __p1) __extension__ ({ \
23402  int8x8_t __s0 = __p0; \
23403  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23404  int16x8_t __ret; \
23405  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
23406  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23407  __ret; \
23408 })
23409 #define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
23410  int8x8_t __s0 = __p0; \
23411  int16x8_t __ret; \
23412  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23413  __ret; \
23414 })
23415 #endif
23416 
23417 #ifdef __LITTLE_ENDIAN__
23418 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
23419  int32x2_t __s0 = __p0; \
23420  int64x2_t __ret; \
23421  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23422  __ret; \
23423 })
23424 #else
23425 #define vshll_n_s32(__p0, __p1) __extension__ ({ \
23426  int32x2_t __s0 = __p0; \
23427  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23428  int64x2_t __ret; \
23429  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
23430  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23431  __ret; \
23432 })
23433 #define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
23434  int32x2_t __s0 = __p0; \
23435  int64x2_t __ret; \
23436  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23437  __ret; \
23438 })
23439 #endif
23440 
23441 #ifdef __LITTLE_ENDIAN__
23442 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
23443  int16x4_t __s0 = __p0; \
23444  int32x4_t __ret; \
23445  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23446  __ret; \
23447 })
23448 #else
23449 #define vshll_n_s16(__p0, __p1) __extension__ ({ \
23450  int16x4_t __s0 = __p0; \
23451  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23452  int32x4_t __ret; \
23453  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
23454  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23455  __ret; \
23456 })
23457 #define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
23458  int16x4_t __s0 = __p0; \
23459  int32x4_t __ret; \
23460  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23461  __ret; \
23462 })
23463 #endif
23464 
23465 #ifdef __LITTLE_ENDIAN__
23466 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23467  uint8x16_t __s0 = __p0; \
23468  uint8x16_t __ret; \
23469  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
23470  __ret; \
23471 })
23472 #else
23473 #define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23474  uint8x16_t __s0 = __p0; \
23475  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23476  uint8x16_t __ret; \
23477  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
23478  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23479  __ret; \
23480 })
23481 #endif
23482 
23483 #ifdef __LITTLE_ENDIAN__
23484 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23485  uint32x4_t __s0 = __p0; \
23486  uint32x4_t __ret; \
23487  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
23488  __ret; \
23489 })
23490 #else
23491 #define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23492  uint32x4_t __s0 = __p0; \
23493  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23494  uint32x4_t __ret; \
23495  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
23496  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23497  __ret; \
23498 })
23499 #endif
23500 
23501 #ifdef __LITTLE_ENDIAN__
23502 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23503  uint64x2_t __s0 = __p0; \
23504  uint64x2_t __ret; \
23505  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
23506  __ret; \
23507 })
23508 #else
23509 #define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23510  uint64x2_t __s0 = __p0; \
23511  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23512  uint64x2_t __ret; \
23513  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
23514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23515  __ret; \
23516 })
23517 #endif
23518 
23519 #ifdef __LITTLE_ENDIAN__
23520 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23521  uint16x8_t __s0 = __p0; \
23522  uint16x8_t __ret; \
23523  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
23524  __ret; \
23525 })
23526 #else
23527 #define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23528  uint16x8_t __s0 = __p0; \
23529  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23530  uint16x8_t __ret; \
23531  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
23532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23533  __ret; \
23534 })
23535 #endif
23536 
23537 #ifdef __LITTLE_ENDIAN__
23538 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23539  int8x16_t __s0 = __p0; \
23540  int8x16_t __ret; \
23541  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
23542  __ret; \
23543 })
23544 #else
23545 #define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23546  int8x16_t __s0 = __p0; \
23547  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23548  int8x16_t __ret; \
23549  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
23550  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23551  __ret; \
23552 })
23553 #endif
23554 
23555 #ifdef __LITTLE_ENDIAN__
23556 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23557  int32x4_t __s0 = __p0; \
23558  int32x4_t __ret; \
23559  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
23560  __ret; \
23561 })
23562 #else
23563 #define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23564  int32x4_t __s0 = __p0; \
23565  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23566  int32x4_t __ret; \
23567  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
23568  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23569  __ret; \
23570 })
23571 #endif
23572 
23573 #ifdef __LITTLE_ENDIAN__
23574 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23575  int64x2_t __s0 = __p0; \
23576  int64x2_t __ret; \
23577  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
23578  __ret; \
23579 })
23580 #else
23581 #define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23582  int64x2_t __s0 = __p0; \
23583  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23584  int64x2_t __ret; \
23585  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
23586  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23587  __ret; \
23588 })
23589 #endif
23590 
23591 #ifdef __LITTLE_ENDIAN__
23592 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23593  int16x8_t __s0 = __p0; \
23594  int16x8_t __ret; \
23595  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
23596  __ret; \
23597 })
23598 #else
23599 #define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23600  int16x8_t __s0 = __p0; \
23601  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23602  int16x8_t __ret; \
23603  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
23604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23605  __ret; \
23606 })
23607 #endif
23608 
23609 #ifdef __LITTLE_ENDIAN__
23610 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
23611  uint8x8_t __s0 = __p0; \
23612  uint8x8_t __ret; \
23613  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
23614  __ret; \
23615 })
23616 #else
23617 #define vshr_n_u8(__p0, __p1) __extension__ ({ \
23618  uint8x8_t __s0 = __p0; \
23619  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23620  uint8x8_t __ret; \
23621  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
23622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23623  __ret; \
23624 })
23625 #endif
23626 
23627 #ifdef __LITTLE_ENDIAN__
23628 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
23629  uint32x2_t __s0 = __p0; \
23630  uint32x2_t __ret; \
23631  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
23632  __ret; \
23633 })
23634 #else
23635 #define vshr_n_u32(__p0, __p1) __extension__ ({ \
23636  uint32x2_t __s0 = __p0; \
23637  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23638  uint32x2_t __ret; \
23639  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
23640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23641  __ret; \
23642 })
23643 #endif
23644 
23645 #ifdef __LITTLE_ENDIAN__
23646 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
23647  uint64x1_t __s0 = __p0; \
23648  uint64x1_t __ret; \
23649  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23650  __ret; \
23651 })
23652 #else
23653 #define vshr_n_u64(__p0, __p1) __extension__ ({ \
23654  uint64x1_t __s0 = __p0; \
23655  uint64x1_t __ret; \
23656  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23657  __ret; \
23658 })
23659 #endif
23660 
23661 #ifdef __LITTLE_ENDIAN__
23662 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
23663  uint16x4_t __s0 = __p0; \
23664  uint16x4_t __ret; \
23665  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
23666  __ret; \
23667 })
23668 #else
23669 #define vshr_n_u16(__p0, __p1) __extension__ ({ \
23670  uint16x4_t __s0 = __p0; \
23671  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23672  uint16x4_t __ret; \
23673  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
23674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23675  __ret; \
23676 })
23677 #endif
23678 
23679 #ifdef __LITTLE_ENDIAN__
23680 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
23681  int8x8_t __s0 = __p0; \
23682  int8x8_t __ret; \
23683  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
23684  __ret; \
23685 })
23686 #else
23687 #define vshr_n_s8(__p0, __p1) __extension__ ({ \
23688  int8x8_t __s0 = __p0; \
23689  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23690  int8x8_t __ret; \
23691  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
23692  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23693  __ret; \
23694 })
23695 #endif
23696 
23697 #ifdef __LITTLE_ENDIAN__
23698 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
23699  int32x2_t __s0 = __p0; \
23700  int32x2_t __ret; \
23701  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
23702  __ret; \
23703 })
23704 #else
23705 #define vshr_n_s32(__p0, __p1) __extension__ ({ \
23706  int32x2_t __s0 = __p0; \
23707  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23708  int32x2_t __ret; \
23709  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
23710  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23711  __ret; \
23712 })
23713 #endif
23714 
23715 #ifdef __LITTLE_ENDIAN__
23716 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
23717  int64x1_t __s0 = __p0; \
23718  int64x1_t __ret; \
23719  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23720  __ret; \
23721 })
23722 #else
23723 #define vshr_n_s64(__p0, __p1) __extension__ ({ \
23724  int64x1_t __s0 = __p0; \
23725  int64x1_t __ret; \
23726  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23727  __ret; \
23728 })
23729 #endif
23730 
23731 #ifdef __LITTLE_ENDIAN__
23732 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
23733  int16x4_t __s0 = __p0; \
23734  int16x4_t __ret; \
23735  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
23736  __ret; \
23737 })
23738 #else
23739 #define vshr_n_s16(__p0, __p1) __extension__ ({ \
23740  int16x4_t __s0 = __p0; \
23741  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23742  int16x4_t __ret; \
23743  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
23744  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23745  __ret; \
23746 })
23747 #endif
23748 
23749 #ifdef __LITTLE_ENDIAN__
23750 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23751  uint32x4_t __s0 = __p0; \
23752  uint16x4_t __ret; \
23753  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23754  __ret; \
23755 })
23756 #else
23757 #define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23758  uint32x4_t __s0 = __p0; \
23759  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23760  uint16x4_t __ret; \
23761  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23762  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23763  __ret; \
23764 })
23765 #define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
23766  uint32x4_t __s0 = __p0; \
23767  uint16x4_t __ret; \
23768  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23769  __ret; \
23770 })
23771 #endif
23772 
23773 #ifdef __LITTLE_ENDIAN__
23774 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23775  uint64x2_t __s0 = __p0; \
23776  uint32x2_t __ret; \
23777  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23778  __ret; \
23779 })
23780 #else
23781 #define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23782  uint64x2_t __s0 = __p0; \
23783  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23784  uint32x2_t __ret; \
23785  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23787  __ret; \
23788 })
23789 #define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
23790  uint64x2_t __s0 = __p0; \
23791  uint32x2_t __ret; \
23792  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23793  __ret; \
23794 })
23795 #endif
23796 
23797 #ifdef __LITTLE_ENDIAN__
23798 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23799  uint16x8_t __s0 = __p0; \
23800  uint8x8_t __ret; \
23801  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23802  __ret; \
23803 })
23804 #else
23805 #define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23806  uint16x8_t __s0 = __p0; \
23807  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23808  uint8x8_t __ret; \
23809  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23810  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23811  __ret; \
23812 })
23813 #define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
23814  uint16x8_t __s0 = __p0; \
23815  uint8x8_t __ret; \
23816  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23817  __ret; \
23818 })
23819 #endif
23820 
23821 #ifdef __LITTLE_ENDIAN__
23822 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23823  int32x4_t __s0 = __p0; \
23824  int16x4_t __ret; \
23825  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23826  __ret; \
23827 })
23828 #else
23829 #define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23830  int32x4_t __s0 = __p0; \
23831  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23832  int16x4_t __ret; \
23833  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23834  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23835  __ret; \
23836 })
23837 #define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
23838  int32x4_t __s0 = __p0; \
23839  int16x4_t __ret; \
23840  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23841  __ret; \
23842 })
23843 #endif
23844 
23845 #ifdef __LITTLE_ENDIAN__
23846 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23847  int64x2_t __s0 = __p0; \
23848  int32x2_t __ret; \
23849  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23850  __ret; \
23851 })
23852 #else
23853 #define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23854  int64x2_t __s0 = __p0; \
23855  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23856  int32x2_t __ret; \
23857  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23859  __ret; \
23860 })
23861 #define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
23862  int64x2_t __s0 = __p0; \
23863  int32x2_t __ret; \
23864  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23865  __ret; \
23866 })
23867 #endif
23868 
23869 #ifdef __LITTLE_ENDIAN__
23870 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23871  int16x8_t __s0 = __p0; \
23872  int8x8_t __ret; \
23873  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23874  __ret; \
23875 })
23876 #else
23877 #define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23878  int16x8_t __s0 = __p0; \
23879  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23880  int8x8_t __ret; \
23881  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23882  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23883  __ret; \
23884 })
23885 #define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
23886  int16x8_t __s0 = __p0; \
23887  int8x8_t __ret; \
23888  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23889  __ret; \
23890 })
23891 #endif
23892 
23893 #ifdef __LITTLE_ENDIAN__
23894 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23895  poly8x8_t __s0 = __p0; \
23896  poly8x8_t __s1 = __p1; \
23897  poly8x8_t __ret; \
23898  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
23899  __ret; \
23900 })
23901 #else
23902 #define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23903  poly8x8_t __s0 = __p0; \
23904  poly8x8_t __s1 = __p1; \
23905  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23906  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23907  poly8x8_t __ret; \
23908  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
23909  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23910  __ret; \
23911 })
23912 #endif
23913 
23914 #ifdef __LITTLE_ENDIAN__
23915 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23916  poly16x4_t __s0 = __p0; \
23917  poly16x4_t __s1 = __p1; \
23918  poly16x4_t __ret; \
23919  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
23920  __ret; \
23921 })
23922 #else
23923 #define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23924  poly16x4_t __s0 = __p0; \
23925  poly16x4_t __s1 = __p1; \
23926  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23927  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23928  poly16x4_t __ret; \
23929  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
23930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23931  __ret; \
23932 })
23933 #endif
23934 
23935 #ifdef __LITTLE_ENDIAN__
23936 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23937  poly8x16_t __s0 = __p0; \
23938  poly8x16_t __s1 = __p1; \
23939  poly8x16_t __ret; \
23940  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
23941  __ret; \
23942 })
23943 #else
23944 #define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23945  poly8x16_t __s0 = __p0; \
23946  poly8x16_t __s1 = __p1; \
23947  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23948  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23949  poly8x16_t __ret; \
23950  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
23951  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23952  __ret; \
23953 })
23954 #endif
23955 
23956 #ifdef __LITTLE_ENDIAN__
23957 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23958  poly16x8_t __s0 = __p0; \
23959  poly16x8_t __s1 = __p1; \
23960  poly16x8_t __ret; \
23961  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
23962  __ret; \
23963 })
23964 #else
23965 #define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23966  poly16x8_t __s0 = __p0; \
23967  poly16x8_t __s1 = __p1; \
23968  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23969  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23970  poly16x8_t __ret; \
23971  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
23972  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23973  __ret; \
23974 })
23975 #endif
23976 
23977 #ifdef __LITTLE_ENDIAN__
23978 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23979  uint8x16_t __s0 = __p0; \
23980  uint8x16_t __s1 = __p1; \
23981  uint8x16_t __ret; \
23982  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23983  __ret; \
23984 })
23985 #else
23986 #define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23987  uint8x16_t __s0 = __p0; \
23988  uint8x16_t __s1 = __p1; \
23989  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23990  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23991  uint8x16_t __ret; \
23992  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23993  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23994  __ret; \
23995 })
23996 #endif
23997 
23998 #ifdef __LITTLE_ENDIAN__
23999 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24000  uint32x4_t __s0 = __p0; \
24001  uint32x4_t __s1 = __p1; \
24002  uint32x4_t __ret; \
24003  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24004  __ret; \
24005 })
24006 #else
24007 #define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24008  uint32x4_t __s0 = __p0; \
24009  uint32x4_t __s1 = __p1; \
24010  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24011  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24012  uint32x4_t __ret; \
24013  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24014  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24015  __ret; \
24016 })
24017 #endif
24018 
24019 #ifdef __LITTLE_ENDIAN__
24020 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24021  uint64x2_t __s0 = __p0; \
24022  uint64x2_t __s1 = __p1; \
24023  uint64x2_t __ret; \
24024  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24025  __ret; \
24026 })
24027 #else
24028 #define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24029  uint64x2_t __s0 = __p0; \
24030  uint64x2_t __s1 = __p1; \
24031  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24032  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24033  uint64x2_t __ret; \
24034  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24035  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24036  __ret; \
24037 })
24038 #endif
24039 
24040 #ifdef __LITTLE_ENDIAN__
24041 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24042  uint16x8_t __s0 = __p0; \
24043  uint16x8_t __s1 = __p1; \
24044  uint16x8_t __ret; \
24045  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24046  __ret; \
24047 })
24048 #else
24049 #define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24050  uint16x8_t __s0 = __p0; \
24051  uint16x8_t __s1 = __p1; \
24052  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24053  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24054  uint16x8_t __ret; \
24055  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24056  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24057  __ret; \
24058 })
24059 #endif
24060 
24061 #ifdef __LITTLE_ENDIAN__
24062 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24063  int8x16_t __s0 = __p0; \
24064  int8x16_t __s1 = __p1; \
24065  int8x16_t __ret; \
24066  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24067  __ret; \
24068 })
24069 #else
24070 #define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24071  int8x16_t __s0 = __p0; \
24072  int8x16_t __s1 = __p1; \
24073  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24074  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24075  int8x16_t __ret; \
24076  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24077  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24078  __ret; \
24079 })
24080 #endif
24081 
24082 #ifdef __LITTLE_ENDIAN__
24083 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24084  int32x4_t __s0 = __p0; \
24085  int32x4_t __s1 = __p1; \
24086  int32x4_t __ret; \
24087  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24088  __ret; \
24089 })
24090 #else
24091 #define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24092  int32x4_t __s0 = __p0; \
24093  int32x4_t __s1 = __p1; \
24094  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24095  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24096  int32x4_t __ret; \
24097  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24099  __ret; \
24100 })
24101 #endif
24102 
24103 #ifdef __LITTLE_ENDIAN__
24104 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24105  int64x2_t __s0 = __p0; \
24106  int64x2_t __s1 = __p1; \
24107  int64x2_t __ret; \
24108  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24109  __ret; \
24110 })
24111 #else
24112 #define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24113  int64x2_t __s0 = __p0; \
24114  int64x2_t __s1 = __p1; \
24115  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24116  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24117  int64x2_t __ret; \
24118  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24119  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24120  __ret; \
24121 })
24122 #endif
24123 
24124 #ifdef __LITTLE_ENDIAN__
24125 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24126  int16x8_t __s0 = __p0; \
24127  int16x8_t __s1 = __p1; \
24128  int16x8_t __ret; \
24129  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24130  __ret; \
24131 })
24132 #else
24133 #define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24134  int16x8_t __s0 = __p0; \
24135  int16x8_t __s1 = __p1; \
24136  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24137  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24138  int16x8_t __ret; \
24139  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24141  __ret; \
24142 })
24143 #endif
24144 
24145 #ifdef __LITTLE_ENDIAN__
24146 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24147  uint8x8_t __s0 = __p0; \
24148  uint8x8_t __s1 = __p1; \
24149  uint8x8_t __ret; \
24150  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24151  __ret; \
24152 })
24153 #else
24154 #define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24155  uint8x8_t __s0 = __p0; \
24156  uint8x8_t __s1 = __p1; \
24157  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24158  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24159  uint8x8_t __ret; \
24160  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24161  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24162  __ret; \
24163 })
24164 #endif
24165 
24166 #ifdef __LITTLE_ENDIAN__
24167 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24168  uint32x2_t __s0 = __p0; \
24169  uint32x2_t __s1 = __p1; \
24170  uint32x2_t __ret; \
24171  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24172  __ret; \
24173 })
24174 #else
24175 #define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24176  uint32x2_t __s0 = __p0; \
24177  uint32x2_t __s1 = __p1; \
24178  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24179  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24180  uint32x2_t __ret; \
24181  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24182  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24183  __ret; \
24184 })
24185 #endif
24186 
24187 #ifdef __LITTLE_ENDIAN__
24188 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24189  uint64x1_t __s0 = __p0; \
24190  uint64x1_t __s1 = __p1; \
24191  uint64x1_t __ret; \
24192  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24193  __ret; \
24194 })
24195 #else
24196 #define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24197  uint64x1_t __s0 = __p0; \
24198  uint64x1_t __s1 = __p1; \
24199  uint64x1_t __ret; \
24200  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24201  __ret; \
24202 })
24203 #endif
24204 
24205 #ifdef __LITTLE_ENDIAN__
24206 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24207  uint16x4_t __s0 = __p0; \
24208  uint16x4_t __s1 = __p1; \
24209  uint16x4_t __ret; \
24210  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24211  __ret; \
24212 })
24213 #else
24214 #define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24215  uint16x4_t __s0 = __p0; \
24216  uint16x4_t __s1 = __p1; \
24217  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24218  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24219  uint16x4_t __ret; \
24220  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24221  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24222  __ret; \
24223 })
24224 #endif
24225 
24226 #ifdef __LITTLE_ENDIAN__
24227 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24228  int8x8_t __s0 = __p0; \
24229  int8x8_t __s1 = __p1; \
24230  int8x8_t __ret; \
24231  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24232  __ret; \
24233 })
24234 #else
24235 #define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24236  int8x8_t __s0 = __p0; \
24237  int8x8_t __s1 = __p1; \
24238  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24239  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24240  int8x8_t __ret; \
24241  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24243  __ret; \
24244 })
24245 #endif
24246 
24247 #ifdef __LITTLE_ENDIAN__
24248 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24249  int32x2_t __s0 = __p0; \
24250  int32x2_t __s1 = __p1; \
24251  int32x2_t __ret; \
24252  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24253  __ret; \
24254 })
24255 #else
24256 #define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24257  int32x2_t __s0 = __p0; \
24258  int32x2_t __s1 = __p1; \
24259  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24260  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24261  int32x2_t __ret; \
24262  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24264  __ret; \
24265 })
24266 #endif
24267 
24268 #ifdef __LITTLE_ENDIAN__
24269 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24270  int64x1_t __s0 = __p0; \
24271  int64x1_t __s1 = __p1; \
24272  int64x1_t __ret; \
24273  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24274  __ret; \
24275 })
24276 #else
24277 #define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24278  int64x1_t __s0 = __p0; \
24279  int64x1_t __s1 = __p1; \
24280  int64x1_t __ret; \
24281  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24282  __ret; \
24283 })
24284 #endif
24285 
24286 #ifdef __LITTLE_ENDIAN__
24287 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24288  int16x4_t __s0 = __p0; \
24289  int16x4_t __s1 = __p1; \
24290  int16x4_t __ret; \
24291  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24292  __ret; \
24293 })
24294 #else
24295 #define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24296  int16x4_t __s0 = __p0; \
24297  int16x4_t __s1 = __p1; \
24298  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24299  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24300  int16x4_t __ret; \
24301  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24302  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24303  __ret; \
24304 })
24305 #endif
24306 
24307 #ifdef __LITTLE_ENDIAN__
24308 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24309  uint8x16_t __s0 = __p0; \
24310  uint8x16_t __s1 = __p1; \
24311  uint8x16_t __ret; \
24312  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24313  __ret; \
24314 })
24315 #else
24316 #define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24317  uint8x16_t __s0 = __p0; \
24318  uint8x16_t __s1 = __p1; \
24319  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24320  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24321  uint8x16_t __ret; \
24322  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24323  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24324  __ret; \
24325 })
24326 #endif
24327 
24328 #ifdef __LITTLE_ENDIAN__
24329 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24330  uint32x4_t __s0 = __p0; \
24331  uint32x4_t __s1 = __p1; \
24332  uint32x4_t __ret; \
24333  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24334  __ret; \
24335 })
24336 #else
24337 #define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24338  uint32x4_t __s0 = __p0; \
24339  uint32x4_t __s1 = __p1; \
24340  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24341  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24342  uint32x4_t __ret; \
24343  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24344  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24345  __ret; \
24346 })
24347 #endif
24348 
24349 #ifdef __LITTLE_ENDIAN__
24350 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24351  uint64x2_t __s0 = __p0; \
24352  uint64x2_t __s1 = __p1; \
24353  uint64x2_t __ret; \
24354  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24355  __ret; \
24356 })
24357 #else
24358 #define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24359  uint64x2_t __s0 = __p0; \
24360  uint64x2_t __s1 = __p1; \
24361  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24362  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24363  uint64x2_t __ret; \
24364  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24365  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24366  __ret; \
24367 })
24368 #endif
24369 
24370 #ifdef __LITTLE_ENDIAN__
24371 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24372  uint16x8_t __s0 = __p0; \
24373  uint16x8_t __s1 = __p1; \
24374  uint16x8_t __ret; \
24375  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24376  __ret; \
24377 })
24378 #else
24379 #define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24380  uint16x8_t __s0 = __p0; \
24381  uint16x8_t __s1 = __p1; \
24382  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24383  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24384  uint16x8_t __ret; \
24385  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24386  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24387  __ret; \
24388 })
24389 #endif
24390 
24391 #ifdef __LITTLE_ENDIAN__
24392 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24393  int8x16_t __s0 = __p0; \
24394  int8x16_t __s1 = __p1; \
24395  int8x16_t __ret; \
24396  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24397  __ret; \
24398 })
24399 #else
24400 #define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24401  int8x16_t __s0 = __p0; \
24402  int8x16_t __s1 = __p1; \
24403  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24404  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24405  int8x16_t __ret; \
24406  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24407  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24408  __ret; \
24409 })
24410 #endif
24411 
24412 #ifdef __LITTLE_ENDIAN__
24413 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24414  int32x4_t __s0 = __p0; \
24415  int32x4_t __s1 = __p1; \
24416  int32x4_t __ret; \
24417  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24418  __ret; \
24419 })
24420 #else
24421 #define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24422  int32x4_t __s0 = __p0; \
24423  int32x4_t __s1 = __p1; \
24424  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24425  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24426  int32x4_t __ret; \
24427  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24429  __ret; \
24430 })
24431 #endif
24432 
24433 #ifdef __LITTLE_ENDIAN__
24434 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24435  int64x2_t __s0 = __p0; \
24436  int64x2_t __s1 = __p1; \
24437  int64x2_t __ret; \
24438  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24439  __ret; \
24440 })
24441 #else
24442 #define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24443  int64x2_t __s0 = __p0; \
24444  int64x2_t __s1 = __p1; \
24445  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24446  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24447  int64x2_t __ret; \
24448  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24449  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24450  __ret; \
24451 })
24452 #endif
24453 
24454 #ifdef __LITTLE_ENDIAN__
24455 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24456  int16x8_t __s0 = __p0; \
24457  int16x8_t __s1 = __p1; \
24458  int16x8_t __ret; \
24459  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24460  __ret; \
24461 })
24462 #else
24463 #define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24464  int16x8_t __s0 = __p0; \
24465  int16x8_t __s1 = __p1; \
24466  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24467  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24468  int16x8_t __ret; \
24469  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24470  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24471  __ret; \
24472 })
24473 #endif
24474 
24475 #ifdef __LITTLE_ENDIAN__
24476 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24477  uint8x8_t __s0 = __p0; \
24478  uint8x8_t __s1 = __p1; \
24479  uint8x8_t __ret; \
24480  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24481  __ret; \
24482 })
24483 #else
24484 #define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24485  uint8x8_t __s0 = __p0; \
24486  uint8x8_t __s1 = __p1; \
24487  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24488  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24489  uint8x8_t __ret; \
24490  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24492  __ret; \
24493 })
24494 #endif
24495 
24496 #ifdef __LITTLE_ENDIAN__
24497 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24498  uint32x2_t __s0 = __p0; \
24499  uint32x2_t __s1 = __p1; \
24500  uint32x2_t __ret; \
24501  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24502  __ret; \
24503 })
24504 #else
24505 #define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24506  uint32x2_t __s0 = __p0; \
24507  uint32x2_t __s1 = __p1; \
24508  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24509  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24510  uint32x2_t __ret; \
24511  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24512  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24513  __ret; \
24514 })
24515 #endif
24516 
24517 #ifdef __LITTLE_ENDIAN__
24518 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24519  uint64x1_t __s0 = __p0; \
24520  uint64x1_t __s1 = __p1; \
24521  uint64x1_t __ret; \
24522  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24523  __ret; \
24524 })
24525 #else
24526 #define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24527  uint64x1_t __s0 = __p0; \
24528  uint64x1_t __s1 = __p1; \
24529  uint64x1_t __ret; \
24530  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24531  __ret; \
24532 })
24533 #endif
24534 
24535 #ifdef __LITTLE_ENDIAN__
24536 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24537  uint16x4_t __s0 = __p0; \
24538  uint16x4_t __s1 = __p1; \
24539  uint16x4_t __ret; \
24540  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24541  __ret; \
24542 })
24543 #else
24544 #define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24545  uint16x4_t __s0 = __p0; \
24546  uint16x4_t __s1 = __p1; \
24547  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24548  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24549  uint16x4_t __ret; \
24550  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24552  __ret; \
24553 })
24554 #endif
24555 
24556 #ifdef __LITTLE_ENDIAN__
24557 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24558  int8x8_t __s0 = __p0; \
24559  int8x8_t __s1 = __p1; \
24560  int8x8_t __ret; \
24561  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24562  __ret; \
24563 })
24564 #else
24565 #define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24566  int8x8_t __s0 = __p0; \
24567  int8x8_t __s1 = __p1; \
24568  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24569  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24570  int8x8_t __ret; \
24571  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24572  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24573  __ret; \
24574 })
24575 #endif
24576 
24577 #ifdef __LITTLE_ENDIAN__
24578 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24579  int32x2_t __s0 = __p0; \
24580  int32x2_t __s1 = __p1; \
24581  int32x2_t __ret; \
24582  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24583  __ret; \
24584 })
24585 #else
24586 #define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24587  int32x2_t __s0 = __p0; \
24588  int32x2_t __s1 = __p1; \
24589  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24590  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24591  int32x2_t __ret; \
24592  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24594  __ret; \
24595 })
24596 #endif
24597 
24598 #ifdef __LITTLE_ENDIAN__
24599 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24600  int64x1_t __s0 = __p0; \
24601  int64x1_t __s1 = __p1; \
24602  int64x1_t __ret; \
24603  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24604  __ret; \
24605 })
24606 #else
24607 #define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24608  int64x1_t __s0 = __p0; \
24609  int64x1_t __s1 = __p1; \
24610  int64x1_t __ret; \
24611  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24612  __ret; \
24613 })
24614 #endif
24615 
24616 #ifdef __LITTLE_ENDIAN__
24617 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24618  int16x4_t __s0 = __p0; \
24619  int16x4_t __s1 = __p1; \
24620  int16x4_t __ret; \
24621  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24622  __ret; \
24623 })
24624 #else
24625 #define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24626  int16x4_t __s0 = __p0; \
24627  int16x4_t __s1 = __p1; \
24628  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24629  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24630  int16x4_t __ret; \
24631  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24633  __ret; \
24634 })
24635 #endif
24636 
24637 #ifdef __LITTLE_ENDIAN__
24638 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24639  poly8x8_t __s0 = __p0; \
24640  poly8x8_t __s1 = __p1; \
24641  poly8x8_t __ret; \
24642  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
24643  __ret; \
24644 })
24645 #else
24646 #define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24647  poly8x8_t __s0 = __p0; \
24648  poly8x8_t __s1 = __p1; \
24649  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24650  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24651  poly8x8_t __ret; \
24652  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
24653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24654  __ret; \
24655 })
24656 #endif
24657 
24658 #ifdef __LITTLE_ENDIAN__
24659 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24660  poly16x4_t __s0 = __p0; \
24661  poly16x4_t __s1 = __p1; \
24662  poly16x4_t __ret; \
24663  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
24664  __ret; \
24665 })
24666 #else
24667 #define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24668  poly16x4_t __s0 = __p0; \
24669  poly16x4_t __s1 = __p1; \
24670  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24671  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24672  poly16x4_t __ret; \
24673  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
24674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24675  __ret; \
24676 })
24677 #endif
24678 
24679 #ifdef __LITTLE_ENDIAN__
24680 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24681  poly8x16_t __s0 = __p0; \
24682  poly8x16_t __s1 = __p1; \
24683  poly8x16_t __ret; \
24684  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
24685  __ret; \
24686 })
24687 #else
24688 #define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24689  poly8x16_t __s0 = __p0; \
24690  poly8x16_t __s1 = __p1; \
24691  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24692  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24693  poly8x16_t __ret; \
24694  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
24695  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24696  __ret; \
24697 })
24698 #endif
24699 
24700 #ifdef __LITTLE_ENDIAN__
24701 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24702  poly16x8_t __s0 = __p0; \
24703  poly16x8_t __s1 = __p1; \
24704  poly16x8_t __ret; \
24705  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
24706  __ret; \
24707 })
24708 #else
24709 #define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24710  poly16x8_t __s0 = __p0; \
24711  poly16x8_t __s1 = __p1; \
24712  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24713  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24714  poly16x8_t __ret; \
24715  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
24716  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24717  __ret; \
24718 })
24719 #endif
24720 
24721 #ifdef __LITTLE_ENDIAN__
24722 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24723  uint8x16_t __s0 = __p0; \
24724  uint8x16_t __s1 = __p1; \
24725  uint8x16_t __ret; \
24726  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24727  __ret; \
24728 })
24729 #else
24730 #define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24731  uint8x16_t __s0 = __p0; \
24732  uint8x16_t __s1 = __p1; \
24733  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24734  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24735  uint8x16_t __ret; \
24736  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24737  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24738  __ret; \
24739 })
24740 #endif
24741 
24742 #ifdef __LITTLE_ENDIAN__
24743 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24744  uint32x4_t __s0 = __p0; \
24745  uint32x4_t __s1 = __p1; \
24746  uint32x4_t __ret; \
24747  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24748  __ret; \
24749 })
24750 #else
24751 #define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24752  uint32x4_t __s0 = __p0; \
24753  uint32x4_t __s1 = __p1; \
24754  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24755  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24756  uint32x4_t __ret; \
24757  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24758  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24759  __ret; \
24760 })
24761 #endif
24762 
24763 #ifdef __LITTLE_ENDIAN__
24764 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24765  uint64x2_t __s0 = __p0; \
24766  uint64x2_t __s1 = __p1; \
24767  uint64x2_t __ret; \
24768  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24769  __ret; \
24770 })
24771 #else
24772 #define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24773  uint64x2_t __s0 = __p0; \
24774  uint64x2_t __s1 = __p1; \
24775  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24776  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24777  uint64x2_t __ret; \
24778  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24779  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24780  __ret; \
24781 })
24782 #endif
24783 
24784 #ifdef __LITTLE_ENDIAN__
24785 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24786  uint16x8_t __s0 = __p0; \
24787  uint16x8_t __s1 = __p1; \
24788  uint16x8_t __ret; \
24789  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24790  __ret; \
24791 })
24792 #else
24793 #define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24794  uint16x8_t __s0 = __p0; \
24795  uint16x8_t __s1 = __p1; \
24796  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24797  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24798  uint16x8_t __ret; \
24799  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24801  __ret; \
24802 })
24803 #endif
24804 
24805 #ifdef __LITTLE_ENDIAN__
24806 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24807  int8x16_t __s0 = __p0; \
24808  int8x16_t __s1 = __p1; \
24809  int8x16_t __ret; \
24810  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24811  __ret; \
24812 })
24813 #else
24814 #define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24815  int8x16_t __s0 = __p0; \
24816  int8x16_t __s1 = __p1; \
24817  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24818  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24819  int8x16_t __ret; \
24820  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24821  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24822  __ret; \
24823 })
24824 #endif
24825 
24826 #ifdef __LITTLE_ENDIAN__
24827 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24828  int32x4_t __s0 = __p0; \
24829  int32x4_t __s1 = __p1; \
24830  int32x4_t __ret; \
24831  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24832  __ret; \
24833 })
24834 #else
24835 #define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24836  int32x4_t __s0 = __p0; \
24837  int32x4_t __s1 = __p1; \
24838  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24839  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24840  int32x4_t __ret; \
24841  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24843  __ret; \
24844 })
24845 #endif
24846 
24847 #ifdef __LITTLE_ENDIAN__
24848 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24849  int64x2_t __s0 = __p0; \
24850  int64x2_t __s1 = __p1; \
24851  int64x2_t __ret; \
24852  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24853  __ret; \
24854 })
24855 #else
24856 #define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24857  int64x2_t __s0 = __p0; \
24858  int64x2_t __s1 = __p1; \
24859  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24860  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24861  int64x2_t __ret; \
24862  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24863  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24864  __ret; \
24865 })
24866 #endif
24867 
24868 #ifdef __LITTLE_ENDIAN__
24869 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24870  int16x8_t __s0 = __p0; \
24871  int16x8_t __s1 = __p1; \
24872  int16x8_t __ret; \
24873  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24874  __ret; \
24875 })
24876 #else
24877 #define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24878  int16x8_t __s0 = __p0; \
24879  int16x8_t __s1 = __p1; \
24880  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24881  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24882  int16x8_t __ret; \
24883  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24884  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24885  __ret; \
24886 })
24887 #endif
24888 
24889 #ifdef __LITTLE_ENDIAN__
24890 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24891  uint8x8_t __s0 = __p0; \
24892  uint8x8_t __s1 = __p1; \
24893  uint8x8_t __ret; \
24894  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24895  __ret; \
24896 })
24897 #else
24898 #define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24899  uint8x8_t __s0 = __p0; \
24900  uint8x8_t __s1 = __p1; \
24901  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24902  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24903  uint8x8_t __ret; \
24904  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24905  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24906  __ret; \
24907 })
24908 #endif
24909 
24910 #ifdef __LITTLE_ENDIAN__
24911 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24912  uint32x2_t __s0 = __p0; \
24913  uint32x2_t __s1 = __p1; \
24914  uint32x2_t __ret; \
24915  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24916  __ret; \
24917 })
24918 #else
24919 #define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24920  uint32x2_t __s0 = __p0; \
24921  uint32x2_t __s1 = __p1; \
24922  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24923  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24924  uint32x2_t __ret; \
24925  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24927  __ret; \
24928 })
24929 #endif
24930 
24931 #ifdef __LITTLE_ENDIAN__
24932 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24933  uint64x1_t __s0 = __p0; \
24934  uint64x1_t __s1 = __p1; \
24935  uint64x1_t __ret; \
24936  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24937  __ret; \
24938 })
24939 #else
24940 #define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24941  uint64x1_t __s0 = __p0; \
24942  uint64x1_t __s1 = __p1; \
24943  uint64x1_t __ret; \
24944  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24945  __ret; \
24946 })
24947 #endif
24948 
24949 #ifdef __LITTLE_ENDIAN__
24950 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24951  uint16x4_t __s0 = __p0; \
24952  uint16x4_t __s1 = __p1; \
24953  uint16x4_t __ret; \
24954  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24955  __ret; \
24956 })
24957 #else
24958 #define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24959  uint16x4_t __s0 = __p0; \
24960  uint16x4_t __s1 = __p1; \
24961  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24962  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24963  uint16x4_t __ret; \
24964  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24966  __ret; \
24967 })
24968 #endif
24969 
24970 #ifdef __LITTLE_ENDIAN__
24971 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24972  int8x8_t __s0 = __p0; \
24973  int8x8_t __s1 = __p1; \
24974  int8x8_t __ret; \
24975  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24976  __ret; \
24977 })
24978 #else
24979 #define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24980  int8x8_t __s0 = __p0; \
24981  int8x8_t __s1 = __p1; \
24982  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24983  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24984  int8x8_t __ret; \
24985  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24986  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24987  __ret; \
24988 })
24989 #endif
24990 
24991 #ifdef __LITTLE_ENDIAN__
24992 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24993  int32x2_t __s0 = __p0; \
24994  int32x2_t __s1 = __p1; \
24995  int32x2_t __ret; \
24996  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24997  __ret; \
24998 })
24999 #else
25000 #define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
25001  int32x2_t __s0 = __p0; \
25002  int32x2_t __s1 = __p1; \
25003  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25004  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25005  int32x2_t __ret; \
25006  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
25007  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25008  __ret; \
25009 })
25010 #endif
25011 
25012 #ifdef __LITTLE_ENDIAN__
25013 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
25014  int64x1_t __s0 = __p0; \
25015  int64x1_t __s1 = __p1; \
25016  int64x1_t __ret; \
25017  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25018  __ret; \
25019 })
25020 #else
25021 #define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
25022  int64x1_t __s0 = __p0; \
25023  int64x1_t __s1 = __p1; \
25024  int64x1_t __ret; \
25025  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25026  __ret; \
25027 })
25028 #endif
25029 
25030 #ifdef __LITTLE_ENDIAN__
25031 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
25032  int16x4_t __s0 = __p0; \
25033  int16x4_t __s1 = __p1; \
25034  int16x4_t __ret; \
25035  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
25036  __ret; \
25037 })
25038 #else
25039 #define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
25040  int16x4_t __s0 = __p0; \
25041  int16x4_t __s1 = __p1; \
25042  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25043  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25044  int16x4_t __ret; \
25045  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
25046  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25047  __ret; \
25048 })
25049 #endif
25050 
25051 #ifdef __LITTLE_ENDIAN__
25052 #define vst1_p8(__p0, __p1) __extension__ ({ \
25053  poly8x8_t __s1 = __p1; \
25054  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
25055 })
25056 #else
25057 #define vst1_p8(__p0, __p1) __extension__ ({ \
25058  poly8x8_t __s1 = __p1; \
25059  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25060  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
25061 })
25062 #endif
25063 
25064 #ifdef __LITTLE_ENDIAN__
25065 #define vst1_p16(__p0, __p1) __extension__ ({ \
25066  poly16x4_t __s1 = __p1; \
25067  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
25068 })
25069 #else
25070 #define vst1_p16(__p0, __p1) __extension__ ({ \
25071  poly16x4_t __s1 = __p1; \
25072  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25073  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
25074 })
25075 #endif
25076 
25077 #ifdef __LITTLE_ENDIAN__
25078 #define vst1q_p8(__p0, __p1) __extension__ ({ \
25079  poly8x16_t __s1 = __p1; \
25080  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
25081 })
25082 #else
25083 #define vst1q_p8(__p0, __p1) __extension__ ({ \
25084  poly8x16_t __s1 = __p1; \
25085  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25086  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
25087 })
25088 #endif
25089 
25090 #ifdef __LITTLE_ENDIAN__
25091 #define vst1q_p16(__p0, __p1) __extension__ ({ \
25092  poly16x8_t __s1 = __p1; \
25093  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
25094 })
25095 #else
25096 #define vst1q_p16(__p0, __p1) __extension__ ({ \
25097  poly16x8_t __s1 = __p1; \
25098  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25099  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
25100 })
25101 #endif
25102 
25103 #ifdef __LITTLE_ENDIAN__
25104 #define vst1q_u8(__p0, __p1) __extension__ ({ \
25105  uint8x16_t __s1 = __p1; \
25106  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
25107 })
25108 #else
25109 #define vst1q_u8(__p0, __p1) __extension__ ({ \
25110  uint8x16_t __s1 = __p1; \
25111  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25112  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
25113 })
25114 #endif
25115 
25116 #ifdef __LITTLE_ENDIAN__
25117 #define vst1q_u32(__p0, __p1) __extension__ ({ \
25118  uint32x4_t __s1 = __p1; \
25119  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
25120 })
25121 #else
25122 #define vst1q_u32(__p0, __p1) __extension__ ({ \
25123  uint32x4_t __s1 = __p1; \
25124  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25125  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
25126 })
25127 #endif
25128 
25129 #ifdef __LITTLE_ENDIAN__
25130 #define vst1q_u64(__p0, __p1) __extension__ ({ \
25131  uint64x2_t __s1 = __p1; \
25132  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
25133 })
25134 #else
25135 #define vst1q_u64(__p0, __p1) __extension__ ({ \
25136  uint64x2_t __s1 = __p1; \
25137  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25138  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
25139 })
25140 #endif
25141 
25142 #ifdef __LITTLE_ENDIAN__
25143 #define vst1q_u16(__p0, __p1) __extension__ ({ \
25144  uint16x8_t __s1 = __p1; \
25145  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
25146 })
25147 #else
25148 #define vst1q_u16(__p0, __p1) __extension__ ({ \
25149  uint16x8_t __s1 = __p1; \
25150  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25151  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
25152 })
25153 #endif
25154 
25155 #ifdef __LITTLE_ENDIAN__
25156 #define vst1q_s8(__p0, __p1) __extension__ ({ \
25157  int8x16_t __s1 = __p1; \
25158  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
25159 })
25160 #else
25161 #define vst1q_s8(__p0, __p1) __extension__ ({ \
25162  int8x16_t __s1 = __p1; \
25163  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25164  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
25165 })
25166 #endif
25167 
25168 #ifdef __LITTLE_ENDIAN__
25169 #define vst1q_f32(__p0, __p1) __extension__ ({ \
25170  float32x4_t __s1 = __p1; \
25171  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
25172 })
25173 #else
25174 #define vst1q_f32(__p0, __p1) __extension__ ({ \
25175  float32x4_t __s1 = __p1; \
25176  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25177  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
25178 })
25179 #endif
25180 
25181 #ifdef __LITTLE_ENDIAN__
25182 #define vst1q_f16(__p0, __p1) __extension__ ({ \
25183  float16x8_t __s1 = __p1; \
25184  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
25185 })
25186 #else
25187 #define vst1q_f16(__p0, __p1) __extension__ ({ \
25188  float16x8_t __s1 = __p1; \
25189  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25190  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
25191 })
25192 #endif
25193 
25194 #ifdef __LITTLE_ENDIAN__
25195 #define vst1q_s32(__p0, __p1) __extension__ ({ \
25196  int32x4_t __s1 = __p1; \
25197  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
25198 })
25199 #else
25200 #define vst1q_s32(__p0, __p1) __extension__ ({ \
25201  int32x4_t __s1 = __p1; \
25202  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25203  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
25204 })
25205 #endif
25206 
25207 #ifdef __LITTLE_ENDIAN__
25208 #define vst1q_s64(__p0, __p1) __extension__ ({ \
25209  int64x2_t __s1 = __p1; \
25210  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
25211 })
25212 #else
25213 #define vst1q_s64(__p0, __p1) __extension__ ({ \
25214  int64x2_t __s1 = __p1; \
25215  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25216  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
25217 })
25218 #endif
25219 
25220 #ifdef __LITTLE_ENDIAN__
25221 #define vst1q_s16(__p0, __p1) __extension__ ({ \
25222  int16x8_t __s1 = __p1; \
25223  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
25224 })
25225 #else
25226 #define vst1q_s16(__p0, __p1) __extension__ ({ \
25227  int16x8_t __s1 = __p1; \
25228  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25229  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
25230 })
25231 #endif
25232 
25233 #ifdef __LITTLE_ENDIAN__
25234 #define vst1_u8(__p0, __p1) __extension__ ({ \
25235  uint8x8_t __s1 = __p1; \
25236  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
25237 })
25238 #else
25239 #define vst1_u8(__p0, __p1) __extension__ ({ \
25240  uint8x8_t __s1 = __p1; \
25241  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25242  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
25243 })
25244 #endif
25245 
25246 #ifdef __LITTLE_ENDIAN__
25247 #define vst1_u32(__p0, __p1) __extension__ ({ \
25248  uint32x2_t __s1 = __p1; \
25249  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
25250 })
25251 #else
25252 #define vst1_u32(__p0, __p1) __extension__ ({ \
25253  uint32x2_t __s1 = __p1; \
25254  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25255  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
25256 })
25257 #endif
25258 
25259 #ifdef __LITTLE_ENDIAN__
25260 #define vst1_u64(__p0, __p1) __extension__ ({ \
25261  uint64x1_t __s1 = __p1; \
25262  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25263 })
25264 #else
25265 #define vst1_u64(__p0, __p1) __extension__ ({ \
25266  uint64x1_t __s1 = __p1; \
25267  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25268 })
25269 #endif
25270 
25271 #ifdef __LITTLE_ENDIAN__
25272 #define vst1_u16(__p0, __p1) __extension__ ({ \
25273  uint16x4_t __s1 = __p1; \
25274  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
25275 })
25276 #else
25277 #define vst1_u16(__p0, __p1) __extension__ ({ \
25278  uint16x4_t __s1 = __p1; \
25279  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25280  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
25281 })
25282 #endif
25283 
25284 #ifdef __LITTLE_ENDIAN__
25285 #define vst1_s8(__p0, __p1) __extension__ ({ \
25286  int8x8_t __s1 = __p1; \
25287  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
25288 })
25289 #else
25290 #define vst1_s8(__p0, __p1) __extension__ ({ \
25291  int8x8_t __s1 = __p1; \
25292  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25293  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
25294 })
25295 #endif
25296 
25297 #ifdef __LITTLE_ENDIAN__
25298 #define vst1_f32(__p0, __p1) __extension__ ({ \
25299  float32x2_t __s1 = __p1; \
25300  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
25301 })
25302 #else
25303 #define vst1_f32(__p0, __p1) __extension__ ({ \
25304  float32x2_t __s1 = __p1; \
25305  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25306  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
25307 })
25308 #endif
25309 
25310 #ifdef __LITTLE_ENDIAN__
25311 #define vst1_f16(__p0, __p1) __extension__ ({ \
25312  float16x4_t __s1 = __p1; \
25313  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
25314 })
25315 #else
25316 #define vst1_f16(__p0, __p1) __extension__ ({ \
25317  float16x4_t __s1 = __p1; \
25318  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25319  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
25320 })
25321 #endif
25322 
25323 #ifdef __LITTLE_ENDIAN__
25324 #define vst1_s32(__p0, __p1) __extension__ ({ \
25325  int32x2_t __s1 = __p1; \
25326  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
25327 })
25328 #else
25329 #define vst1_s32(__p0, __p1) __extension__ ({ \
25330  int32x2_t __s1 = __p1; \
25331  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25332  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
25333 })
25334 #endif
25335 
25336 #ifdef __LITTLE_ENDIAN__
25337 #define vst1_s64(__p0, __p1) __extension__ ({ \
25338  int64x1_t __s1 = __p1; \
25339  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25340 })
25341 #else
25342 #define vst1_s64(__p0, __p1) __extension__ ({ \
25343  int64x1_t __s1 = __p1; \
25344  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25345 })
25346 #endif
25347 
25348 #ifdef __LITTLE_ENDIAN__
25349 #define vst1_s16(__p0, __p1) __extension__ ({ \
25350  int16x4_t __s1 = __p1; \
25351  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
25352 })
25353 #else
25354 #define vst1_s16(__p0, __p1) __extension__ ({ \
25355  int16x4_t __s1 = __p1; \
25356  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25357  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
25358 })
25359 #endif
25360 
25361 #ifdef __LITTLE_ENDIAN__
25362 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25363  poly8x8_t __s1 = __p1; \
25364  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
25365 })
25366 #else
25367 #define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25368  poly8x8_t __s1 = __p1; \
25369  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25370  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
25371 })
25372 #endif
25373 
25374 #ifdef __LITTLE_ENDIAN__
25375 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25376  poly16x4_t __s1 = __p1; \
25377  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
25378 })
25379 #else
25380 #define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25381  poly16x4_t __s1 = __p1; \
25382  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25383  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
25384 })
25385 #endif
25386 
25387 #ifdef __LITTLE_ENDIAN__
25388 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25389  poly8x16_t __s1 = __p1; \
25390  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
25391 })
25392 #else
25393 #define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25394  poly8x16_t __s1 = __p1; \
25395  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25396  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
25397 })
25398 #endif
25399 
25400 #ifdef __LITTLE_ENDIAN__
25401 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25402  poly16x8_t __s1 = __p1; \
25403  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
25404 })
25405 #else
25406 #define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25407  poly16x8_t __s1 = __p1; \
25408  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25409  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
25410 })
25411 #endif
25412 
25413 #ifdef __LITTLE_ENDIAN__
25414 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25415  uint8x16_t __s1 = __p1; \
25416  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
25417 })
25418 #else
25419 #define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25420  uint8x16_t __s1 = __p1; \
25421  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25422  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
25423 })
25424 #endif
25425 
25426 #ifdef __LITTLE_ENDIAN__
25427 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25428  uint32x4_t __s1 = __p1; \
25429  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
25430 })
25431 #else
25432 #define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25433  uint32x4_t __s1 = __p1; \
25434  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25435  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
25436 })
25437 #endif
25438 
25439 #ifdef __LITTLE_ENDIAN__
25440 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25441  uint64x2_t __s1 = __p1; \
25442  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
25443 })
25444 #else
25445 #define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25446  uint64x2_t __s1 = __p1; \
25447  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25448  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
25449 })
25450 #endif
25451 
25452 #ifdef __LITTLE_ENDIAN__
25453 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25454  uint16x8_t __s1 = __p1; \
25455  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
25456 })
25457 #else
25458 #define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25459  uint16x8_t __s1 = __p1; \
25460  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25461  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
25462 })
25463 #endif
25464 
25465 #ifdef __LITTLE_ENDIAN__
25466 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25467  int8x16_t __s1 = __p1; \
25468  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
25469 })
25470 #else
25471 #define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25472  int8x16_t __s1 = __p1; \
25473  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25474  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
25475 })
25476 #endif
25477 
25478 #ifdef __LITTLE_ENDIAN__
25479 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25480  float32x4_t __s1 = __p1; \
25481  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
25482 })
25483 #else
25484 #define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25485  float32x4_t __s1 = __p1; \
25486  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25487  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
25488 })
25489 #endif
25490 
25491 #ifdef __LITTLE_ENDIAN__
25492 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25493  float16x8_t __s1 = __p1; \
25494  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
25495 })
25496 #else
25497 #define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25498  float16x8_t __s1 = __p1; \
25499  float16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25500  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
25501 })
25502 #endif
25503 
25504 #ifdef __LITTLE_ENDIAN__
25505 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25506  int32x4_t __s1 = __p1; \
25507  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
25508 })
25509 #else
25510 #define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25511  int32x4_t __s1 = __p1; \
25512  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25513  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
25514 })
25515 #endif
25516 
25517 #ifdef __LITTLE_ENDIAN__
25518 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25519  int64x2_t __s1 = __p1; \
25520  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
25521 })
25522 #else
25523 #define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25524  int64x2_t __s1 = __p1; \
25525  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25526  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
25527 })
25528 #endif
25529 
25530 #ifdef __LITTLE_ENDIAN__
25531 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25532  int16x8_t __s1 = __p1; \
25533  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
25534 })
25535 #else
25536 #define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25537  int16x8_t __s1 = __p1; \
25538  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25539  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
25540 })
25541 #endif
25542 
25543 #ifdef __LITTLE_ENDIAN__
25544 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25545  uint8x8_t __s1 = __p1; \
25546  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
25547 })
25548 #else
25549 #define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25550  uint8x8_t __s1 = __p1; \
25551  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25552  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
25553 })
25554 #endif
25555 
25556 #ifdef __LITTLE_ENDIAN__
25557 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25558  uint32x2_t __s1 = __p1; \
25559  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
25560 })
25561 #else
25562 #define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25563  uint32x2_t __s1 = __p1; \
25564  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25565  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
25566 })
25567 #endif
25568 
25569 #ifdef __LITTLE_ENDIAN__
25570 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25571  uint64x1_t __s1 = __p1; \
25572  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25573 })
25574 #else
25575 #define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25576  uint64x1_t __s1 = __p1; \
25577  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25578 })
25579 #endif
25580 
25581 #ifdef __LITTLE_ENDIAN__
25582 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25583  uint16x4_t __s1 = __p1; \
25584  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
25585 })
25586 #else
25587 #define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25588  uint16x4_t __s1 = __p1; \
25589  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25590  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
25591 })
25592 #endif
25593 
25594 #ifdef __LITTLE_ENDIAN__
25595 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25596  int8x8_t __s1 = __p1; \
25597  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
25598 })
25599 #else
25600 #define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25601  int8x8_t __s1 = __p1; \
25602  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25603  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
25604 })
25605 #endif
25606 
25607 #ifdef __LITTLE_ENDIAN__
25608 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25609  float32x2_t __s1 = __p1; \
25610  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
25611 })
25612 #else
25613 #define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25614  float32x2_t __s1 = __p1; \
25615  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25616  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
25617 })
25618 #endif
25619 
25620 #ifdef __LITTLE_ENDIAN__
25621 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25622  float16x4_t __s1 = __p1; \
25623  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
25624 })
25625 #else
25626 #define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25627  float16x4_t __s1 = __p1; \
25628  float16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25629  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
25630 })
25631 #endif
25632 
25633 #ifdef __LITTLE_ENDIAN__
25634 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25635  int32x2_t __s1 = __p1; \
25636  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
25637 })
25638 #else
25639 #define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25640  int32x2_t __s1 = __p1; \
25641  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25642  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
25643 })
25644 #endif
25645 
25646 #ifdef __LITTLE_ENDIAN__
25647 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25648  int64x1_t __s1 = __p1; \
25649  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25650 })
25651 #else
25652 #define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25653  int64x1_t __s1 = __p1; \
25654  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25655 })
25656 #endif
25657 
25658 #ifdef __LITTLE_ENDIAN__
25659 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25660  int16x4_t __s1 = __p1; \
25661  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
25662 })
25663 #else
25664 #define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25665  int16x4_t __s1 = __p1; \
25666  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25667  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
25668 })
25669 #endif
25670 
25671 #ifdef __LITTLE_ENDIAN__
25672 #define vst2_p8(__p0, __p1) __extension__ ({ \
25673  poly8x8x2_t __s1 = __p1; \
25674  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
25675 })
25676 #else
25677 #define vst2_p8(__p0, __p1) __extension__ ({ \
25678  poly8x8x2_t __s1 = __p1; \
25679  poly8x8x2_t __rev1; \
25680  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25681  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25682  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
25683 })
25684 #endif
25685 
25686 #ifdef __LITTLE_ENDIAN__
25687 #define vst2_p16(__p0, __p1) __extension__ ({ \
25688  poly16x4x2_t __s1 = __p1; \
25689  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
25690 })
25691 #else
25692 #define vst2_p16(__p0, __p1) __extension__ ({ \
25693  poly16x4x2_t __s1 = __p1; \
25694  poly16x4x2_t __rev1; \
25695  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25696  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25697  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
25698 })
25699 #endif
25700 
25701 #ifdef __LITTLE_ENDIAN__
25702 #define vst2q_p8(__p0, __p1) __extension__ ({ \
25703  poly8x16x2_t __s1 = __p1; \
25704  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
25705 })
25706 #else
25707 #define vst2q_p8(__p0, __p1) __extension__ ({ \
25708  poly8x16x2_t __s1 = __p1; \
25709  poly8x16x2_t __rev1; \
25710  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25711  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25712  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
25713 })
25714 #endif
25715 
25716 #ifdef __LITTLE_ENDIAN__
25717 #define vst2q_p16(__p0, __p1) __extension__ ({ \
25718  poly16x8x2_t __s1 = __p1; \
25719  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
25720 })
25721 #else
25722 #define vst2q_p16(__p0, __p1) __extension__ ({ \
25723  poly16x8x2_t __s1 = __p1; \
25724  poly16x8x2_t __rev1; \
25725  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25726  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25727  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
25728 })
25729 #endif
25730 
25731 #ifdef __LITTLE_ENDIAN__
25732 #define vst2q_u8(__p0, __p1) __extension__ ({ \
25733  uint8x16x2_t __s1 = __p1; \
25734  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
25735 })
25736 #else
25737 #define vst2q_u8(__p0, __p1) __extension__ ({ \
25738  uint8x16x2_t __s1 = __p1; \
25739  uint8x16x2_t __rev1; \
25740  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25741  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25742  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
25743 })
25744 #endif
25745 
25746 #ifdef __LITTLE_ENDIAN__
25747 #define vst2q_u32(__p0, __p1) __extension__ ({ \
25748  uint32x4x2_t __s1 = __p1; \
25749  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
25750 })
25751 #else
25752 #define vst2q_u32(__p0, __p1) __extension__ ({ \
25753  uint32x4x2_t __s1 = __p1; \
25754  uint32x4x2_t __rev1; \
25755  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25756  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25757  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
25758 })
25759 #endif
25760 
25761 #ifdef __LITTLE_ENDIAN__
25762 #define vst2q_u16(__p0, __p1) __extension__ ({ \
25763  uint16x8x2_t __s1 = __p1; \
25764  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
25765 })
25766 #else
25767 #define vst2q_u16(__p0, __p1) __extension__ ({ \
25768  uint16x8x2_t __s1 = __p1; \
25769  uint16x8x2_t __rev1; \
25770  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25771  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25772  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
25773 })
25774 #endif
25775 
25776 #ifdef __LITTLE_ENDIAN__
25777 #define vst2q_s8(__p0, __p1) __extension__ ({ \
25778  int8x16x2_t __s1 = __p1; \
25779  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
25780 })
25781 #else
25782 #define vst2q_s8(__p0, __p1) __extension__ ({ \
25783  int8x16x2_t __s1 = __p1; \
25784  int8x16x2_t __rev1; \
25785  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25786  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25787  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
25788 })
25789 #endif
25790 
25791 #ifdef __LITTLE_ENDIAN__
25792 #define vst2q_f32(__p0, __p1) __extension__ ({ \
25793  float32x4x2_t __s1 = __p1; \
25794  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 41); \
25795 })
25796 #else
25797 #define vst2q_f32(__p0, __p1) __extension__ ({ \
25798  float32x4x2_t __s1 = __p1; \
25799  float32x4x2_t __rev1; \
25800  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25801  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25802  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
25803 })
25804 #endif
25805 
25806 #ifdef __LITTLE_ENDIAN__
25807 #define vst2q_f16(__p0, __p1) __extension__ ({ \
25808  float16x8x2_t __s1 = __p1; \
25809  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 40); \
25810 })
25811 #else
25812 #define vst2q_f16(__p0, __p1) __extension__ ({ \
25813  float16x8x2_t __s1 = __p1; \
25814  float16x8x2_t __rev1; \
25815  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25816  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25817  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
25818 })
25819 #endif
25820 
25821 #ifdef __LITTLE_ENDIAN__
25822 #define vst2q_s32(__p0, __p1) __extension__ ({ \
25823  int32x4x2_t __s1 = __p1; \
25824  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 34); \
25825 })
25826 #else
25827 #define vst2q_s32(__p0, __p1) __extension__ ({ \
25828  int32x4x2_t __s1 = __p1; \
25829  int32x4x2_t __rev1; \
25830  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25831  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25832  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
25833 })
25834 #endif
25835 
25836 #ifdef __LITTLE_ENDIAN__
25837 #define vst2q_s16(__p0, __p1) __extension__ ({ \
25838  int16x8x2_t __s1 = __p1; \
25839  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 33); \
25840 })
25841 #else
25842 #define vst2q_s16(__p0, __p1) __extension__ ({ \
25843  int16x8x2_t __s1 = __p1; \
25844  int16x8x2_t __rev1; \
25845  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25846  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25847  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
25848 })
25849 #endif
25850 
25851 #ifdef __LITTLE_ENDIAN__
25852 #define vst2_u8(__p0, __p1) __extension__ ({ \
25853  uint8x8x2_t __s1 = __p1; \
25854  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
25855 })
25856 #else
25857 #define vst2_u8(__p0, __p1) __extension__ ({ \
25858  uint8x8x2_t __s1 = __p1; \
25859  uint8x8x2_t __rev1; \
25860  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25861  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25862  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
25863 })
25864 #endif
25865 
25866 #ifdef __LITTLE_ENDIAN__
25867 #define vst2_u32(__p0, __p1) __extension__ ({ \
25868  uint32x2x2_t __s1 = __p1; \
25869  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
25870 })
25871 #else
25872 #define vst2_u32(__p0, __p1) __extension__ ({ \
25873  uint32x2x2_t __s1 = __p1; \
25874  uint32x2x2_t __rev1; \
25875  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25876  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25877  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
25878 })
25879 #endif
25880 
25881 #ifdef __LITTLE_ENDIAN__
25882 #define vst2_u64(__p0, __p1) __extension__ ({ \
25883  uint64x1x2_t __s1 = __p1; \
25884  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25885 })
25886 #else
25887 #define vst2_u64(__p0, __p1) __extension__ ({ \
25888  uint64x1x2_t __s1 = __p1; \
25889  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25890 })
25891 #endif
25892 
25893 #ifdef __LITTLE_ENDIAN__
25894 #define vst2_u16(__p0, __p1) __extension__ ({ \
25895  uint16x4x2_t __s1 = __p1; \
25896  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
25897 })
25898 #else
25899 #define vst2_u16(__p0, __p1) __extension__ ({ \
25900  uint16x4x2_t __s1 = __p1; \
25901  uint16x4x2_t __rev1; \
25902  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25903  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25904  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
25905 })
25906 #endif
25907 
25908 #ifdef __LITTLE_ENDIAN__
25909 #define vst2_s8(__p0, __p1) __extension__ ({ \
25910  int8x8x2_t __s1 = __p1; \
25911  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
25912 })
25913 #else
25914 #define vst2_s8(__p0, __p1) __extension__ ({ \
25915  int8x8x2_t __s1 = __p1; \
25916  int8x8x2_t __rev1; \
25917  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25918  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25919  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
25920 })
25921 #endif
25922 
25923 #ifdef __LITTLE_ENDIAN__
25924 #define vst2_f32(__p0, __p1) __extension__ ({ \
25925  float32x2x2_t __s1 = __p1; \
25926  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 9); \
25927 })
25928 #else
25929 #define vst2_f32(__p0, __p1) __extension__ ({ \
25930  float32x2x2_t __s1 = __p1; \
25931  float32x2x2_t __rev1; \
25932  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25933  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25934  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
25935 })
25936 #endif
25937 
25938 #ifdef __LITTLE_ENDIAN__
25939 #define vst2_f16(__p0, __p1) __extension__ ({ \
25940  float16x4x2_t __s1 = __p1; \
25941  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 8); \
25942 })
25943 #else
25944 #define vst2_f16(__p0, __p1) __extension__ ({ \
25945  float16x4x2_t __s1 = __p1; \
25946  float16x4x2_t __rev1; \
25947  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25948  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25949  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
25950 })
25951 #endif
25952 
25953 #ifdef __LITTLE_ENDIAN__
25954 #define vst2_s32(__p0, __p1) __extension__ ({ \
25955  int32x2x2_t __s1 = __p1; \
25956  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 2); \
25957 })
25958 #else
25959 #define vst2_s32(__p0, __p1) __extension__ ({ \
25960  int32x2x2_t __s1 = __p1; \
25961  int32x2x2_t __rev1; \
25962  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25963  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25964  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
25965 })
25966 #endif
25967 
25968 #ifdef __LITTLE_ENDIAN__
25969 #define vst2_s64(__p0, __p1) __extension__ ({ \
25970  int64x1x2_t __s1 = __p1; \
25971  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25972 })
25973 #else
25974 #define vst2_s64(__p0, __p1) __extension__ ({ \
25975  int64x1x2_t __s1 = __p1; \
25976  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25977 })
25978 #endif
25979 
25980 #ifdef __LITTLE_ENDIAN__
25981 #define vst2_s16(__p0, __p1) __extension__ ({ \
25982  int16x4x2_t __s1 = __p1; \
25983  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 1); \
25984 })
25985 #else
25986 #define vst2_s16(__p0, __p1) __extension__ ({ \
25987  int16x4x2_t __s1 = __p1; \
25988  int16x4x2_t __rev1; \
25989  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25990  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25991  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
25992 })
25993 #endif
25994 
25995 #ifdef __LITTLE_ENDIAN__
25996 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25997  poly8x8x2_t __s1 = __p1; \
25998  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
25999 })
26000 #else
26001 #define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26002  poly8x8x2_t __s1 = __p1; \
26003  poly8x8x2_t __rev1; \
26004  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26005  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26006  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
26007 })
26008 #endif
26009 
26010 #ifdef __LITTLE_ENDIAN__
26011 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26012  poly16x4x2_t __s1 = __p1; \
26013  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
26014 })
26015 #else
26016 #define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26017  poly16x4x2_t __s1 = __p1; \
26018  poly16x4x2_t __rev1; \
26019  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26020  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26021  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
26022 })
26023 #endif
26024 
26025 #ifdef __LITTLE_ENDIAN__
26026 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26027  poly16x8x2_t __s1 = __p1; \
26028  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
26029 })
26030 #else
26031 #define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26032  poly16x8x2_t __s1 = __p1; \
26033  poly16x8x2_t __rev1; \
26034  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26035  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26036  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
26037 })
26038 #endif
26039 
26040 #ifdef __LITTLE_ENDIAN__
26041 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26042  uint32x4x2_t __s1 = __p1; \
26043  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
26044 })
26045 #else
26046 #define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26047  uint32x4x2_t __s1 = __p1; \
26048  uint32x4x2_t __rev1; \
26049  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26050  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26051  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
26052 })
26053 #endif
26054 
26055 #ifdef __LITTLE_ENDIAN__
26056 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26057  uint16x8x2_t __s1 = __p1; \
26058  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
26059 })
26060 #else
26061 #define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26062  uint16x8x2_t __s1 = __p1; \
26063  uint16x8x2_t __rev1; \
26064  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26065  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26066  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
26067 })
26068 #endif
26069 
26070 #ifdef __LITTLE_ENDIAN__
26071 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26072  float32x4x2_t __s1 = __p1; \
26073  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 41); \
26074 })
26075 #else
26076 #define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26077  float32x4x2_t __s1 = __p1; \
26078  float32x4x2_t __rev1; \
26079  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26080  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26081  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
26082 })
26083 #endif
26084 
26085 #ifdef __LITTLE_ENDIAN__
26086 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26087  float16x8x2_t __s1 = __p1; \
26088  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 40); \
26089 })
26090 #else
26091 #define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26092  float16x8x2_t __s1 = __p1; \
26093  float16x8x2_t __rev1; \
26094  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26095  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26096  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
26097 })
26098 #endif
26099 
26100 #ifdef __LITTLE_ENDIAN__
26101 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26102  int32x4x2_t __s1 = __p1; \
26103  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 34); \
26104 })
26105 #else
26106 #define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26107  int32x4x2_t __s1 = __p1; \
26108  int32x4x2_t __rev1; \
26109  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26110  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26111  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
26112 })
26113 #endif
26114 
26115 #ifdef __LITTLE_ENDIAN__
26116 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26117  int16x8x2_t __s1 = __p1; \
26118  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 33); \
26119 })
26120 #else
26121 #define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26122  int16x8x2_t __s1 = __p1; \
26123  int16x8x2_t __rev1; \
26124  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26125  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26126  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
26127 })
26128 #endif
26129 
26130 #ifdef __LITTLE_ENDIAN__
26131 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26132  uint8x8x2_t __s1 = __p1; \
26133  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
26134 })
26135 #else
26136 #define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26137  uint8x8x2_t __s1 = __p1; \
26138  uint8x8x2_t __rev1; \
26139  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26140  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26141  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
26142 })
26143 #endif
26144 
26145 #ifdef __LITTLE_ENDIAN__
26146 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26147  uint32x2x2_t __s1 = __p1; \
26148  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
26149 })
26150 #else
26151 #define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26152  uint32x2x2_t __s1 = __p1; \
26153  uint32x2x2_t __rev1; \
26154  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26155  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26156  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
26157 })
26158 #endif
26159 
26160 #ifdef __LITTLE_ENDIAN__
26161 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26162  uint16x4x2_t __s1 = __p1; \
26163  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
26164 })
26165 #else
26166 #define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26167  uint16x4x2_t __s1 = __p1; \
26168  uint16x4x2_t __rev1; \
26169  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26170  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26171  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
26172 })
26173 #endif
26174 
26175 #ifdef __LITTLE_ENDIAN__
26176 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26177  int8x8x2_t __s1 = __p1; \
26178  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
26179 })
26180 #else
26181 #define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26182  int8x8x2_t __s1 = __p1; \
26183  int8x8x2_t __rev1; \
26184  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26185  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26186  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
26187 })
26188 #endif
26189 
26190 #ifdef __LITTLE_ENDIAN__
26191 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26192  float32x2x2_t __s1 = __p1; \
26193  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 9); \
26194 })
26195 #else
26196 #define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26197  float32x2x2_t __s1 = __p1; \
26198  float32x2x2_t __rev1; \
26199  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26200  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26201  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
26202 })
26203 #endif
26204 
26205 #ifdef __LITTLE_ENDIAN__
26206 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26207  float16x4x2_t __s1 = __p1; \
26208  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 8); \
26209 })
26210 #else
26211 #define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26212  float16x4x2_t __s1 = __p1; \
26213  float16x4x2_t __rev1; \
26214  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26215  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26216  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
26217 })
26218 #endif
26219 
26220 #ifdef __LITTLE_ENDIAN__
26221 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26222  int32x2x2_t __s1 = __p1; \
26223  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 2); \
26224 })
26225 #else
26226 #define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26227  int32x2x2_t __s1 = __p1; \
26228  int32x2x2_t __rev1; \
26229  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26230  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26231  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
26232 })
26233 #endif
26234 
26235 #ifdef __LITTLE_ENDIAN__
26236 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26237  int16x4x2_t __s1 = __p1; \
26238  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 1); \
26239 })
26240 #else
26241 #define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26242  int16x4x2_t __s1 = __p1; \
26243  int16x4x2_t __rev1; \
26244  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26245  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26246  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
26247 })
26248 #endif
26249 
26250 #ifdef __LITTLE_ENDIAN__
26251 #define vst3_p8(__p0, __p1) __extension__ ({ \
26252  poly8x8x3_t __s1 = __p1; \
26253  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
26254 })
26255 #else
26256 #define vst3_p8(__p0, __p1) __extension__ ({ \
26257  poly8x8x3_t __s1 = __p1; \
26258  poly8x8x3_t __rev1; \
26259  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26260  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26261  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26262  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
26263 })
26264 #endif
26265 
26266 #ifdef __LITTLE_ENDIAN__
26267 #define vst3_p16(__p0, __p1) __extension__ ({ \
26268  poly16x4x3_t __s1 = __p1; \
26269  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
26270 })
26271 #else
26272 #define vst3_p16(__p0, __p1) __extension__ ({ \
26273  poly16x4x3_t __s1 = __p1; \
26274  poly16x4x3_t __rev1; \
26275  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26276  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26277  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26278  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
26279 })
26280 #endif
26281 
26282 #ifdef __LITTLE_ENDIAN__
26283 #define vst3q_p8(__p0, __p1) __extension__ ({ \
26284  poly8x16x3_t __s1 = __p1; \
26285  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
26286 })
26287 #else
26288 #define vst3q_p8(__p0, __p1) __extension__ ({ \
26289  poly8x16x3_t __s1 = __p1; \
26290  poly8x16x3_t __rev1; \
26291  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26292  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26293  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26294  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
26295 })
26296 #endif
26297 
26298 #ifdef __LITTLE_ENDIAN__
26299 #define vst3q_p16(__p0, __p1) __extension__ ({ \
26300  poly16x8x3_t __s1 = __p1; \
26301  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
26302 })
26303 #else
26304 #define vst3q_p16(__p0, __p1) __extension__ ({ \
26305  poly16x8x3_t __s1 = __p1; \
26306  poly16x8x3_t __rev1; \
26307  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26308  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26309  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26310  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
26311 })
26312 #endif
26313 
26314 #ifdef __LITTLE_ENDIAN__
26315 #define vst3q_u8(__p0, __p1) __extension__ ({ \
26316  uint8x16x3_t __s1 = __p1; \
26317  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
26318 })
26319 #else
26320 #define vst3q_u8(__p0, __p1) __extension__ ({ \
26321  uint8x16x3_t __s1 = __p1; \
26322  uint8x16x3_t __rev1; \
26323  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26324  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26325  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26326  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
26327 })
26328 #endif
26329 
26330 #ifdef __LITTLE_ENDIAN__
26331 #define vst3q_u32(__p0, __p1) __extension__ ({ \
26332  uint32x4x3_t __s1 = __p1; \
26333  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
26334 })
26335 #else
26336 #define vst3q_u32(__p0, __p1) __extension__ ({ \
26337  uint32x4x3_t __s1 = __p1; \
26338  uint32x4x3_t __rev1; \
26339  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26340  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26341  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26342  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
26343 })
26344 #endif
26345 
26346 #ifdef __LITTLE_ENDIAN__
26347 #define vst3q_u16(__p0, __p1) __extension__ ({ \
26348  uint16x8x3_t __s1 = __p1; \
26349  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
26350 })
26351 #else
26352 #define vst3q_u16(__p0, __p1) __extension__ ({ \
26353  uint16x8x3_t __s1 = __p1; \
26354  uint16x8x3_t __rev1; \
26355  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26356  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26357  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26358  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
26359 })
26360 #endif
26361 
26362 #ifdef __LITTLE_ENDIAN__
26363 #define vst3q_s8(__p0, __p1) __extension__ ({ \
26364  int8x16x3_t __s1 = __p1; \
26365  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
26366 })
26367 #else
26368 #define vst3q_s8(__p0, __p1) __extension__ ({ \
26369  int8x16x3_t __s1 = __p1; \
26370  int8x16x3_t __rev1; \
26371  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26372  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26373  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26374  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
26375 })
26376 #endif
26377 
26378 #ifdef __LITTLE_ENDIAN__
26379 #define vst3q_f32(__p0, __p1) __extension__ ({ \
26380  float32x4x3_t __s1 = __p1; \
26381  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
26382 })
26383 #else
26384 #define vst3q_f32(__p0, __p1) __extension__ ({ \
26385  float32x4x3_t __s1 = __p1; \
26386  float32x4x3_t __rev1; \
26387  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26388  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26389  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26390  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
26391 })
26392 #endif
26393 
26394 #ifdef __LITTLE_ENDIAN__
26395 #define vst3q_f16(__p0, __p1) __extension__ ({ \
26396  float16x8x3_t __s1 = __p1; \
26397  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
26398 })
26399 #else
26400 #define vst3q_f16(__p0, __p1) __extension__ ({ \
26401  float16x8x3_t __s1 = __p1; \
26402  float16x8x3_t __rev1; \
26403  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26404  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26405  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26406  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
26407 })
26408 #endif
26409 
26410 #ifdef __LITTLE_ENDIAN__
26411 #define vst3q_s32(__p0, __p1) __extension__ ({ \
26412  int32x4x3_t __s1 = __p1; \
26413  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
26414 })
26415 #else
26416 #define vst3q_s32(__p0, __p1) __extension__ ({ \
26417  int32x4x3_t __s1 = __p1; \
26418  int32x4x3_t __rev1; \
26419  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26420  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26421  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26422  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
26423 })
26424 #endif
26425 
26426 #ifdef __LITTLE_ENDIAN__
26427 #define vst3q_s16(__p0, __p1) __extension__ ({ \
26428  int16x8x3_t __s1 = __p1; \
26429  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
26430 })
26431 #else
26432 #define vst3q_s16(__p0, __p1) __extension__ ({ \
26433  int16x8x3_t __s1 = __p1; \
26434  int16x8x3_t __rev1; \
26435  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26436  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26437  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26438  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
26439 })
26440 #endif
26441 
26442 #ifdef __LITTLE_ENDIAN__
26443 #define vst3_u8(__p0, __p1) __extension__ ({ \
26444  uint8x8x3_t __s1 = __p1; \
26445  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
26446 })
26447 #else
26448 #define vst3_u8(__p0, __p1) __extension__ ({ \
26449  uint8x8x3_t __s1 = __p1; \
26450  uint8x8x3_t __rev1; \
26451  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26452  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26453  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26454  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
26455 })
26456 #endif
26457 
26458 #ifdef __LITTLE_ENDIAN__
26459 #define vst3_u32(__p0, __p1) __extension__ ({ \
26460  uint32x2x3_t __s1 = __p1; \
26461  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
26462 })
26463 #else
26464 #define vst3_u32(__p0, __p1) __extension__ ({ \
26465  uint32x2x3_t __s1 = __p1; \
26466  uint32x2x3_t __rev1; \
26467  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26468  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26469  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26470  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
26471 })
26472 #endif
26473 
26474 #ifdef __LITTLE_ENDIAN__
26475 #define vst3_u64(__p0, __p1) __extension__ ({ \
26476  uint64x1x3_t __s1 = __p1; \
26477  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26478 })
26479 #else
26480 #define vst3_u64(__p0, __p1) __extension__ ({ \
26481  uint64x1x3_t __s1 = __p1; \
26482  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26483 })
26484 #endif
26485 
26486 #ifdef __LITTLE_ENDIAN__
26487 #define vst3_u16(__p0, __p1) __extension__ ({ \
26488  uint16x4x3_t __s1 = __p1; \
26489  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
26490 })
26491 #else
26492 #define vst3_u16(__p0, __p1) __extension__ ({ \
26493  uint16x4x3_t __s1 = __p1; \
26494  uint16x4x3_t __rev1; \
26495  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26496  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26497  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26498  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
26499 })
26500 #endif
26501 
26502 #ifdef __LITTLE_ENDIAN__
26503 #define vst3_s8(__p0, __p1) __extension__ ({ \
26504  int8x8x3_t __s1 = __p1; \
26505  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
26506 })
26507 #else
26508 #define vst3_s8(__p0, __p1) __extension__ ({ \
26509  int8x8x3_t __s1 = __p1; \
26510  int8x8x3_t __rev1; \
26511  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26512  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26513  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26514  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
26515 })
26516 #endif
26517 
26518 #ifdef __LITTLE_ENDIAN__
26519 #define vst3_f32(__p0, __p1) __extension__ ({ \
26520  float32x2x3_t __s1 = __p1; \
26521  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
26522 })
26523 #else
26524 #define vst3_f32(__p0, __p1) __extension__ ({ \
26525  float32x2x3_t __s1 = __p1; \
26526  float32x2x3_t __rev1; \
26527  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26528  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26529  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26530  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
26531 })
26532 #endif
26533 
26534 #ifdef __LITTLE_ENDIAN__
26535 #define vst3_f16(__p0, __p1) __extension__ ({ \
26536  float16x4x3_t __s1 = __p1; \
26537  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
26538 })
26539 #else
26540 #define vst3_f16(__p0, __p1) __extension__ ({ \
26541  float16x4x3_t __s1 = __p1; \
26542  float16x4x3_t __rev1; \
26543  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26544  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26545  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26546  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
26547 })
26548 #endif
26549 
26550 #ifdef __LITTLE_ENDIAN__
26551 #define vst3_s32(__p0, __p1) __extension__ ({ \
26552  int32x2x3_t __s1 = __p1; \
26553  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
26554 })
26555 #else
26556 #define vst3_s32(__p0, __p1) __extension__ ({ \
26557  int32x2x3_t __s1 = __p1; \
26558  int32x2x3_t __rev1; \
26559  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26560  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26561  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26562  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
26563 })
26564 #endif
26565 
26566 #ifdef __LITTLE_ENDIAN__
26567 #define vst3_s64(__p0, __p1) __extension__ ({ \
26568  int64x1x3_t __s1 = __p1; \
26569  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26570 })
26571 #else
26572 #define vst3_s64(__p0, __p1) __extension__ ({ \
26573  int64x1x3_t __s1 = __p1; \
26574  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26575 })
26576 #endif
26577 
26578 #ifdef __LITTLE_ENDIAN__
26579 #define vst3_s16(__p0, __p1) __extension__ ({ \
26580  int16x4x3_t __s1 = __p1; \
26581  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
26582 })
26583 #else
26584 #define vst3_s16(__p0, __p1) __extension__ ({ \
26585  int16x4x3_t __s1 = __p1; \
26586  int16x4x3_t __rev1; \
26587  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26588  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26589  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26590  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
26591 })
26592 #endif
26593 
26594 #ifdef __LITTLE_ENDIAN__
26595 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26596  poly8x8x3_t __s1 = __p1; \
26597  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
26598 })
26599 #else
26600 #define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26601  poly8x8x3_t __s1 = __p1; \
26602  poly8x8x3_t __rev1; \
26603  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26604  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26605  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26606  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
26607 })
26608 #endif
26609 
26610 #ifdef __LITTLE_ENDIAN__
26611 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26612  poly16x4x3_t __s1 = __p1; \
26613  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
26614 })
26615 #else
26616 #define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26617  poly16x4x3_t __s1 = __p1; \
26618  poly16x4x3_t __rev1; \
26619  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26620  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26621  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26622  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
26623 })
26624 #endif
26625 
26626 #ifdef __LITTLE_ENDIAN__
26627 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26628  poly16x8x3_t __s1 = __p1; \
26629  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
26630 })
26631 #else
26632 #define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26633  poly16x8x3_t __s1 = __p1; \
26634  poly16x8x3_t __rev1; \
26635  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26636  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26637  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26638  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
26639 })
26640 #endif
26641 
26642 #ifdef __LITTLE_ENDIAN__
26643 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26644  uint32x4x3_t __s1 = __p1; \
26645  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
26646 })
26647 #else
26648 #define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26649  uint32x4x3_t __s1 = __p1; \
26650  uint32x4x3_t __rev1; \
26651  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26652  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26653  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26654  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
26655 })
26656 #endif
26657 
26658 #ifdef __LITTLE_ENDIAN__
26659 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26660  uint16x8x3_t __s1 = __p1; \
26661  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
26662 })
26663 #else
26664 #define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26665  uint16x8x3_t __s1 = __p1; \
26666  uint16x8x3_t __rev1; \
26667  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26668  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26669  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26670  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
26671 })
26672 #endif
26673 
26674 #ifdef __LITTLE_ENDIAN__
26675 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26676  float32x4x3_t __s1 = __p1; \
26677  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
26678 })
26679 #else
26680 #define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26681  float32x4x3_t __s1 = __p1; \
26682  float32x4x3_t __rev1; \
26683  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26684  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26685  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26686  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
26687 })
26688 #endif
26689 
26690 #ifdef __LITTLE_ENDIAN__
26691 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26692  float16x8x3_t __s1 = __p1; \
26693  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
26694 })
26695 #else
26696 #define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26697  float16x8x3_t __s1 = __p1; \
26698  float16x8x3_t __rev1; \
26699  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26700  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26701  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26702  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
26703 })
26704 #endif
26705 
26706 #ifdef __LITTLE_ENDIAN__
26707 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26708  int32x4x3_t __s1 = __p1; \
26709  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
26710 })
26711 #else
26712 #define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26713  int32x4x3_t __s1 = __p1; \
26714  int32x4x3_t __rev1; \
26715  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26716  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26717  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26718  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
26719 })
26720 #endif
26721 
26722 #ifdef __LITTLE_ENDIAN__
26723 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26724  int16x8x3_t __s1 = __p1; \
26725  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
26726 })
26727 #else
26728 #define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26729  int16x8x3_t __s1 = __p1; \
26730  int16x8x3_t __rev1; \
26731  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26732  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26733  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26734  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
26735 })
26736 #endif
26737 
26738 #ifdef __LITTLE_ENDIAN__
26739 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26740  uint8x8x3_t __s1 = __p1; \
26741  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
26742 })
26743 #else
26744 #define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26745  uint8x8x3_t __s1 = __p1; \
26746  uint8x8x3_t __rev1; \
26747  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26748  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26749  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26750  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
26751 })
26752 #endif
26753 
26754 #ifdef __LITTLE_ENDIAN__
26755 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26756  uint32x2x3_t __s1 = __p1; \
26757  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
26758 })
26759 #else
26760 #define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26761  uint32x2x3_t __s1 = __p1; \
26762  uint32x2x3_t __rev1; \
26763  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26764  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26765  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26766  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
26767 })
26768 #endif
26769 
26770 #ifdef __LITTLE_ENDIAN__
26771 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26772  uint16x4x3_t __s1 = __p1; \
26773  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
26774 })
26775 #else
26776 #define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26777  uint16x4x3_t __s1 = __p1; \
26778  uint16x4x3_t __rev1; \
26779  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26780  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26781  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26782  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
26783 })
26784 #endif
26785 
26786 #ifdef __LITTLE_ENDIAN__
26787 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26788  int8x8x3_t __s1 = __p1; \
26789  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
26790 })
26791 #else
26792 #define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26793  int8x8x3_t __s1 = __p1; \
26794  int8x8x3_t __rev1; \
26795  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26796  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26797  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26798  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
26799 })
26800 #endif
26801 
26802 #ifdef __LITTLE_ENDIAN__
26803 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26804  float32x2x3_t __s1 = __p1; \
26805  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
26806 })
26807 #else
26808 #define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26809  float32x2x3_t __s1 = __p1; \
26810  float32x2x3_t __rev1; \
26811  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26812  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26813  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26814  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
26815 })
26816 #endif
26817 
26818 #ifdef __LITTLE_ENDIAN__
26819 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26820  float16x4x3_t __s1 = __p1; \
26821  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
26822 })
26823 #else
26824 #define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26825  float16x4x3_t __s1 = __p1; \
26826  float16x4x3_t __rev1; \
26827  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26828  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26829  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26830  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
26831 })
26832 #endif
26833 
26834 #ifdef __LITTLE_ENDIAN__
26835 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26836  int32x2x3_t __s1 = __p1; \
26837  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
26838 })
26839 #else
26840 #define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26841  int32x2x3_t __s1 = __p1; \
26842  int32x2x3_t __rev1; \
26843  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26844  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26845  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26846  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
26847 })
26848 #endif
26849 
26850 #ifdef __LITTLE_ENDIAN__
26851 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26852  int16x4x3_t __s1 = __p1; \
26853  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
26854 })
26855 #else
26856 #define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26857  int16x4x3_t __s1 = __p1; \
26858  int16x4x3_t __rev1; \
26859  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26860  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26861  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26862  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
26863 })
26864 #endif
26865 
26866 #ifdef __LITTLE_ENDIAN__
26867 #define vst4_p8(__p0, __p1) __extension__ ({ \
26868  poly8x8x4_t __s1 = __p1; \
26869  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
26870 })
26871 #else
26872 #define vst4_p8(__p0, __p1) __extension__ ({ \
26873  poly8x8x4_t __s1 = __p1; \
26874  poly8x8x4_t __rev1; \
26875  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26876  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26877  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26878  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26879  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
26880 })
26881 #endif
26882 
26883 #ifdef __LITTLE_ENDIAN__
26884 #define vst4_p16(__p0, __p1) __extension__ ({ \
26885  poly16x4x4_t __s1 = __p1; \
26886  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
26887 })
26888 #else
26889 #define vst4_p16(__p0, __p1) __extension__ ({ \
26890  poly16x4x4_t __s1 = __p1; \
26891  poly16x4x4_t __rev1; \
26892  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26893  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26894  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26895  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26896  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
26897 })
26898 #endif
26899 
26900 #ifdef __LITTLE_ENDIAN__
26901 #define vst4q_p8(__p0, __p1) __extension__ ({ \
26902  poly8x16x4_t __s1 = __p1; \
26903  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
26904 })
26905 #else
26906 #define vst4q_p8(__p0, __p1) __extension__ ({ \
26907  poly8x16x4_t __s1 = __p1; \
26908  poly8x16x4_t __rev1; \
26909  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26910  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26911  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26912  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26913  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
26914 })
26915 #endif
26916 
26917 #ifdef __LITTLE_ENDIAN__
26918 #define vst4q_p16(__p0, __p1) __extension__ ({ \
26919  poly16x8x4_t __s1 = __p1; \
26920  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
26921 })
26922 #else
26923 #define vst4q_p16(__p0, __p1) __extension__ ({ \
26924  poly16x8x4_t __s1 = __p1; \
26925  poly16x8x4_t __rev1; \
26926  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26927  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26928  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26929  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26930  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
26931 })
26932 #endif
26933 
26934 #ifdef __LITTLE_ENDIAN__
26935 #define vst4q_u8(__p0, __p1) __extension__ ({ \
26936  uint8x16x4_t __s1 = __p1; \
26937  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
26938 })
26939 #else
26940 #define vst4q_u8(__p0, __p1) __extension__ ({ \
26941  uint8x16x4_t __s1 = __p1; \
26942  uint8x16x4_t __rev1; \
26943  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26944  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26945  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26946  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26947  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
26948 })
26949 #endif
26950 
26951 #ifdef __LITTLE_ENDIAN__
26952 #define vst4q_u32(__p0, __p1) __extension__ ({ \
26953  uint32x4x4_t __s1 = __p1; \
26954  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
26955 })
26956 #else
26957 #define vst4q_u32(__p0, __p1) __extension__ ({ \
26958  uint32x4x4_t __s1 = __p1; \
26959  uint32x4x4_t __rev1; \
26960  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26961  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26962  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26963  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26964  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
26965 })
26966 #endif
26967 
26968 #ifdef __LITTLE_ENDIAN__
26969 #define vst4q_u16(__p0, __p1) __extension__ ({ \
26970  uint16x8x4_t __s1 = __p1; \
26971  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
26972 })
26973 #else
26974 #define vst4q_u16(__p0, __p1) __extension__ ({ \
26975  uint16x8x4_t __s1 = __p1; \
26976  uint16x8x4_t __rev1; \
26977  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26978  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26979  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26980  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26981  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
26982 })
26983 #endif
26984 
26985 #ifdef __LITTLE_ENDIAN__
26986 #define vst4q_s8(__p0, __p1) __extension__ ({ \
26987  int8x16x4_t __s1 = __p1; \
26988  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
26989 })
26990 #else
26991 #define vst4q_s8(__p0, __p1) __extension__ ({ \
26992  int8x16x4_t __s1 = __p1; \
26993  int8x16x4_t __rev1; \
26994  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26995  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26996  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26997  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26998  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
26999 })
27000 #endif
27001 
27002 #ifdef __LITTLE_ENDIAN__
27003 #define vst4q_f32(__p0, __p1) __extension__ ({ \
27004  float32x4x4_t __s1 = __p1; \
27005  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
27006 })
27007 #else
27008 #define vst4q_f32(__p0, __p1) __extension__ ({ \
27009  float32x4x4_t __s1 = __p1; \
27010  float32x4x4_t __rev1; \
27011  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27012  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27013  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27014  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27015  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
27016 })
27017 #endif
27018 
27019 #ifdef __LITTLE_ENDIAN__
27020 #define vst4q_f16(__p0, __p1) __extension__ ({ \
27021  float16x8x4_t __s1 = __p1; \
27022  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
27023 })
27024 #else
27025 #define vst4q_f16(__p0, __p1) __extension__ ({ \
27026  float16x8x4_t __s1 = __p1; \
27027  float16x8x4_t __rev1; \
27028  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27029  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27030  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27031  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27032  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
27033 })
27034 #endif
27035 
27036 #ifdef __LITTLE_ENDIAN__
27037 #define vst4q_s32(__p0, __p1) __extension__ ({ \
27038  int32x4x4_t __s1 = __p1; \
27039  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
27040 })
27041 #else
27042 #define vst4q_s32(__p0, __p1) __extension__ ({ \
27043  int32x4x4_t __s1 = __p1; \
27044  int32x4x4_t __rev1; \
27045  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27046  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27047  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27048  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27049  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
27050 })
27051 #endif
27052 
27053 #ifdef __LITTLE_ENDIAN__
27054 #define vst4q_s16(__p0, __p1) __extension__ ({ \
27055  int16x8x4_t __s1 = __p1; \
27056  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
27057 })
27058 #else
27059 #define vst4q_s16(__p0, __p1) __extension__ ({ \
27060  int16x8x4_t __s1 = __p1; \
27061  int16x8x4_t __rev1; \
27062  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27063  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27064  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27065  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27066  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
27067 })
27068 #endif
27069 
27070 #ifdef __LITTLE_ENDIAN__
27071 #define vst4_u8(__p0, __p1) __extension__ ({ \
27072  uint8x8x4_t __s1 = __p1; \
27073  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
27074 })
27075 #else
27076 #define vst4_u8(__p0, __p1) __extension__ ({ \
27077  uint8x8x4_t __s1 = __p1; \
27078  uint8x8x4_t __rev1; \
27079  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27080  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27081  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27082  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27083  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
27084 })
27085 #endif
27086 
27087 #ifdef __LITTLE_ENDIAN__
27088 #define vst4_u32(__p0, __p1) __extension__ ({ \
27089  uint32x2x4_t __s1 = __p1; \
27090  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27091 })
27092 #else
27093 #define vst4_u32(__p0, __p1) __extension__ ({ \
27094  uint32x2x4_t __s1 = __p1; \
27095  uint32x2x4_t __rev1; \
27096  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27097  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27098  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27099  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27100  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27101 })
27102 #endif
27103 
27104 #ifdef __LITTLE_ENDIAN__
27105 #define vst4_u64(__p0, __p1) __extension__ ({ \
27106  uint64x1x4_t __s1 = __p1; \
27107  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27108 })
27109 #else
27110 #define vst4_u64(__p0, __p1) __extension__ ({ \
27111  uint64x1x4_t __s1 = __p1; \
27112  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27113 })
27114 #endif
27115 
27116 #ifdef __LITTLE_ENDIAN__
27117 #define vst4_u16(__p0, __p1) __extension__ ({ \
27118  uint16x4x4_t __s1 = __p1; \
27119  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
27120 })
27121 #else
27122 #define vst4_u16(__p0, __p1) __extension__ ({ \
27123  uint16x4x4_t __s1 = __p1; \
27124  uint16x4x4_t __rev1; \
27125  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27126  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27127  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27128  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27129  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
27130 })
27131 #endif
27132 
27133 #ifdef __LITTLE_ENDIAN__
27134 #define vst4_s8(__p0, __p1) __extension__ ({ \
27135  int8x8x4_t __s1 = __p1; \
27136  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
27137 })
27138 #else
27139 #define vst4_s8(__p0, __p1) __extension__ ({ \
27140  int8x8x4_t __s1 = __p1; \
27141  int8x8x4_t __rev1; \
27142  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27143  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27144  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27145  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27146  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
27147 })
27148 #endif
27149 
27150 #ifdef __LITTLE_ENDIAN__
27151 #define vst4_f32(__p0, __p1) __extension__ ({ \
27152  float32x2x4_t __s1 = __p1; \
27153  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
27154 })
27155 #else
27156 #define vst4_f32(__p0, __p1) __extension__ ({ \
27157  float32x2x4_t __s1 = __p1; \
27158  float32x2x4_t __rev1; \
27159  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27160  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27161  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27162  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27163  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
27164 })
27165 #endif
27166 
27167 #ifdef __LITTLE_ENDIAN__
27168 #define vst4_f16(__p0, __p1) __extension__ ({ \
27169  float16x4x4_t __s1 = __p1; \
27170  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
27171 })
27172 #else
27173 #define vst4_f16(__p0, __p1) __extension__ ({ \
27174  float16x4x4_t __s1 = __p1; \
27175  float16x4x4_t __rev1; \
27176  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27177  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27178  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27179  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27180  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
27181 })
27182 #endif
27183 
27184 #ifdef __LITTLE_ENDIAN__
27185 #define vst4_s32(__p0, __p1) __extension__ ({ \
27186  int32x2x4_t __s1 = __p1; \
27187  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
27188 })
27189 #else
27190 #define vst4_s32(__p0, __p1) __extension__ ({ \
27191  int32x2x4_t __s1 = __p1; \
27192  int32x2x4_t __rev1; \
27193  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27194  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27195  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27196  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27197  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
27198 })
27199 #endif
27200 
27201 #ifdef __LITTLE_ENDIAN__
27202 #define vst4_s64(__p0, __p1) __extension__ ({ \
27203  int64x1x4_t __s1 = __p1; \
27204  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27205 })
27206 #else
27207 #define vst4_s64(__p0, __p1) __extension__ ({ \
27208  int64x1x4_t __s1 = __p1; \
27209  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27210 })
27211 #endif
27212 
27213 #ifdef __LITTLE_ENDIAN__
27214 #define vst4_s16(__p0, __p1) __extension__ ({ \
27215  int16x4x4_t __s1 = __p1; \
27216  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
27217 })
27218 #else
27219 #define vst4_s16(__p0, __p1) __extension__ ({ \
27220  int16x4x4_t __s1 = __p1; \
27221  int16x4x4_t __rev1; \
27222  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27223  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27224  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27225  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27226  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
27227 })
27228 #endif
27229 
27230 #ifdef __LITTLE_ENDIAN__
27231 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27232  poly8x8x4_t __s1 = __p1; \
27233  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
27234 })
27235 #else
27236 #define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27237  poly8x8x4_t __s1 = __p1; \
27238  poly8x8x4_t __rev1; \
27239  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27240  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27241  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27242  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27243  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
27244 })
27245 #endif
27246 
27247 #ifdef __LITTLE_ENDIAN__
27248 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27249  poly16x4x4_t __s1 = __p1; \
27250  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
27251 })
27252 #else
27253 #define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27254  poly16x4x4_t __s1 = __p1; \
27255  poly16x4x4_t __rev1; \
27256  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27257  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27258  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27259  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27260  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
27261 })
27262 #endif
27263 
27264 #ifdef __LITTLE_ENDIAN__
27265 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27266  poly16x8x4_t __s1 = __p1; \
27267  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
27268 })
27269 #else
27270 #define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27271  poly16x8x4_t __s1 = __p1; \
27272  poly16x8x4_t __rev1; \
27273  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27274  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27275  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27276  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27277  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
27278 })
27279 #endif
27280 
27281 #ifdef __LITTLE_ENDIAN__
27282 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27283  uint32x4x4_t __s1 = __p1; \
27284  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
27285 })
27286 #else
27287 #define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27288  uint32x4x4_t __s1 = __p1; \
27289  uint32x4x4_t __rev1; \
27290  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27291  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27292  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27293  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27294  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
27295 })
27296 #endif
27297 
27298 #ifdef __LITTLE_ENDIAN__
27299 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27300  uint16x8x4_t __s1 = __p1; \
27301  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
27302 })
27303 #else
27304 #define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27305  uint16x8x4_t __s1 = __p1; \
27306  uint16x8x4_t __rev1; \
27307  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27308  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27309  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27310  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27311  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
27312 })
27313 #endif
27314 
27315 #ifdef __LITTLE_ENDIAN__
27316 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27317  float32x4x4_t __s1 = __p1; \
27318  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
27319 })
27320 #else
27321 #define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27322  float32x4x4_t __s1 = __p1; \
27323  float32x4x4_t __rev1; \
27324  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27325  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27326  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27327  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27328  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
27329 })
27330 #endif
27331 
27332 #ifdef __LITTLE_ENDIAN__
27333 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27334  float16x8x4_t __s1 = __p1; \
27335  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
27336 })
27337 #else
27338 #define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27339  float16x8x4_t __s1 = __p1; \
27340  float16x8x4_t __rev1; \
27341  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27342  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27343  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27344  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27345  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
27346 })
27347 #endif
27348 
27349 #ifdef __LITTLE_ENDIAN__
27350 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27351  int32x4x4_t __s1 = __p1; \
27352  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
27353 })
27354 #else
27355 #define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27356  int32x4x4_t __s1 = __p1; \
27357  int32x4x4_t __rev1; \
27358  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27359  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27360  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27361  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27362  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
27363 })
27364 #endif
27365 
27366 #ifdef __LITTLE_ENDIAN__
27367 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27368  int16x8x4_t __s1 = __p1; \
27369  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
27370 })
27371 #else
27372 #define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27373  int16x8x4_t __s1 = __p1; \
27374  int16x8x4_t __rev1; \
27375  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27376  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27377  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27378  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27379  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
27380 })
27381 #endif
27382 
27383 #ifdef __LITTLE_ENDIAN__
27384 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27385  uint8x8x4_t __s1 = __p1; \
27386  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
27387 })
27388 #else
27389 #define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27390  uint8x8x4_t __s1 = __p1; \
27391  uint8x8x4_t __rev1; \
27392  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27393  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27394  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27395  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27396  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
27397 })
27398 #endif
27399 
27400 #ifdef __LITTLE_ENDIAN__
27401 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27402  uint32x2x4_t __s1 = __p1; \
27403  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
27404 })
27405 #else
27406 #define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27407  uint32x2x4_t __s1 = __p1; \
27408  uint32x2x4_t __rev1; \
27409  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27410  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27411  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27412  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27413  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
27414 })
27415 #endif
27416 
27417 #ifdef __LITTLE_ENDIAN__
27418 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27419  uint16x4x4_t __s1 = __p1; \
27420  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
27421 })
27422 #else
27423 #define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27424  uint16x4x4_t __s1 = __p1; \
27425  uint16x4x4_t __rev1; \
27426  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27427  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27428  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27429  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27430  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
27431 })
27432 #endif
27433 
27434 #ifdef __LITTLE_ENDIAN__
27435 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27436  int8x8x4_t __s1 = __p1; \
27437  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
27438 })
27439 #else
27440 #define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27441  int8x8x4_t __s1 = __p1; \
27442  int8x8x4_t __rev1; \
27443  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27444  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27445  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27446  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27447  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
27448 })
27449 #endif
27450 
27451 #ifdef __LITTLE_ENDIAN__
27452 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27453  float32x2x4_t __s1 = __p1; \
27454  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
27455 })
27456 #else
27457 #define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27458  float32x2x4_t __s1 = __p1; \
27459  float32x2x4_t __rev1; \
27460  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27461  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27462  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27463  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27464  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
27465 })
27466 #endif
27467 
27468 #ifdef __LITTLE_ENDIAN__
27469 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27470  float16x4x4_t __s1 = __p1; \
27471  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
27472 })
27473 #else
27474 #define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27475  float16x4x4_t __s1 = __p1; \
27476  float16x4x4_t __rev1; \
27477  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27478  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27479  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27480  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27481  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
27482 })
27483 #endif
27484 
27485 #ifdef __LITTLE_ENDIAN__
27486 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27487  int32x2x4_t __s1 = __p1; \
27488  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
27489 })
27490 #else
27491 #define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27492  int32x2x4_t __s1 = __p1; \
27493  int32x2x4_t __rev1; \
27494  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27495  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27496  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27497  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27498  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
27499 })
27500 #endif
27501 
27502 #ifdef __LITTLE_ENDIAN__
27503 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27504  int16x4x4_t __s1 = __p1; \
27505  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
27506 })
27507 #else
27508 #define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27509  int16x4x4_t __s1 = __p1; \
27510  int16x4x4_t __rev1; \
27511  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27512  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27513  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27514  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27515  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
27516 })
27517 #endif
27518 
27519 #ifdef __LITTLE_ENDIAN__
27520 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27521  uint8x16_t __ret;
27522  __ret = __p0 - __p1;
27523  return __ret;
27524 }
27525 #else
27526 __ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27527  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27528  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27529  uint8x16_t __ret;
27530  __ret = __rev0 - __rev1;
27531  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27532  return __ret;
27533 }
27534 #endif
27535 
27536 #ifdef __LITTLE_ENDIAN__
27537 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27538  uint32x4_t __ret;
27539  __ret = __p0 - __p1;
27540  return __ret;
27541 }
27542 #else
27543 __ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27544  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27545  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27546  uint32x4_t __ret;
27547  __ret = __rev0 - __rev1;
27548  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27549  return __ret;
27550 }
27551 #endif
27552 
27553 #ifdef __LITTLE_ENDIAN__
27554 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27555  uint64x2_t __ret;
27556  __ret = __p0 - __p1;
27557  return __ret;
27558 }
27559 #else
27560 __ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27561  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27562  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27563  uint64x2_t __ret;
27564  __ret = __rev0 - __rev1;
27565  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27566  return __ret;
27567 }
27568 #endif
27569 
27570 #ifdef __LITTLE_ENDIAN__
27571 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27572  uint16x8_t __ret;
27573  __ret = __p0 - __p1;
27574  return __ret;
27575 }
27576 #else
27577 __ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27578  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27579  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27580  uint16x8_t __ret;
27581  __ret = __rev0 - __rev1;
27582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27583  return __ret;
27584 }
27585 #endif
27586 
27587 #ifdef __LITTLE_ENDIAN__
27588 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27589  int8x16_t __ret;
27590  __ret = __p0 - __p1;
27591  return __ret;
27592 }
27593 #else
27594 __ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27595  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27596  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27597  int8x16_t __ret;
27598  __ret = __rev0 - __rev1;
27599  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27600  return __ret;
27601 }
27602 #endif
27603 
27604 #ifdef __LITTLE_ENDIAN__
27605 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27606  float32x4_t __ret;
27607  __ret = __p0 - __p1;
27608  return __ret;
27609 }
27610 #else
27611 __ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27612  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27613  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27614  float32x4_t __ret;
27615  __ret = __rev0 - __rev1;
27616  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27617  return __ret;
27618 }
27619 #endif
27620 
27621 #ifdef __LITTLE_ENDIAN__
27622 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27623  int32x4_t __ret;
27624  __ret = __p0 - __p1;
27625  return __ret;
27626 }
27627 #else
27628 __ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27629  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27630  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27631  int32x4_t __ret;
27632  __ret = __rev0 - __rev1;
27633  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27634  return __ret;
27635 }
27636 #endif
27637 
27638 #ifdef __LITTLE_ENDIAN__
27639 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27640  int64x2_t __ret;
27641  __ret = __p0 - __p1;
27642  return __ret;
27643 }
27644 #else
27645 __ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27646  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27647  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27648  int64x2_t __ret;
27649  __ret = __rev0 - __rev1;
27650  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27651  return __ret;
27652 }
27653 #endif
27654 
27655 #ifdef __LITTLE_ENDIAN__
27656 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27657  int16x8_t __ret;
27658  __ret = __p0 - __p1;
27659  return __ret;
27660 }
27661 #else
27662 __ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27663  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27664  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27665  int16x8_t __ret;
27666  __ret = __rev0 - __rev1;
27667  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27668  return __ret;
27669 }
27670 #endif
27671 
27672 #ifdef __LITTLE_ENDIAN__
27673 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27674  uint8x8_t __ret;
27675  __ret = __p0 - __p1;
27676  return __ret;
27677 }
27678 #else
27679 __ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27680  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27681  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27682  uint8x8_t __ret;
27683  __ret = __rev0 - __rev1;
27684  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27685  return __ret;
27686 }
27687 #endif
27688 
27689 #ifdef __LITTLE_ENDIAN__
27690 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27691  uint32x2_t __ret;
27692  __ret = __p0 - __p1;
27693  return __ret;
27694 }
27695 #else
27696 __ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27697  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27698  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27699  uint32x2_t __ret;
27700  __ret = __rev0 - __rev1;
27701  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27702  return __ret;
27703 }
27704 #endif
27705 
27706 #ifdef __LITTLE_ENDIAN__
27707 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27708  uint64x1_t __ret;
27709  __ret = __p0 - __p1;
27710  return __ret;
27711 }
27712 #else
27713 __ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27714  uint64x1_t __ret;
27715  __ret = __p0 - __p1;
27716  return __ret;
27717 }
27718 #endif
27719 
27720 #ifdef __LITTLE_ENDIAN__
27721 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27722  uint16x4_t __ret;
27723  __ret = __p0 - __p1;
27724  return __ret;
27725 }
27726 #else
27727 __ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27728  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27729  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27730  uint16x4_t __ret;
27731  __ret = __rev0 - __rev1;
27732  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27733  return __ret;
27734 }
27735 #endif
27736 
27737 #ifdef __LITTLE_ENDIAN__
27738 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27739  int8x8_t __ret;
27740  __ret = __p0 - __p1;
27741  return __ret;
27742 }
27743 #else
27744 __ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27745  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27746  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27747  int8x8_t __ret;
27748  __ret = __rev0 - __rev1;
27749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27750  return __ret;
27751 }
27752 #endif
27753 
27754 #ifdef __LITTLE_ENDIAN__
27755 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27756  float32x2_t __ret;
27757  __ret = __p0 - __p1;
27758  return __ret;
27759 }
27760 #else
27761 __ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27762  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27763  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27764  float32x2_t __ret;
27765  __ret = __rev0 - __rev1;
27766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27767  return __ret;
27768 }
27769 #endif
27770 
27771 #ifdef __LITTLE_ENDIAN__
27772 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27773  int32x2_t __ret;
27774  __ret = __p0 - __p1;
27775  return __ret;
27776 }
27777 #else
27778 __ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27779  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27780  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27781  int32x2_t __ret;
27782  __ret = __rev0 - __rev1;
27783  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27784  return __ret;
27785 }
27786 #endif
27787 
27788 #ifdef __LITTLE_ENDIAN__
27789 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27790  int64x1_t __ret;
27791  __ret = __p0 - __p1;
27792  return __ret;
27793 }
27794 #else
27795 __ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27796  int64x1_t __ret;
27797  __ret = __p0 - __p1;
27798  return __ret;
27799 }
27800 #endif
27801 
27802 #ifdef __LITTLE_ENDIAN__
27803 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27804  int16x4_t __ret;
27805  __ret = __p0 - __p1;
27806  return __ret;
27807 }
27808 #else
27809 __ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27810  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27811  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27812  int16x4_t __ret;
27813  __ret = __rev0 - __rev1;
27814  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27815  return __ret;
27816 }
27817 #endif
27818 
27819 #ifdef __LITTLE_ENDIAN__
27820 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27821  uint16x4_t __ret;
27822  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27823  return __ret;
27824 }
27825 #else
27826 __ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27827  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27828  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27829  uint16x4_t __ret;
27830  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
27831  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27832  return __ret;
27833 }
27834 __ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27835  uint16x4_t __ret;
27836  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27837  return __ret;
27838 }
27839 #endif
27840 
27841 #ifdef __LITTLE_ENDIAN__
27842 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27843  uint32x2_t __ret;
27844  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27845  return __ret;
27846 }
27847 #else
27848 __ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27849  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27850  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27851  uint32x2_t __ret;
27852  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
27853  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27854  return __ret;
27855 }
27856 __ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27857  uint32x2_t __ret;
27858  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27859  return __ret;
27860 }
27861 #endif
27862 
27863 #ifdef __LITTLE_ENDIAN__
27864 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27865  uint8x8_t __ret;
27866  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27867  return __ret;
27868 }
27869 #else
27870 __ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27871  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27872  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27873  uint8x8_t __ret;
27874  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
27875  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27876  return __ret;
27877 }
27878 __ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27879  uint8x8_t __ret;
27880  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27881  return __ret;
27882 }
27883 #endif
27884 
27885 #ifdef __LITTLE_ENDIAN__
27886 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27887  int16x4_t __ret;
27888  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27889  return __ret;
27890 }
27891 #else
27892 __ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27893  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27894  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27895  int16x4_t __ret;
27896  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
27897  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27898  return __ret;
27899 }
27900 __ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27901  int16x4_t __ret;
27902  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27903  return __ret;
27904 }
27905 #endif
27906 
27907 #ifdef __LITTLE_ENDIAN__
27908 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27909  int32x2_t __ret;
27910  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27911  return __ret;
27912 }
27913 #else
27914 __ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27915  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27916  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27917  int32x2_t __ret;
27918  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
27919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27920  return __ret;
27921 }
27922 __ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27923  int32x2_t __ret;
27924  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27925  return __ret;
27926 }
27927 #endif
27928 
27929 #ifdef __LITTLE_ENDIAN__
27930 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27931  int8x8_t __ret;
27932  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27933  return __ret;
27934 }
27935 #else
27936 __ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27937  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27938  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27939  int8x8_t __ret;
27940  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
27941  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27942  return __ret;
27943 }
27944 __ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27945  int8x8_t __ret;
27946  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27947  return __ret;
27948 }
27949 #endif
27950 
27951 #ifdef __LITTLE_ENDIAN__
27952 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27953  uint16x8_t __ret;
27954  __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
27955  return __ret;
27956 }
27957 #else
27958 __ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27959  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27960  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27961  uint16x8_t __ret;
27962  __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
27963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27964  return __ret;
27965 }
27966 #endif
27967 
27968 #ifdef __LITTLE_ENDIAN__
27969 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27970  uint64x2_t __ret;
27971  __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
27972  return __ret;
27973 }
27974 #else
27975 __ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27976  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27977  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27978  uint64x2_t __ret;
27979  __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
27980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27981  return __ret;
27982 }
27983 #endif
27984 
27985 #ifdef __LITTLE_ENDIAN__
27986 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27987  uint32x4_t __ret;
27988  __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
27989  return __ret;
27990 }
27991 #else
27992 __ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27993  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27994  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27995  uint32x4_t __ret;
27996  __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
27997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27998  return __ret;
27999 }
28000 #endif
28001 
28002 #ifdef __LITTLE_ENDIAN__
28003 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
28004  int16x8_t __ret;
28005  __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
28006  return __ret;
28007 }
28008 #else
28009 __ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
28010  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28011  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28012  int16x8_t __ret;
28013  __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
28014  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28015  return __ret;
28016 }
28017 #endif
28018 
28019 #ifdef __LITTLE_ENDIAN__
28020 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
28021  int64x2_t __ret;
28022  __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
28023  return __ret;
28024 }
28025 #else
28026 __ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
28027  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28028  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28029  int64x2_t __ret;
28030  __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
28031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28032  return __ret;
28033 }
28034 #endif
28035 
28036 #ifdef __LITTLE_ENDIAN__
28037 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
28038  int32x4_t __ret;
28039  __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
28040  return __ret;
28041 }
28042 #else
28043 __ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
28044  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28045  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28046  int32x4_t __ret;
28047  __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
28048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28049  return __ret;
28050 }
28051 #endif
28052 
28053 #ifdef __LITTLE_ENDIAN__
28054 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28055  uint16x8_t __ret;
28056  __ret = __p0 - vmovl_u8(__p1);
28057  return __ret;
28058 }
28059 #else
28060 __ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28061  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28062  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28063  uint16x8_t __ret;
28064  __ret = __rev0 - __noswap_vmovl_u8(__rev1);
28065  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28066  return __ret;
28067 }
28068 #endif
28069 
28070 #ifdef __LITTLE_ENDIAN__
28071 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28072  uint64x2_t __ret;
28073  __ret = __p0 - vmovl_u32(__p1);
28074  return __ret;
28075 }
28076 #else
28077 __ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28078  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28079  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28080  uint64x2_t __ret;
28081  __ret = __rev0 - __noswap_vmovl_u32(__rev1);
28082  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28083  return __ret;
28084 }
28085 #endif
28086 
28087 #ifdef __LITTLE_ENDIAN__
28088 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28089  uint32x4_t __ret;
28090  __ret = __p0 - vmovl_u16(__p1);
28091  return __ret;
28092 }
28093 #else
28094 __ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28095  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28096  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28097  uint32x4_t __ret;
28098  __ret = __rev0 - __noswap_vmovl_u16(__rev1);
28099  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28100  return __ret;
28101 }
28102 #endif
28103 
28104 #ifdef __LITTLE_ENDIAN__
28105 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28106  int16x8_t __ret;
28107  __ret = __p0 - vmovl_s8(__p1);
28108  return __ret;
28109 }
28110 #else
28111 __ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28112  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28113  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28114  int16x8_t __ret;
28115  __ret = __rev0 - __noswap_vmovl_s8(__rev1);
28116  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28117  return __ret;
28118 }
28119 #endif
28120 
28121 #ifdef __LITTLE_ENDIAN__
28122 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28123  int64x2_t __ret;
28124  __ret = __p0 - vmovl_s32(__p1);
28125  return __ret;
28126 }
28127 #else
28128 __ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28129  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28130  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28131  int64x2_t __ret;
28132  __ret = __rev0 - __noswap_vmovl_s32(__rev1);
28133  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28134  return __ret;
28135 }
28136 #endif
28137 
28138 #ifdef __LITTLE_ENDIAN__
28139 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28140  int32x4_t __ret;
28141  __ret = __p0 - vmovl_s16(__p1);
28142  return __ret;
28143 }
28144 #else
28145 __ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28146  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28147  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28148  int32x4_t __ret;
28149  __ret = __rev0 - __noswap_vmovl_s16(__rev1);
28150  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28151  return __ret;
28152 }
28153 #endif
28154 
28155 #ifdef __LITTLE_ENDIAN__
28156 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28157  poly8x8_t __ret;
28158  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
28159  return __ret;
28160 }
28161 #else
28162 __ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28163  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28164  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28165  poly8x8_t __ret;
28166  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28167  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28168  return __ret;
28169 }
28170 #endif
28171 
28172 #ifdef __LITTLE_ENDIAN__
28173 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28174  uint8x8_t __ret;
28175  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28176  return __ret;
28177 }
28178 #else
28179 __ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28180  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28181  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28182  uint8x8_t __ret;
28183  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28185  return __ret;
28186 }
28187 #endif
28188 
28189 #ifdef __LITTLE_ENDIAN__
28190 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28191  int8x8_t __ret;
28192  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
28193  return __ret;
28194 }
28195 #else
28196 __ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28197  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28198  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28199  int8x8_t __ret;
28200  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28201  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28202  return __ret;
28203 }
28204 #endif
28205 
28206 #ifdef __LITTLE_ENDIAN__
28207 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28208  poly8x8_t __ret;
28209  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
28210  return __ret;
28211 }
28212 #else
28213 __ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28214  poly8x8x2_t __rev0;
28215  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28216  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28217  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28218  poly8x8_t __ret;
28219  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
28220  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28221  return __ret;
28222 }
28223 #endif
28224 
28225 #ifdef __LITTLE_ENDIAN__
28226 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28227  uint8x8_t __ret;
28228  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
28229  return __ret;
28230 }
28231 #else
28232 __ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28233  uint8x8x2_t __rev0;
28234  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28235  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28236  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28237  uint8x8_t __ret;
28238  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
28239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28240  return __ret;
28241 }
28242 #endif
28243 
28244 #ifdef __LITTLE_ENDIAN__
28245 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28246  int8x8_t __ret;
28247  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
28248  return __ret;
28249 }
28250 #else
28251 __ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28252  int8x8x2_t __rev0;
28253  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28254  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28255  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28256  int8x8_t __ret;
28257  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
28258  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28259  return __ret;
28260 }
28261 #endif
28262 
28263 #ifdef __LITTLE_ENDIAN__
28264 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28265  poly8x8_t __ret;
28266  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
28267  return __ret;
28268 }
28269 #else
28270 __ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28271  poly8x8x3_t __rev0;
28272  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28273  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28274  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28275  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28276  poly8x8_t __ret;
28277  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
28278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28279  return __ret;
28280 }
28281 #endif
28282 
28283 #ifdef __LITTLE_ENDIAN__
28284 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28285  uint8x8_t __ret;
28286  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
28287  return __ret;
28288 }
28289 #else
28290 __ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28291  uint8x8x3_t __rev0;
28292  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28293  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28294  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28295  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28296  uint8x8_t __ret;
28297  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
28298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28299  return __ret;
28300 }
28301 #endif
28302 
28303 #ifdef __LITTLE_ENDIAN__
28304 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28305  int8x8_t __ret;
28306  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
28307  return __ret;
28308 }
28309 #else
28310 __ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28311  int8x8x3_t __rev0;
28312  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28313  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28314  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28315  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28316  int8x8_t __ret;
28317  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
28318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28319  return __ret;
28320 }
28321 #endif
28322 
28323 #ifdef __LITTLE_ENDIAN__
28324 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28325  poly8x8_t __ret;
28326  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
28327  return __ret;
28328 }
28329 #else
28330 __ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28331  poly8x8x4_t __rev0;
28332  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28333  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28334  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28335  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28336  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28337  poly8x8_t __ret;
28338  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
28339  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28340  return __ret;
28341 }
28342 #endif
28343 
28344 #ifdef __LITTLE_ENDIAN__
28345 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28346  uint8x8_t __ret;
28347  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
28348  return __ret;
28349 }
28350 #else
28351 __ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28352  uint8x8x4_t __rev0;
28353  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28354  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28355  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28356  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28357  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28358  uint8x8_t __ret;
28359  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
28360  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28361  return __ret;
28362 }
28363 #endif
28364 
28365 #ifdef __LITTLE_ENDIAN__
28366 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28367  int8x8_t __ret;
28368  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
28369  return __ret;
28370 }
28371 #else
28372 __ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28373  int8x8x4_t __rev0;
28374  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28375  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28376  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28377  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28378  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28379  int8x8_t __ret;
28380  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
28381  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28382  return __ret;
28383 }
28384 #endif
28385 
28386 #ifdef __LITTLE_ENDIAN__
28387 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28388  poly8x8_t __ret;
28389  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
28390  return __ret;
28391 }
28392 #else
28393 __ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28394  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28395  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28396  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28397  poly8x8_t __ret;
28398  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
28399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28400  return __ret;
28401 }
28402 #endif
28403 
28404 #ifdef __LITTLE_ENDIAN__
28405 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28406  uint8x8_t __ret;
28407  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
28408  return __ret;
28409 }
28410 #else
28411 __ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28412  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28413  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28414  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28415  uint8x8_t __ret;
28416  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
28417  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28418  return __ret;
28419 }
28420 #endif
28421 
28422 #ifdef __LITTLE_ENDIAN__
28423 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28424  int8x8_t __ret;
28425  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
28426  return __ret;
28427 }
28428 #else
28429 __ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28430  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28431  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28432  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28433  int8x8_t __ret;
28434  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
28435  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28436  return __ret;
28437 }
28438 #endif
28439 
28440 #ifdef __LITTLE_ENDIAN__
28441 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28442  poly8x8_t __ret;
28443  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
28444  return __ret;
28445 }
28446 #else
28447 __ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28448  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28449  poly8x8x2_t __rev1;
28450  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28451  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28452  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28453  poly8x8_t __ret;
28454  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
28455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28456  return __ret;
28457 }
28458 #endif
28459 
28460 #ifdef __LITTLE_ENDIAN__
28461 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28462  uint8x8_t __ret;
28463  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
28464  return __ret;
28465 }
28466 #else
28467 __ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28468  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28469  uint8x8x2_t __rev1;
28470  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28471  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28472  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28473  uint8x8_t __ret;
28474  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
28475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28476  return __ret;
28477 }
28478 #endif
28479 
28480 #ifdef __LITTLE_ENDIAN__
28481 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28482  int8x8_t __ret;
28483  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
28484  return __ret;
28485 }
28486 #else
28487 __ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28488  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28489  int8x8x2_t __rev1;
28490  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28491  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28492  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28493  int8x8_t __ret;
28494  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
28495  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28496  return __ret;
28497 }
28498 #endif
28499 
28500 #ifdef __LITTLE_ENDIAN__
28501 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28502  poly8x8_t __ret;
28503  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
28504  return __ret;
28505 }
28506 #else
28507 __ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28508  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28509  poly8x8x3_t __rev1;
28510  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28511  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28512  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28513  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28514  poly8x8_t __ret;
28515  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
28516  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28517  return __ret;
28518 }
28519 #endif
28520 
28521 #ifdef __LITTLE_ENDIAN__
28522 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28523  uint8x8_t __ret;
28524  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
28525  return __ret;
28526 }
28527 #else
28528 __ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28529  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28530  uint8x8x3_t __rev1;
28531  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28532  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28533  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28534  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28535  uint8x8_t __ret;
28536  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
28537  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28538  return __ret;
28539 }
28540 #endif
28541 
28542 #ifdef __LITTLE_ENDIAN__
28543 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28544  int8x8_t __ret;
28545  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
28546  return __ret;
28547 }
28548 #else
28549 __ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28550  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28551  int8x8x3_t __rev1;
28552  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28553  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28554  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28555  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28556  int8x8_t __ret;
28557  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
28558  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28559  return __ret;
28560 }
28561 #endif
28562 
28563 #ifdef __LITTLE_ENDIAN__
28564 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28565  poly8x8_t __ret;
28566  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
28567  return __ret;
28568 }
28569 #else
28570 __ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28571  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28572  poly8x8x4_t __rev1;
28573  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28574  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28575  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28576  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28577  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28578  poly8x8_t __ret;
28579  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
28580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28581  return __ret;
28582 }
28583 #endif
28584 
28585 #ifdef __LITTLE_ENDIAN__
28586 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28587  uint8x8_t __ret;
28588  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
28589  return __ret;
28590 }
28591 #else
28592 __ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28593  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28594  uint8x8x4_t __rev1;
28595  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28596  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28597  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28598  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28599  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28600  uint8x8_t __ret;
28601  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
28602  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28603  return __ret;
28604 }
28605 #endif
28606 
28607 #ifdef __LITTLE_ENDIAN__
28608 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28609  int8x8_t __ret;
28610  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
28611  return __ret;
28612 }
28613 #else
28614 __ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28615  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28616  int8x8x4_t __rev1;
28617  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28618  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28619  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28620  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28621  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28622  int8x8_t __ret;
28623  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
28624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28625  return __ret;
28626 }
28627 #endif
28628 
28629 #ifdef __LITTLE_ENDIAN__
28630 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28631  poly8x8x2_t __ret;
28632  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
28633  return __ret;
28634 }
28635 #else
28636 __ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28637  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28638  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28639  poly8x8x2_t __ret;
28640  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28641 
28642  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28643  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28644  return __ret;
28645 }
28646 #endif
28647 
28648 #ifdef __LITTLE_ENDIAN__
28649 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28650  poly16x4x2_t __ret;
28651  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
28652  return __ret;
28653 }
28654 #else
28655 __ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28656  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28657  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28658  poly16x4x2_t __ret;
28659  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
28660 
28661  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28662  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28663  return __ret;
28664 }
28665 #endif
28666 
28667 #ifdef __LITTLE_ENDIAN__
28668 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28669  poly8x16x2_t __ret;
28670  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
28671  return __ret;
28672 }
28673 #else
28674 __ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28675  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28676  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28677  poly8x16x2_t __ret;
28678  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
28679 
28680  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28681  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28682  return __ret;
28683 }
28684 #endif
28685 
28686 #ifdef __LITTLE_ENDIAN__
28687 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28688  poly16x8x2_t __ret;
28689  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
28690  return __ret;
28691 }
28692 #else
28693 __ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28694  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28695  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28696  poly16x8x2_t __ret;
28697  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
28698 
28699  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28700  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28701  return __ret;
28702 }
28703 #endif
28704 
28705 #ifdef __LITTLE_ENDIAN__
28706 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28707  uint8x16x2_t __ret;
28708  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
28709  return __ret;
28710 }
28711 #else
28712 __ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28713  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28714  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28715  uint8x16x2_t __ret;
28716  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28717 
28718  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28719  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28720  return __ret;
28721 }
28722 #endif
28723 
28724 #ifdef __LITTLE_ENDIAN__
28725 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28726  uint32x4x2_t __ret;
28727  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
28728  return __ret;
28729 }
28730 #else
28731 __ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28732  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28733  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28734  uint32x4x2_t __ret;
28735  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
28736 
28737  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28738  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28739  return __ret;
28740 }
28741 #endif
28742 
28743 #ifdef __LITTLE_ENDIAN__
28744 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28745  uint16x8x2_t __ret;
28746  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
28747  return __ret;
28748 }
28749 #else
28750 __ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28751  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28752  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28753  uint16x8x2_t __ret;
28754  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28755 
28756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28758  return __ret;
28759 }
28760 #endif
28761 
28762 #ifdef __LITTLE_ENDIAN__
28763 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28764  int8x16x2_t __ret;
28765  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
28766  return __ret;
28767 }
28768 #else
28769 __ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28770  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28771  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28772  int8x16x2_t __ret;
28773  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
28774 
28775  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28776  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28777  return __ret;
28778 }
28779 #endif
28780 
28781 #ifdef __LITTLE_ENDIAN__
28782 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28783  float32x4x2_t __ret;
28784  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
28785  return __ret;
28786 }
28787 #else
28788 __ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28789  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28790  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28791  float32x4x2_t __ret;
28792  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
28793 
28794  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28795  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28796  return __ret;
28797 }
28798 #endif
28799 
28800 #ifdef __LITTLE_ENDIAN__
28801 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28802  int32x4x2_t __ret;
28803  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
28804  return __ret;
28805 }
28806 #else
28807 __ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28808  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28809  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28810  int32x4x2_t __ret;
28811  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
28812 
28813  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28814  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28815  return __ret;
28816 }
28817 #endif
28818 
28819 #ifdef __LITTLE_ENDIAN__
28820 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28821  int16x8x2_t __ret;
28822  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
28823  return __ret;
28824 }
28825 #else
28826 __ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28827  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28828  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28829  int16x8x2_t __ret;
28830  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
28831 
28832  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28833  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28834  return __ret;
28835 }
28836 #endif
28837 
28838 #ifdef __LITTLE_ENDIAN__
28839 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28840  uint8x8x2_t __ret;
28841  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
28842  return __ret;
28843 }
28844 #else
28845 __ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28846  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28847  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28848  uint8x8x2_t __ret;
28849  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28850 
28851  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28852  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28853  return __ret;
28854 }
28855 #endif
28856 
28857 #ifdef __LITTLE_ENDIAN__
28858 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28859  uint32x2x2_t __ret;
28860  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
28861  return __ret;
28862 }
28863 #else
28864 __ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28865  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28866  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28867  uint32x2x2_t __ret;
28868  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
28869 
28870  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28871  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28872  return __ret;
28873 }
28874 #endif
28875 
28876 #ifdef __LITTLE_ENDIAN__
28877 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28878  uint16x4x2_t __ret;
28879  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
28880  return __ret;
28881 }
28882 #else
28883 __ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28884  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28885  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28886  uint16x4x2_t __ret;
28887  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28888 
28889  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28890  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28891  return __ret;
28892 }
28893 #endif
28894 
28895 #ifdef __LITTLE_ENDIAN__
28896 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28897  int8x8x2_t __ret;
28898  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
28899  return __ret;
28900 }
28901 #else
28902 __ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28903  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28904  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28905  int8x8x2_t __ret;
28906  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28907 
28908  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28909  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28910  return __ret;
28911 }
28912 #endif
28913 
28914 #ifdef __LITTLE_ENDIAN__
28915 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28916  float32x2x2_t __ret;
28917  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
28918  return __ret;
28919 }
28920 #else
28921 __ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28922  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28923  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28924  float32x2x2_t __ret;
28925  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
28926 
28927  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28928  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28929  return __ret;
28930 }
28931 #endif
28932 
28933 #ifdef __LITTLE_ENDIAN__
28934 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28935  int32x2x2_t __ret;
28936  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
28937  return __ret;
28938 }
28939 #else
28940 __ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28941  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28942  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28943  int32x2x2_t __ret;
28944  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
28945 
28946  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28947  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28948  return __ret;
28949 }
28950 #endif
28951 
28952 #ifdef __LITTLE_ENDIAN__
28953 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28954  int16x4x2_t __ret;
28955  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
28956  return __ret;
28957 }
28958 #else
28959 __ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28960  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28961  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28962  int16x4x2_t __ret;
28963  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
28964 
28965  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28966  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28967  return __ret;
28968 }
28969 #endif
28970 
28971 #ifdef __LITTLE_ENDIAN__
28972 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28973  uint8x8_t __ret;
28974  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28975  return __ret;
28976 }
28977 #else
28978 __ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28979  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28980  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28981  uint8x8_t __ret;
28982  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28984  return __ret;
28985 }
28986 #endif
28987 
28988 #ifdef __LITTLE_ENDIAN__
28989 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28990  uint16x4_t __ret;
28991  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
28992  return __ret;
28993 }
28994 #else
28995 __ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28996  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28997  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28998  uint16x4_t __ret;
28999  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29001  return __ret;
29002 }
29003 #endif
29004 
29005 #ifdef __LITTLE_ENDIAN__
29006 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29007  uint8x16_t __ret;
29008  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29009  return __ret;
29010 }
29011 #else
29012 __ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29013  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29014  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29015  uint8x16_t __ret;
29016  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29017  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29018  return __ret;
29019 }
29020 #endif
29021 
29022 #ifdef __LITTLE_ENDIAN__
29023 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29024  uint16x8_t __ret;
29025  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29026  return __ret;
29027 }
29028 #else
29029 __ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29030  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29031  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29032  uint16x8_t __ret;
29033  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29034  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29035  return __ret;
29036 }
29037 #endif
29038 
29039 #ifdef __LITTLE_ENDIAN__
29040 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29041  uint8x16_t __ret;
29042  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29043  return __ret;
29044 }
29045 #else
29046 __ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29047  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29048  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29049  uint8x16_t __ret;
29050  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29051  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29052  return __ret;
29053 }
29054 #endif
29055 
29056 #ifdef __LITTLE_ENDIAN__
29057 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29058  uint32x4_t __ret;
29059  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29060  return __ret;
29061 }
29062 #else
29063 __ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29064  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29065  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29066  uint32x4_t __ret;
29067  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29068  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29069  return __ret;
29070 }
29071 #endif
29072 
29073 #ifdef __LITTLE_ENDIAN__
29074 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29075  uint16x8_t __ret;
29076  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29077  return __ret;
29078 }
29079 #else
29080 __ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29081  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29082  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29083  uint16x8_t __ret;
29084  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29085  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29086  return __ret;
29087 }
29088 #endif
29089 
29090 #ifdef __LITTLE_ENDIAN__
29091 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29092  uint8x16_t __ret;
29093  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29094  return __ret;
29095 }
29096 #else
29097 __ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29098  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29099  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29100  uint8x16_t __ret;
29101  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29102  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29103  return __ret;
29104 }
29105 #endif
29106 
29107 #ifdef __LITTLE_ENDIAN__
29108 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29109  uint32x4_t __ret;
29110  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29111  return __ret;
29112 }
29113 #else
29114 __ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29115  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29116  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29117  uint32x4_t __ret;
29118  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29120  return __ret;
29121 }
29122 #endif
29123 
29124 #ifdef __LITTLE_ENDIAN__
29125 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29126  uint16x8_t __ret;
29127  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29128  return __ret;
29129 }
29130 #else
29131 __ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29132  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29133  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29134  uint16x8_t __ret;
29135  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29136  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29137  return __ret;
29138 }
29139 #endif
29140 
29141 #ifdef __LITTLE_ENDIAN__
29142 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29143  uint8x8_t __ret;
29144  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29145  return __ret;
29146 }
29147 #else
29148 __ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29149  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29150  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29151  uint8x8_t __ret;
29152  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29153  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29154  return __ret;
29155 }
29156 #endif
29157 
29158 #ifdef __LITTLE_ENDIAN__
29159 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29160  uint32x2_t __ret;
29161  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29162  return __ret;
29163 }
29164 #else
29165 __ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29166  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29167  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29168  uint32x2_t __ret;
29169  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29170  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29171  return __ret;
29172 }
29173 #endif
29174 
29175 #ifdef __LITTLE_ENDIAN__
29176 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29177  uint16x4_t __ret;
29178  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29179  return __ret;
29180 }
29181 #else
29182 __ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29183  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29184  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29185  uint16x4_t __ret;
29186  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29187  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29188  return __ret;
29189 }
29190 #endif
29191 
29192 #ifdef __LITTLE_ENDIAN__
29193 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29194  uint8x8_t __ret;
29195  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29196  return __ret;
29197 }
29198 #else
29199 __ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29200  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29201  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29202  uint8x8_t __ret;
29203  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29204  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29205  return __ret;
29206 }
29207 #endif
29208 
29209 #ifdef __LITTLE_ENDIAN__
29210 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29211  uint32x2_t __ret;
29212  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29213  return __ret;
29214 }
29215 #else
29216 __ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29217  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29218  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29219  uint32x2_t __ret;
29220  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29221  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29222  return __ret;
29223 }
29224 #endif
29225 
29226 #ifdef __LITTLE_ENDIAN__
29227 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29228  uint16x4_t __ret;
29229  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29230  return __ret;
29231 }
29232 #else
29233 __ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29234  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29235  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29236  uint16x4_t __ret;
29237  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29239  return __ret;
29240 }
29241 #endif
29242 
29243 #ifdef __LITTLE_ENDIAN__
29244 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29245  poly8x8x2_t __ret;
29246  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29247  return __ret;
29248 }
29249 #else
29250 __ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29251  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29252  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29253  poly8x8x2_t __ret;
29254  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29255 
29256  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29257  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29258  return __ret;
29259 }
29260 #endif
29261 
29262 #ifdef __LITTLE_ENDIAN__
29263 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29264  poly16x4x2_t __ret;
29265  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29266  return __ret;
29267 }
29268 #else
29269 __ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29270  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29271  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29272  poly16x4x2_t __ret;
29273  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29274 
29275  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29276  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29277  return __ret;
29278 }
29279 #endif
29280 
29281 #ifdef __LITTLE_ENDIAN__
29282 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29283  poly8x16x2_t __ret;
29284  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29285  return __ret;
29286 }
29287 #else
29288 __ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29289  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29290  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29291  poly8x16x2_t __ret;
29292  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29293 
29294  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29295  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29296  return __ret;
29297 }
29298 #endif
29299 
29300 #ifdef __LITTLE_ENDIAN__
29301 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29302  poly16x8x2_t __ret;
29303  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29304  return __ret;
29305 }
29306 #else
29307 __ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29308  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29309  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29310  poly16x8x2_t __ret;
29311  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29312 
29313  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29314  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29315  return __ret;
29316 }
29317 #endif
29318 
29319 #ifdef __LITTLE_ENDIAN__
29320 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29321  uint8x16x2_t __ret;
29322  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29323  return __ret;
29324 }
29325 #else
29326 __ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29327  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29328  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29329  uint8x16x2_t __ret;
29330  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29331 
29332  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29333  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29334  return __ret;
29335 }
29336 #endif
29337 
29338 #ifdef __LITTLE_ENDIAN__
29339 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29340  uint32x4x2_t __ret;
29341  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29342  return __ret;
29343 }
29344 #else
29345 __ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29346  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29347  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29348  uint32x4x2_t __ret;
29349  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29350 
29351  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29352  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29353  return __ret;
29354 }
29355 #endif
29356 
29357 #ifdef __LITTLE_ENDIAN__
29358 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29359  uint16x8x2_t __ret;
29360  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29361  return __ret;
29362 }
29363 #else
29364 __ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29365  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29366  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29367  uint16x8x2_t __ret;
29368  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29369 
29370  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29371  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29372  return __ret;
29373 }
29374 #endif
29375 
29376 #ifdef __LITTLE_ENDIAN__
29377 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29378  int8x16x2_t __ret;
29379  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29380  return __ret;
29381 }
29382 #else
29383 __ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29384  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29385  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29386  int8x16x2_t __ret;
29387  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29388 
29389  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29390  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29391  return __ret;
29392 }
29393 #endif
29394 
29395 #ifdef __LITTLE_ENDIAN__
29396 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29397  float32x4x2_t __ret;
29398  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29399  return __ret;
29400 }
29401 #else
29402 __ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29403  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29404  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29405  float32x4x2_t __ret;
29406  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29407 
29408  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29409  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29410  return __ret;
29411 }
29412 #endif
29413 
29414 #ifdef __LITTLE_ENDIAN__
29415 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29416  int32x4x2_t __ret;
29417  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29418  return __ret;
29419 }
29420 #else
29421 __ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29422  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29423  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29424  int32x4x2_t __ret;
29425  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29426 
29427  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29428  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29429  return __ret;
29430 }
29431 #endif
29432 
29433 #ifdef __LITTLE_ENDIAN__
29434 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29435  int16x8x2_t __ret;
29436  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29437  return __ret;
29438 }
29439 #else
29440 __ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29441  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29442  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29443  int16x8x2_t __ret;
29444  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29445 
29446  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29447  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29448  return __ret;
29449 }
29450 #endif
29451 
29452 #ifdef __LITTLE_ENDIAN__
29453 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29454  uint8x8x2_t __ret;
29455  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29456  return __ret;
29457 }
29458 #else
29459 __ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29460  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29461  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29462  uint8x8x2_t __ret;
29463  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29464 
29465  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29466  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29467  return __ret;
29468 }
29469 #endif
29470 
29471 #ifdef __LITTLE_ENDIAN__
29472 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29473  uint32x2x2_t __ret;
29474  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29475  return __ret;
29476 }
29477 #else
29478 __ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29479  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29480  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29481  uint32x2x2_t __ret;
29482  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29483 
29484  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29485  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29486  return __ret;
29487 }
29488 #endif
29489 
29490 #ifdef __LITTLE_ENDIAN__
29491 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29492  uint16x4x2_t __ret;
29493  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29494  return __ret;
29495 }
29496 #else
29497 __ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29498  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29499  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29500  uint16x4x2_t __ret;
29501  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29502 
29503  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29504  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29505  return __ret;
29506 }
29507 #endif
29508 
29509 #ifdef __LITTLE_ENDIAN__
29510 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29511  int8x8x2_t __ret;
29512  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29513  return __ret;
29514 }
29515 #else
29516 __ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29517  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29518  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29519  int8x8x2_t __ret;
29520  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29521 
29522  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29523  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29524  return __ret;
29525 }
29526 #endif
29527 
29528 #ifdef __LITTLE_ENDIAN__
29529 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29530  float32x2x2_t __ret;
29531  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29532  return __ret;
29533 }
29534 #else
29535 __ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29536  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29537  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29538  float32x2x2_t __ret;
29539  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29540 
29541  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29542  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29543  return __ret;
29544 }
29545 #endif
29546 
29547 #ifdef __LITTLE_ENDIAN__
29548 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29549  int32x2x2_t __ret;
29550  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29551  return __ret;
29552 }
29553 #else
29554 __ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29555  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29556  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29557  int32x2x2_t __ret;
29558  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29559 
29560  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29561  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29562  return __ret;
29563 }
29564 #endif
29565 
29566 #ifdef __LITTLE_ENDIAN__
29567 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29568  int16x4x2_t __ret;
29569  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29570  return __ret;
29571 }
29572 #else
29573 __ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29574  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29575  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29576  int16x4x2_t __ret;
29577  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29578 
29579  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29580  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29581  return __ret;
29582 }
29583 #endif
29584 
29585 #ifdef __LITTLE_ENDIAN__
29586 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29587  poly8x8x2_t __ret;
29588  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29589  return __ret;
29590 }
29591 #else
29592 __ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29593  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29594  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29595  poly8x8x2_t __ret;
29596  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29597 
29598  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29599  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29600  return __ret;
29601 }
29602 #endif
29603 
29604 #ifdef __LITTLE_ENDIAN__
29605 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29606  poly16x4x2_t __ret;
29607  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29608  return __ret;
29609 }
29610 #else
29611 __ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29612  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29613  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29614  poly16x4x2_t __ret;
29615  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29616 
29617  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29618  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29619  return __ret;
29620 }
29621 #endif
29622 
29623 #ifdef __LITTLE_ENDIAN__
29624 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29625  poly8x16x2_t __ret;
29626  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29627  return __ret;
29628 }
29629 #else
29630 __ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29631  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29632  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29633  poly8x16x2_t __ret;
29634  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29635 
29636  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29637  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29638  return __ret;
29639 }
29640 #endif
29641 
29642 #ifdef __LITTLE_ENDIAN__
29643 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29644  poly16x8x2_t __ret;
29645  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29646  return __ret;
29647 }
29648 #else
29649 __ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29650  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29651  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29652  poly16x8x2_t __ret;
29653  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29654 
29655  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29656  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29657  return __ret;
29658 }
29659 #endif
29660 
29661 #ifdef __LITTLE_ENDIAN__
29662 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29663  uint8x16x2_t __ret;
29664  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29665  return __ret;
29666 }
29667 #else
29668 __ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29669  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29670  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29671  uint8x16x2_t __ret;
29672  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29673 
29674  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29675  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29676  return __ret;
29677 }
29678 #endif
29679 
29680 #ifdef __LITTLE_ENDIAN__
29681 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29682  uint32x4x2_t __ret;
29683  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29684  return __ret;
29685 }
29686 #else
29687 __ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29688  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29689  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29690  uint32x4x2_t __ret;
29691  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29692 
29693  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29694  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29695  return __ret;
29696 }
29697 #endif
29698 
29699 #ifdef __LITTLE_ENDIAN__
29700 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29701  uint16x8x2_t __ret;
29702  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29703  return __ret;
29704 }
29705 #else
29706 __ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29707  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29708  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29709  uint16x8x2_t __ret;
29710  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29711 
29712  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29713  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29714  return __ret;
29715 }
29716 #endif
29717 
29718 #ifdef __LITTLE_ENDIAN__
29719 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29720  int8x16x2_t __ret;
29721  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29722  return __ret;
29723 }
29724 #else
29725 __ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29726  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29727  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29728  int8x16x2_t __ret;
29729  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29730 
29731  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29732  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29733  return __ret;
29734 }
29735 #endif
29736 
29737 #ifdef __LITTLE_ENDIAN__
29738 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29739  float32x4x2_t __ret;
29740  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29741  return __ret;
29742 }
29743 #else
29744 __ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29745  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29746  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29747  float32x4x2_t __ret;
29748  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29749 
29750  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29751  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29752  return __ret;
29753 }
29754 #endif
29755 
29756 #ifdef __LITTLE_ENDIAN__
29757 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29758  int32x4x2_t __ret;
29759  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29760  return __ret;
29761 }
29762 #else
29763 __ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29764  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29765  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29766  int32x4x2_t __ret;
29767  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29768 
29769  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29770  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29771  return __ret;
29772 }
29773 #endif
29774 
29775 #ifdef __LITTLE_ENDIAN__
29776 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29777  int16x8x2_t __ret;
29778  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29779  return __ret;
29780 }
29781 #else
29782 __ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29783  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29784  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29785  int16x8x2_t __ret;
29786  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29787 
29788  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29789  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29790  return __ret;
29791 }
29792 #endif
29793 
29794 #ifdef __LITTLE_ENDIAN__
29795 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29796  uint8x8x2_t __ret;
29797  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29798  return __ret;
29799 }
29800 #else
29801 __ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29802  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29803  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29804  uint8x8x2_t __ret;
29805  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29806 
29807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29809  return __ret;
29810 }
29811 #endif
29812 
29813 #ifdef __LITTLE_ENDIAN__
29814 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29815  uint32x2x2_t __ret;
29816  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29817  return __ret;
29818 }
29819 #else
29820 __ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29821  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29822  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29823  uint32x2x2_t __ret;
29824  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29825 
29826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29828  return __ret;
29829 }
29830 #endif
29831 
29832 #ifdef __LITTLE_ENDIAN__
29833 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29834  uint16x4x2_t __ret;
29835  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29836  return __ret;
29837 }
29838 #else
29839 __ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29840  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29841  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29842  uint16x4x2_t __ret;
29843  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29844 
29845  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29846  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29847  return __ret;
29848 }
29849 #endif
29850 
29851 #ifdef __LITTLE_ENDIAN__
29852 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29853  int8x8x2_t __ret;
29854  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29855  return __ret;
29856 }
29857 #else
29858 __ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29859  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29860  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29861  int8x8x2_t __ret;
29862  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29863 
29864  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29865  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29866  return __ret;
29867 }
29868 #endif
29869 
29870 #ifdef __LITTLE_ENDIAN__
29871 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29872  float32x2x2_t __ret;
29873  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29874  return __ret;
29875 }
29876 #else
29877 __ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29878  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29879  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29880  float32x2x2_t __ret;
29881  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29882 
29883  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29884  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29885  return __ret;
29886 }
29887 #endif
29888 
29889 #ifdef __LITTLE_ENDIAN__
29890 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29891  int32x2x2_t __ret;
29892  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29893  return __ret;
29894 }
29895 #else
29896 __ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29897  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29898  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29899  int32x2x2_t __ret;
29900  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29901 
29902  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29903  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29904  return __ret;
29905 }
29906 #endif
29907 
29908 #ifdef __LITTLE_ENDIAN__
29909 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29910  int16x4x2_t __ret;
29911  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29912  return __ret;
29913 }
29914 #else
29915 __ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29916  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29917  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29918  int16x4x2_t __ret;
29919  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29920 
29921  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29922  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29923  return __ret;
29924 }
29925 #endif
29926 
29927 #if !defined(__aarch64__)
29928 #ifdef __LITTLE_ENDIAN__
29929 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29930  poly8x8_t __ret;
29931  __ret = (poly8x8_t)(__p0);
29932  return __ret;
29933 }
29934 #else
29935 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29936  poly8x8_t __ret;
29937  __ret = (poly8x8_t)(__p0);
29938  return __ret;
29939 }
29940 #endif
29941 
29942 #ifdef __LITTLE_ENDIAN__
29943 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29944  poly8x8_t __ret;
29945  __ret = (poly8x8_t)(__p0);
29946  return __ret;
29947 }
29948 #else
29949 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29950  poly8x8_t __ret;
29951  __ret = (poly8x8_t)(__p0);
29952  return __ret;
29953 }
29954 #endif
29955 
29956 #ifdef __LITTLE_ENDIAN__
29957 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29958  poly8x8_t __ret;
29959  __ret = (poly8x8_t)(__p0);
29960  return __ret;
29961 }
29962 #else
29963 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29964  poly8x8_t __ret;
29965  __ret = (poly8x8_t)(__p0);
29966  return __ret;
29967 }
29968 #endif
29969 
29970 #ifdef __LITTLE_ENDIAN__
29971 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29972  poly8x8_t __ret;
29973  __ret = (poly8x8_t)(__p0);
29974  return __ret;
29975 }
29976 #else
29977 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29978  poly8x8_t __ret;
29979  __ret = (poly8x8_t)(__p0);
29980  return __ret;
29981 }
29982 #endif
29983 
29984 #ifdef __LITTLE_ENDIAN__
29985 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29986  poly8x8_t __ret;
29987  __ret = (poly8x8_t)(__p0);
29988  return __ret;
29989 }
29990 #else
29991 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29992  poly8x8_t __ret;
29993  __ret = (poly8x8_t)(__p0);
29994  return __ret;
29995 }
29996 #endif
29997 
29998 #ifdef __LITTLE_ENDIAN__
29999 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
30000  poly8x8_t __ret;
30001  __ret = (poly8x8_t)(__p0);
30002  return __ret;
30003 }
30004 #else
30005 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
30006  poly8x8_t __ret;
30007  __ret = (poly8x8_t)(__p0);
30008  return __ret;
30009 }
30010 #endif
30011 
30012 #ifdef __LITTLE_ENDIAN__
30013 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
30014  poly8x8_t __ret;
30015  __ret = (poly8x8_t)(__p0);
30016  return __ret;
30017 }
30018 #else
30019 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
30020  poly8x8_t __ret;
30021  __ret = (poly8x8_t)(__p0);
30022  return __ret;
30023 }
30024 #endif
30025 
30026 #ifdef __LITTLE_ENDIAN__
30027 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
30028  poly8x8_t __ret;
30029  __ret = (poly8x8_t)(__p0);
30030  return __ret;
30031 }
30032 #else
30033 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
30034  poly8x8_t __ret;
30035  __ret = (poly8x8_t)(__p0);
30036  return __ret;
30037 }
30038 #endif
30039 
30040 #ifdef __LITTLE_ENDIAN__
30041 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30042  poly8x8_t __ret;
30043  __ret = (poly8x8_t)(__p0);
30044  return __ret;
30045 }
30046 #else
30047 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30048  poly8x8_t __ret;
30049  __ret = (poly8x8_t)(__p0);
30050  return __ret;
30051 }
30052 #endif
30053 
30054 #ifdef __LITTLE_ENDIAN__
30055 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30056  poly8x8_t __ret;
30057  __ret = (poly8x8_t)(__p0);
30058  return __ret;
30059 }
30060 #else
30061 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30062  poly8x8_t __ret;
30063  __ret = (poly8x8_t)(__p0);
30064  return __ret;
30065 }
30066 #endif
30067 
30068 #ifdef __LITTLE_ENDIAN__
30069 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30070  poly8x8_t __ret;
30071  __ret = (poly8x8_t)(__p0);
30072  return __ret;
30073 }
30074 #else
30075 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30076  poly8x8_t __ret;
30077  __ret = (poly8x8_t)(__p0);
30078  return __ret;
30079 }
30080 #endif
30081 
30082 #ifdef __LITTLE_ENDIAN__
30083 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30084  poly16x4_t __ret;
30085  __ret = (poly16x4_t)(__p0);
30086  return __ret;
30087 }
30088 #else
30089 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30090  poly16x4_t __ret;
30091  __ret = (poly16x4_t)(__p0);
30092  return __ret;
30093 }
30094 #endif
30095 
30096 #ifdef __LITTLE_ENDIAN__
30097 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30098  poly16x4_t __ret;
30099  __ret = (poly16x4_t)(__p0);
30100  return __ret;
30101 }
30102 #else
30103 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30104  poly16x4_t __ret;
30105  __ret = (poly16x4_t)(__p0);
30106  return __ret;
30107 }
30108 #endif
30109 
30110 #ifdef __LITTLE_ENDIAN__
30111 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30112  poly16x4_t __ret;
30113  __ret = (poly16x4_t)(__p0);
30114  return __ret;
30115 }
30116 #else
30117 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30118  poly16x4_t __ret;
30119  __ret = (poly16x4_t)(__p0);
30120  return __ret;
30121 }
30122 #endif
30123 
30124 #ifdef __LITTLE_ENDIAN__
30125 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30126  poly16x4_t __ret;
30127  __ret = (poly16x4_t)(__p0);
30128  return __ret;
30129 }
30130 #else
30131 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30132  poly16x4_t __ret;
30133  __ret = (poly16x4_t)(__p0);
30134  return __ret;
30135 }
30136 #endif
30137 
30138 #ifdef __LITTLE_ENDIAN__
30139 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30140  poly16x4_t __ret;
30141  __ret = (poly16x4_t)(__p0);
30142  return __ret;
30143 }
30144 #else
30145 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30146  poly16x4_t __ret;
30147  __ret = (poly16x4_t)(__p0);
30148  return __ret;
30149 }
30150 #endif
30151 
30152 #ifdef __LITTLE_ENDIAN__
30153 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30154  poly16x4_t __ret;
30155  __ret = (poly16x4_t)(__p0);
30156  return __ret;
30157 }
30158 #else
30159 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30160  poly16x4_t __ret;
30161  __ret = (poly16x4_t)(__p0);
30162  return __ret;
30163 }
30164 #endif
30165 
30166 #ifdef __LITTLE_ENDIAN__
30167 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30168  poly16x4_t __ret;
30169  __ret = (poly16x4_t)(__p0);
30170  return __ret;
30171 }
30172 #else
30173 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30174  poly16x4_t __ret;
30175  __ret = (poly16x4_t)(__p0);
30176  return __ret;
30177 }
30178 #endif
30179 
30180 #ifdef __LITTLE_ENDIAN__
30181 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30182  poly16x4_t __ret;
30183  __ret = (poly16x4_t)(__p0);
30184  return __ret;
30185 }
30186 #else
30187 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30188  poly16x4_t __ret;
30189  __ret = (poly16x4_t)(__p0);
30190  return __ret;
30191 }
30192 #endif
30193 
30194 #ifdef __LITTLE_ENDIAN__
30195 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30196  poly16x4_t __ret;
30197  __ret = (poly16x4_t)(__p0);
30198  return __ret;
30199 }
30200 #else
30201 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30202  poly16x4_t __ret;
30203  __ret = (poly16x4_t)(__p0);
30204  return __ret;
30205 }
30206 #endif
30207 
30208 #ifdef __LITTLE_ENDIAN__
30209 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30210  poly16x4_t __ret;
30211  __ret = (poly16x4_t)(__p0);
30212  return __ret;
30213 }
30214 #else
30215 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30216  poly16x4_t __ret;
30217  __ret = (poly16x4_t)(__p0);
30218  return __ret;
30219 }
30220 #endif
30221 
30222 #ifdef __LITTLE_ENDIAN__
30223 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30224  poly16x4_t __ret;
30225  __ret = (poly16x4_t)(__p0);
30226  return __ret;
30227 }
30228 #else
30229 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30230  poly16x4_t __ret;
30231  __ret = (poly16x4_t)(__p0);
30232  return __ret;
30233 }
30234 #endif
30235 
30236 #ifdef __LITTLE_ENDIAN__
30237 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30238  poly8x16_t __ret;
30239  __ret = (poly8x16_t)(__p0);
30240  return __ret;
30241 }
30242 #else
30243 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30244  poly8x16_t __ret;
30245  __ret = (poly8x16_t)(__p0);
30246  return __ret;
30247 }
30248 #endif
30249 
30250 #ifdef __LITTLE_ENDIAN__
30251 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30252  poly8x16_t __ret;
30253  __ret = (poly8x16_t)(__p0);
30254  return __ret;
30255 }
30256 #else
30257 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30258  poly8x16_t __ret;
30259  __ret = (poly8x16_t)(__p0);
30260  return __ret;
30261 }
30262 #endif
30263 
30264 #ifdef __LITTLE_ENDIAN__
30265 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30266  poly8x16_t __ret;
30267  __ret = (poly8x16_t)(__p0);
30268  return __ret;
30269 }
30270 #else
30271 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30272  poly8x16_t __ret;
30273  __ret = (poly8x16_t)(__p0);
30274  return __ret;
30275 }
30276 #endif
30277 
30278 #ifdef __LITTLE_ENDIAN__
30279 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30280  poly8x16_t __ret;
30281  __ret = (poly8x16_t)(__p0);
30282  return __ret;
30283 }
30284 #else
30285 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30286  poly8x16_t __ret;
30287  __ret = (poly8x16_t)(__p0);
30288  return __ret;
30289 }
30290 #endif
30291 
30292 #ifdef __LITTLE_ENDIAN__
30293 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30294  poly8x16_t __ret;
30295  __ret = (poly8x16_t)(__p0);
30296  return __ret;
30297 }
30298 #else
30299 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30300  poly8x16_t __ret;
30301  __ret = (poly8x16_t)(__p0);
30302  return __ret;
30303 }
30304 #endif
30305 
30306 #ifdef __LITTLE_ENDIAN__
30307 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30308  poly8x16_t __ret;
30309  __ret = (poly8x16_t)(__p0);
30310  return __ret;
30311 }
30312 #else
30313 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30314  poly8x16_t __ret;
30315  __ret = (poly8x16_t)(__p0);
30316  return __ret;
30317 }
30318 #endif
30319 
30320 #ifdef __LITTLE_ENDIAN__
30321 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30322  poly8x16_t __ret;
30323  __ret = (poly8x16_t)(__p0);
30324  return __ret;
30325 }
30326 #else
30327 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30328  poly8x16_t __ret;
30329  __ret = (poly8x16_t)(__p0);
30330  return __ret;
30331 }
30332 #endif
30333 
30334 #ifdef __LITTLE_ENDIAN__
30335 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30336  poly8x16_t __ret;
30337  __ret = (poly8x16_t)(__p0);
30338  return __ret;
30339 }
30340 #else
30341 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30342  poly8x16_t __ret;
30343  __ret = (poly8x16_t)(__p0);
30344  return __ret;
30345 }
30346 #endif
30347 
30348 #ifdef __LITTLE_ENDIAN__
30349 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30350  poly8x16_t __ret;
30351  __ret = (poly8x16_t)(__p0);
30352  return __ret;
30353 }
30354 #else
30355 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30356  poly8x16_t __ret;
30357  __ret = (poly8x16_t)(__p0);
30358  return __ret;
30359 }
30360 #endif
30361 
30362 #ifdef __LITTLE_ENDIAN__
30363 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30364  poly8x16_t __ret;
30365  __ret = (poly8x16_t)(__p0);
30366  return __ret;
30367 }
30368 #else
30369 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30370  poly8x16_t __ret;
30371  __ret = (poly8x16_t)(__p0);
30372  return __ret;
30373 }
30374 #endif
30375 
30376 #ifdef __LITTLE_ENDIAN__
30377 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30378  poly8x16_t __ret;
30379  __ret = (poly8x16_t)(__p0);
30380  return __ret;
30381 }
30382 #else
30383 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30384  poly8x16_t __ret;
30385  __ret = (poly8x16_t)(__p0);
30386  return __ret;
30387 }
30388 #endif
30389 
30390 #ifdef __LITTLE_ENDIAN__
30391 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30392  poly16x8_t __ret;
30393  __ret = (poly16x8_t)(__p0);
30394  return __ret;
30395 }
30396 #else
30397 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30398  poly16x8_t __ret;
30399  __ret = (poly16x8_t)(__p0);
30400  return __ret;
30401 }
30402 #endif
30403 
30404 #ifdef __LITTLE_ENDIAN__
30405 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30406  poly16x8_t __ret;
30407  __ret = (poly16x8_t)(__p0);
30408  return __ret;
30409 }
30410 #else
30411 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30412  poly16x8_t __ret;
30413  __ret = (poly16x8_t)(__p0);
30414  return __ret;
30415 }
30416 #endif
30417 
30418 #ifdef __LITTLE_ENDIAN__
30419 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30420  poly16x8_t __ret;
30421  __ret = (poly16x8_t)(__p0);
30422  return __ret;
30423 }
30424 #else
30425 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30426  poly16x8_t __ret;
30427  __ret = (poly16x8_t)(__p0);
30428  return __ret;
30429 }
30430 #endif
30431 
30432 #ifdef __LITTLE_ENDIAN__
30433 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30434  poly16x8_t __ret;
30435  __ret = (poly16x8_t)(__p0);
30436  return __ret;
30437 }
30438 #else
30439 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30440  poly16x8_t __ret;
30441  __ret = (poly16x8_t)(__p0);
30442  return __ret;
30443 }
30444 #endif
30445 
30446 #ifdef __LITTLE_ENDIAN__
30447 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30448  poly16x8_t __ret;
30449  __ret = (poly16x8_t)(__p0);
30450  return __ret;
30451 }
30452 #else
30453 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30454  poly16x8_t __ret;
30455  __ret = (poly16x8_t)(__p0);
30456  return __ret;
30457 }
30458 #endif
30459 
30460 #ifdef __LITTLE_ENDIAN__
30461 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30462  poly16x8_t __ret;
30463  __ret = (poly16x8_t)(__p0);
30464  return __ret;
30465 }
30466 #else
30467 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30468  poly16x8_t __ret;
30469  __ret = (poly16x8_t)(__p0);
30470  return __ret;
30471 }
30472 #endif
30473 
30474 #ifdef __LITTLE_ENDIAN__
30475 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30476  poly16x8_t __ret;
30477  __ret = (poly16x8_t)(__p0);
30478  return __ret;
30479 }
30480 #else
30481 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30482  poly16x8_t __ret;
30483  __ret = (poly16x8_t)(__p0);
30484  return __ret;
30485 }
30486 #endif
30487 
30488 #ifdef __LITTLE_ENDIAN__
30489 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30490  poly16x8_t __ret;
30491  __ret = (poly16x8_t)(__p0);
30492  return __ret;
30493 }
30494 #else
30495 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30496  poly16x8_t __ret;
30497  __ret = (poly16x8_t)(__p0);
30498  return __ret;
30499 }
30500 #endif
30501 
30502 #ifdef __LITTLE_ENDIAN__
30503 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30504  poly16x8_t __ret;
30505  __ret = (poly16x8_t)(__p0);
30506  return __ret;
30507 }
30508 #else
30509 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30510  poly16x8_t __ret;
30511  __ret = (poly16x8_t)(__p0);
30512  return __ret;
30513 }
30514 #endif
30515 
30516 #ifdef __LITTLE_ENDIAN__
30517 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30518  poly16x8_t __ret;
30519  __ret = (poly16x8_t)(__p0);
30520  return __ret;
30521 }
30522 #else
30523 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30524  poly16x8_t __ret;
30525  __ret = (poly16x8_t)(__p0);
30526  return __ret;
30527 }
30528 #endif
30529 
30530 #ifdef __LITTLE_ENDIAN__
30531 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30532  poly16x8_t __ret;
30533  __ret = (poly16x8_t)(__p0);
30534  return __ret;
30535 }
30536 #else
30537 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30538  poly16x8_t __ret;
30539  __ret = (poly16x8_t)(__p0);
30540  return __ret;
30541 }
30542 #endif
30543 
30544 #ifdef __LITTLE_ENDIAN__
30545 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30546  uint8x16_t __ret;
30547  __ret = (uint8x16_t)(__p0);
30548  return __ret;
30549 }
30550 #else
30551 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30552  uint8x16_t __ret;
30553  __ret = (uint8x16_t)(__p0);
30554  return __ret;
30555 }
30556 #endif
30557 
30558 #ifdef __LITTLE_ENDIAN__
30559 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30560  uint8x16_t __ret;
30561  __ret = (uint8x16_t)(__p0);
30562  return __ret;
30563 }
30564 #else
30565 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30566  uint8x16_t __ret;
30567  __ret = (uint8x16_t)(__p0);
30568  return __ret;
30569 }
30570 #endif
30571 
30572 #ifdef __LITTLE_ENDIAN__
30573 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30574  uint8x16_t __ret;
30575  __ret = (uint8x16_t)(__p0);
30576  return __ret;
30577 }
30578 #else
30579 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30580  uint8x16_t __ret;
30581  __ret = (uint8x16_t)(__p0);
30582  return __ret;
30583 }
30584 #endif
30585 
30586 #ifdef __LITTLE_ENDIAN__
30587 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30588  uint8x16_t __ret;
30589  __ret = (uint8x16_t)(__p0);
30590  return __ret;
30591 }
30592 #else
30593 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30594  uint8x16_t __ret;
30595  __ret = (uint8x16_t)(__p0);
30596  return __ret;
30597 }
30598 #endif
30599 
30600 #ifdef __LITTLE_ENDIAN__
30601 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30602  uint8x16_t __ret;
30603  __ret = (uint8x16_t)(__p0);
30604  return __ret;
30605 }
30606 #else
30607 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30608  uint8x16_t __ret;
30609  __ret = (uint8x16_t)(__p0);
30610  return __ret;
30611 }
30612 #endif
30613 
30614 #ifdef __LITTLE_ENDIAN__
30615 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30616  uint8x16_t __ret;
30617  __ret = (uint8x16_t)(__p0);
30618  return __ret;
30619 }
30620 #else
30621 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30622  uint8x16_t __ret;
30623  __ret = (uint8x16_t)(__p0);
30624  return __ret;
30625 }
30626 #endif
30627 
30628 #ifdef __LITTLE_ENDIAN__
30629 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30630  uint8x16_t __ret;
30631  __ret = (uint8x16_t)(__p0);
30632  return __ret;
30633 }
30634 #else
30635 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30636  uint8x16_t __ret;
30637  __ret = (uint8x16_t)(__p0);
30638  return __ret;
30639 }
30640 #endif
30641 
30642 #ifdef __LITTLE_ENDIAN__
30643 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30644  uint8x16_t __ret;
30645  __ret = (uint8x16_t)(__p0);
30646  return __ret;
30647 }
30648 #else
30649 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30650  uint8x16_t __ret;
30651  __ret = (uint8x16_t)(__p0);
30652  return __ret;
30653 }
30654 #endif
30655 
30656 #ifdef __LITTLE_ENDIAN__
30657 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30658  uint8x16_t __ret;
30659  __ret = (uint8x16_t)(__p0);
30660  return __ret;
30661 }
30662 #else
30663 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30664  uint8x16_t __ret;
30665  __ret = (uint8x16_t)(__p0);
30666  return __ret;
30667 }
30668 #endif
30669 
30670 #ifdef __LITTLE_ENDIAN__
30671 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30672  uint8x16_t __ret;
30673  __ret = (uint8x16_t)(__p0);
30674  return __ret;
30675 }
30676 #else
30677 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30678  uint8x16_t __ret;
30679  __ret = (uint8x16_t)(__p0);
30680  return __ret;
30681 }
30682 #endif
30683 
30684 #ifdef __LITTLE_ENDIAN__
30685 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30686  uint8x16_t __ret;
30687  __ret = (uint8x16_t)(__p0);
30688  return __ret;
30689 }
30690 #else
30691 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30692  uint8x16_t __ret;
30693  __ret = (uint8x16_t)(__p0);
30694  return __ret;
30695 }
30696 #endif
30697 
30698 #ifdef __LITTLE_ENDIAN__
30699 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30700  uint32x4_t __ret;
30701  __ret = (uint32x4_t)(__p0);
30702  return __ret;
30703 }
30704 #else
30705 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30706  uint32x4_t __ret;
30707  __ret = (uint32x4_t)(__p0);
30708  return __ret;
30709 }
30710 #endif
30711 
30712 #ifdef __LITTLE_ENDIAN__
30713 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30714  uint32x4_t __ret;
30715  __ret = (uint32x4_t)(__p0);
30716  return __ret;
30717 }
30718 #else
30719 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30720  uint32x4_t __ret;
30721  __ret = (uint32x4_t)(__p0);
30722  return __ret;
30723 }
30724 #endif
30725 
30726 #ifdef __LITTLE_ENDIAN__
30727 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30728  uint32x4_t __ret;
30729  __ret = (uint32x4_t)(__p0);
30730  return __ret;
30731 }
30732 #else
30733 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30734  uint32x4_t __ret;
30735  __ret = (uint32x4_t)(__p0);
30736  return __ret;
30737 }
30738 #endif
30739 
30740 #ifdef __LITTLE_ENDIAN__
30741 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30742  uint32x4_t __ret;
30743  __ret = (uint32x4_t)(__p0);
30744  return __ret;
30745 }
30746 #else
30747 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30748  uint32x4_t __ret;
30749  __ret = (uint32x4_t)(__p0);
30750  return __ret;
30751 }
30752 #endif
30753 
30754 #ifdef __LITTLE_ENDIAN__
30755 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30756  uint32x4_t __ret;
30757  __ret = (uint32x4_t)(__p0);
30758  return __ret;
30759 }
30760 #else
30761 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30762  uint32x4_t __ret;
30763  __ret = (uint32x4_t)(__p0);
30764  return __ret;
30765 }
30766 #endif
30767 
30768 #ifdef __LITTLE_ENDIAN__
30769 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30770  uint32x4_t __ret;
30771  __ret = (uint32x4_t)(__p0);
30772  return __ret;
30773 }
30774 #else
30775 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30776  uint32x4_t __ret;
30777  __ret = (uint32x4_t)(__p0);
30778  return __ret;
30779 }
30780 #endif
30781 
30782 #ifdef __LITTLE_ENDIAN__
30783 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30784  uint32x4_t __ret;
30785  __ret = (uint32x4_t)(__p0);
30786  return __ret;
30787 }
30788 #else
30789 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30790  uint32x4_t __ret;
30791  __ret = (uint32x4_t)(__p0);
30792  return __ret;
30793 }
30794 #endif
30795 
30796 #ifdef __LITTLE_ENDIAN__
30797 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30798  uint32x4_t __ret;
30799  __ret = (uint32x4_t)(__p0);
30800  return __ret;
30801 }
30802 #else
30803 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30804  uint32x4_t __ret;
30805  __ret = (uint32x4_t)(__p0);
30806  return __ret;
30807 }
30808 #endif
30809 
30810 #ifdef __LITTLE_ENDIAN__
30811 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30812  uint32x4_t __ret;
30813  __ret = (uint32x4_t)(__p0);
30814  return __ret;
30815 }
30816 #else
30817 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30818  uint32x4_t __ret;
30819  __ret = (uint32x4_t)(__p0);
30820  return __ret;
30821 }
30822 #endif
30823 
30824 #ifdef __LITTLE_ENDIAN__
30825 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30826  uint32x4_t __ret;
30827  __ret = (uint32x4_t)(__p0);
30828  return __ret;
30829 }
30830 #else
30831 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30832  uint32x4_t __ret;
30833  __ret = (uint32x4_t)(__p0);
30834  return __ret;
30835 }
30836 #endif
30837 
30838 #ifdef __LITTLE_ENDIAN__
30839 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30840  uint32x4_t __ret;
30841  __ret = (uint32x4_t)(__p0);
30842  return __ret;
30843 }
30844 #else
30845 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30846  uint32x4_t __ret;
30847  __ret = (uint32x4_t)(__p0);
30848  return __ret;
30849 }
30850 #endif
30851 
30852 #ifdef __LITTLE_ENDIAN__
30853 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30854  uint64x2_t __ret;
30855  __ret = (uint64x2_t)(__p0);
30856  return __ret;
30857 }
30858 #else
30859 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30860  uint64x2_t __ret;
30861  __ret = (uint64x2_t)(__p0);
30862  return __ret;
30863 }
30864 #endif
30865 
30866 #ifdef __LITTLE_ENDIAN__
30867 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30868  uint64x2_t __ret;
30869  __ret = (uint64x2_t)(__p0);
30870  return __ret;
30871 }
30872 #else
30873 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30874  uint64x2_t __ret;
30875  __ret = (uint64x2_t)(__p0);
30876  return __ret;
30877 }
30878 #endif
30879 
30880 #ifdef __LITTLE_ENDIAN__
30881 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30882  uint64x2_t __ret;
30883  __ret = (uint64x2_t)(__p0);
30884  return __ret;
30885 }
30886 #else
30887 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30888  uint64x2_t __ret;
30889  __ret = (uint64x2_t)(__p0);
30890  return __ret;
30891 }
30892 #endif
30893 
30894 #ifdef __LITTLE_ENDIAN__
30895 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30896  uint64x2_t __ret;
30897  __ret = (uint64x2_t)(__p0);
30898  return __ret;
30899 }
30900 #else
30901 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30902  uint64x2_t __ret;
30903  __ret = (uint64x2_t)(__p0);
30904  return __ret;
30905 }
30906 #endif
30907 
30908 #ifdef __LITTLE_ENDIAN__
30909 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30910  uint64x2_t __ret;
30911  __ret = (uint64x2_t)(__p0);
30912  return __ret;
30913 }
30914 #else
30915 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30916  uint64x2_t __ret;
30917  __ret = (uint64x2_t)(__p0);
30918  return __ret;
30919 }
30920 #endif
30921 
30922 #ifdef __LITTLE_ENDIAN__
30923 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30924  uint64x2_t __ret;
30925  __ret = (uint64x2_t)(__p0);
30926  return __ret;
30927 }
30928 #else
30929 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30930  uint64x2_t __ret;
30931  __ret = (uint64x2_t)(__p0);
30932  return __ret;
30933 }
30934 #endif
30935 
30936 #ifdef __LITTLE_ENDIAN__
30937 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30938  uint64x2_t __ret;
30939  __ret = (uint64x2_t)(__p0);
30940  return __ret;
30941 }
30942 #else
30943 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30944  uint64x2_t __ret;
30945  __ret = (uint64x2_t)(__p0);
30946  return __ret;
30947 }
30948 #endif
30949 
30950 #ifdef __LITTLE_ENDIAN__
30951 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30952  uint64x2_t __ret;
30953  __ret = (uint64x2_t)(__p0);
30954  return __ret;
30955 }
30956 #else
30957 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30958  uint64x2_t __ret;
30959  __ret = (uint64x2_t)(__p0);
30960  return __ret;
30961 }
30962 #endif
30963 
30964 #ifdef __LITTLE_ENDIAN__
30965 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30966  uint64x2_t __ret;
30967  __ret = (uint64x2_t)(__p0);
30968  return __ret;
30969 }
30970 #else
30971 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30972  uint64x2_t __ret;
30973  __ret = (uint64x2_t)(__p0);
30974  return __ret;
30975 }
30976 #endif
30977 
30978 #ifdef __LITTLE_ENDIAN__
30979 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30980  uint64x2_t __ret;
30981  __ret = (uint64x2_t)(__p0);
30982  return __ret;
30983 }
30984 #else
30985 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30986  uint64x2_t __ret;
30987  __ret = (uint64x2_t)(__p0);
30988  return __ret;
30989 }
30990 #endif
30991 
30992 #ifdef __LITTLE_ENDIAN__
30993 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30994  uint64x2_t __ret;
30995  __ret = (uint64x2_t)(__p0);
30996  return __ret;
30997 }
30998 #else
30999 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
31000  uint64x2_t __ret;
31001  __ret = (uint64x2_t)(__p0);
31002  return __ret;
31003 }
31004 #endif
31005 
31006 #ifdef __LITTLE_ENDIAN__
31007 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
31008  uint16x8_t __ret;
31009  __ret = (uint16x8_t)(__p0);
31010  return __ret;
31011 }
31012 #else
31013 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
31014  uint16x8_t __ret;
31015  __ret = (uint16x8_t)(__p0);
31016  return __ret;
31017 }
31018 #endif
31019 
31020 #ifdef __LITTLE_ENDIAN__
31021 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
31022  uint16x8_t __ret;
31023  __ret = (uint16x8_t)(__p0);
31024  return __ret;
31025 }
31026 #else
31027 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
31028  uint16x8_t __ret;
31029  __ret = (uint16x8_t)(__p0);
31030  return __ret;
31031 }
31032 #endif
31033 
31034 #ifdef __LITTLE_ENDIAN__
31035 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
31036  uint16x8_t __ret;
31037  __ret = (uint16x8_t)(__p0);
31038  return __ret;
31039 }
31040 #else
31041 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
31042  uint16x8_t __ret;
31043  __ret = (uint16x8_t)(__p0);
31044  return __ret;
31045 }
31046 #endif
31047 
31048 #ifdef __LITTLE_ENDIAN__
31049 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31050  uint16x8_t __ret;
31051  __ret = (uint16x8_t)(__p0);
31052  return __ret;
31053 }
31054 #else
31055 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31056  uint16x8_t __ret;
31057  __ret = (uint16x8_t)(__p0);
31058  return __ret;
31059 }
31060 #endif
31061 
31062 #ifdef __LITTLE_ENDIAN__
31063 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31064  uint16x8_t __ret;
31065  __ret = (uint16x8_t)(__p0);
31066  return __ret;
31067 }
31068 #else
31069 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31070  uint16x8_t __ret;
31071  __ret = (uint16x8_t)(__p0);
31072  return __ret;
31073 }
31074 #endif
31075 
31076 #ifdef __LITTLE_ENDIAN__
31077 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31078  uint16x8_t __ret;
31079  __ret = (uint16x8_t)(__p0);
31080  return __ret;
31081 }
31082 #else
31083 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31084  uint16x8_t __ret;
31085  __ret = (uint16x8_t)(__p0);
31086  return __ret;
31087 }
31088 #endif
31089 
31090 #ifdef __LITTLE_ENDIAN__
31091 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31092  uint16x8_t __ret;
31093  __ret = (uint16x8_t)(__p0);
31094  return __ret;
31095 }
31096 #else
31097 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31098  uint16x8_t __ret;
31099  __ret = (uint16x8_t)(__p0);
31100  return __ret;
31101 }
31102 #endif
31103 
31104 #ifdef __LITTLE_ENDIAN__
31105 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31106  uint16x8_t __ret;
31107  __ret = (uint16x8_t)(__p0);
31108  return __ret;
31109 }
31110 #else
31111 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31112  uint16x8_t __ret;
31113  __ret = (uint16x8_t)(__p0);
31114  return __ret;
31115 }
31116 #endif
31117 
31118 #ifdef __LITTLE_ENDIAN__
31119 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31120  uint16x8_t __ret;
31121  __ret = (uint16x8_t)(__p0);
31122  return __ret;
31123 }
31124 #else
31125 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31126  uint16x8_t __ret;
31127  __ret = (uint16x8_t)(__p0);
31128  return __ret;
31129 }
31130 #endif
31131 
31132 #ifdef __LITTLE_ENDIAN__
31133 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31134  uint16x8_t __ret;
31135  __ret = (uint16x8_t)(__p0);
31136  return __ret;
31137 }
31138 #else
31139 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31140  uint16x8_t __ret;
31141  __ret = (uint16x8_t)(__p0);
31142  return __ret;
31143 }
31144 #endif
31145 
31146 #ifdef __LITTLE_ENDIAN__
31147 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31148  uint16x8_t __ret;
31149  __ret = (uint16x8_t)(__p0);
31150  return __ret;
31151 }
31152 #else
31153 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31154  uint16x8_t __ret;
31155  __ret = (uint16x8_t)(__p0);
31156  return __ret;
31157 }
31158 #endif
31159 
31160 #ifdef __LITTLE_ENDIAN__
31161 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31162  int8x16_t __ret;
31163  __ret = (int8x16_t)(__p0);
31164  return __ret;
31165 }
31166 #else
31167 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31168  int8x16_t __ret;
31169  __ret = (int8x16_t)(__p0);
31170  return __ret;
31171 }
31172 #endif
31173 
31174 #ifdef __LITTLE_ENDIAN__
31175 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31176  int8x16_t __ret;
31177  __ret = (int8x16_t)(__p0);
31178  return __ret;
31179 }
31180 #else
31181 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31182  int8x16_t __ret;
31183  __ret = (int8x16_t)(__p0);
31184  return __ret;
31185 }
31186 #endif
31187 
31188 #ifdef __LITTLE_ENDIAN__
31189 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31190  int8x16_t __ret;
31191  __ret = (int8x16_t)(__p0);
31192  return __ret;
31193 }
31194 #else
31195 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31196  int8x16_t __ret;
31197  __ret = (int8x16_t)(__p0);
31198  return __ret;
31199 }
31200 #endif
31201 
31202 #ifdef __LITTLE_ENDIAN__
31203 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31204  int8x16_t __ret;
31205  __ret = (int8x16_t)(__p0);
31206  return __ret;
31207 }
31208 #else
31209 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31210  int8x16_t __ret;
31211  __ret = (int8x16_t)(__p0);
31212  return __ret;
31213 }
31214 #endif
31215 
31216 #ifdef __LITTLE_ENDIAN__
31217 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31218  int8x16_t __ret;
31219  __ret = (int8x16_t)(__p0);
31220  return __ret;
31221 }
31222 #else
31223 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31224  int8x16_t __ret;
31225  __ret = (int8x16_t)(__p0);
31226  return __ret;
31227 }
31228 #endif
31229 
31230 #ifdef __LITTLE_ENDIAN__
31231 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31232  int8x16_t __ret;
31233  __ret = (int8x16_t)(__p0);
31234  return __ret;
31235 }
31236 #else
31237 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31238  int8x16_t __ret;
31239  __ret = (int8x16_t)(__p0);
31240  return __ret;
31241 }
31242 #endif
31243 
31244 #ifdef __LITTLE_ENDIAN__
31245 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31246  int8x16_t __ret;
31247  __ret = (int8x16_t)(__p0);
31248  return __ret;
31249 }
31250 #else
31251 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31252  int8x16_t __ret;
31253  __ret = (int8x16_t)(__p0);
31254  return __ret;
31255 }
31256 #endif
31257 
31258 #ifdef __LITTLE_ENDIAN__
31259 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31260  int8x16_t __ret;
31261  __ret = (int8x16_t)(__p0);
31262  return __ret;
31263 }
31264 #else
31265 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31266  int8x16_t __ret;
31267  __ret = (int8x16_t)(__p0);
31268  return __ret;
31269 }
31270 #endif
31271 
31272 #ifdef __LITTLE_ENDIAN__
31273 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31274  int8x16_t __ret;
31275  __ret = (int8x16_t)(__p0);
31276  return __ret;
31277 }
31278 #else
31279 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31280  int8x16_t __ret;
31281  __ret = (int8x16_t)(__p0);
31282  return __ret;
31283 }
31284 #endif
31285 
31286 #ifdef __LITTLE_ENDIAN__
31287 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31288  int8x16_t __ret;
31289  __ret = (int8x16_t)(__p0);
31290  return __ret;
31291 }
31292 #else
31293 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31294  int8x16_t __ret;
31295  __ret = (int8x16_t)(__p0);
31296  return __ret;
31297 }
31298 #endif
31299 
31300 #ifdef __LITTLE_ENDIAN__
31301 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31302  int8x16_t __ret;
31303  __ret = (int8x16_t)(__p0);
31304  return __ret;
31305 }
31306 #else
31307 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31308  int8x16_t __ret;
31309  __ret = (int8x16_t)(__p0);
31310  return __ret;
31311 }
31312 #endif
31313 
31314 #ifdef __LITTLE_ENDIAN__
31315 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31316  float32x4_t __ret;
31317  __ret = (float32x4_t)(__p0);
31318  return __ret;
31319 }
31320 #else
31321 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31322  float32x4_t __ret;
31323  __ret = (float32x4_t)(__p0);
31324  return __ret;
31325 }
31326 #endif
31327 
31328 #ifdef __LITTLE_ENDIAN__
31329 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31330  float32x4_t __ret;
31331  __ret = (float32x4_t)(__p0);
31332  return __ret;
31333 }
31334 #else
31335 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31336  float32x4_t __ret;
31337  __ret = (float32x4_t)(__p0);
31338  return __ret;
31339 }
31340 #endif
31341 
31342 #ifdef __LITTLE_ENDIAN__
31343 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31344  float32x4_t __ret;
31345  __ret = (float32x4_t)(__p0);
31346  return __ret;
31347 }
31348 #else
31349 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31350  float32x4_t __ret;
31351  __ret = (float32x4_t)(__p0);
31352  return __ret;
31353 }
31354 #endif
31355 
31356 #ifdef __LITTLE_ENDIAN__
31357 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31358  float32x4_t __ret;
31359  __ret = (float32x4_t)(__p0);
31360  return __ret;
31361 }
31362 #else
31363 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31364  float32x4_t __ret;
31365  __ret = (float32x4_t)(__p0);
31366  return __ret;
31367 }
31368 #endif
31369 
31370 #ifdef __LITTLE_ENDIAN__
31371 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31372  float32x4_t __ret;
31373  __ret = (float32x4_t)(__p0);
31374  return __ret;
31375 }
31376 #else
31377 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31378  float32x4_t __ret;
31379  __ret = (float32x4_t)(__p0);
31380  return __ret;
31381 }
31382 #endif
31383 
31384 #ifdef __LITTLE_ENDIAN__
31385 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31386  float32x4_t __ret;
31387  __ret = (float32x4_t)(__p0);
31388  return __ret;
31389 }
31390 #else
31391 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31392  float32x4_t __ret;
31393  __ret = (float32x4_t)(__p0);
31394  return __ret;
31395 }
31396 #endif
31397 
31398 #ifdef __LITTLE_ENDIAN__
31399 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31400  float32x4_t __ret;
31401  __ret = (float32x4_t)(__p0);
31402  return __ret;
31403 }
31404 #else
31405 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31406  float32x4_t __ret;
31407  __ret = (float32x4_t)(__p0);
31408  return __ret;
31409 }
31410 #endif
31411 
31412 #ifdef __LITTLE_ENDIAN__
31413 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31414  float32x4_t __ret;
31415  __ret = (float32x4_t)(__p0);
31416  return __ret;
31417 }
31418 #else
31419 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31420  float32x4_t __ret;
31421  __ret = (float32x4_t)(__p0);
31422  return __ret;
31423 }
31424 #endif
31425 
31426 #ifdef __LITTLE_ENDIAN__
31427 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31428  float32x4_t __ret;
31429  __ret = (float32x4_t)(__p0);
31430  return __ret;
31431 }
31432 #else
31433 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31434  float32x4_t __ret;
31435  __ret = (float32x4_t)(__p0);
31436  return __ret;
31437 }
31438 #endif
31439 
31440 #ifdef __LITTLE_ENDIAN__
31441 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31442  float32x4_t __ret;
31443  __ret = (float32x4_t)(__p0);
31444  return __ret;
31445 }
31446 #else
31447 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31448  float32x4_t __ret;
31449  __ret = (float32x4_t)(__p0);
31450  return __ret;
31451 }
31452 #endif
31453 
31454 #ifdef __LITTLE_ENDIAN__
31455 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31456  float32x4_t __ret;
31457  __ret = (float32x4_t)(__p0);
31458  return __ret;
31459 }
31460 #else
31461 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31462  float32x4_t __ret;
31463  __ret = (float32x4_t)(__p0);
31464  return __ret;
31465 }
31466 #endif
31467 
31468 #ifdef __LITTLE_ENDIAN__
31469 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31470  float16x8_t __ret;
31471  __ret = (float16x8_t)(__p0);
31472  return __ret;
31473 }
31474 #else
31475 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31476  float16x8_t __ret;
31477  __ret = (float16x8_t)(__p0);
31478  return __ret;
31479 }
31480 #endif
31481 
31482 #ifdef __LITTLE_ENDIAN__
31483 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31484  float16x8_t __ret;
31485  __ret = (float16x8_t)(__p0);
31486  return __ret;
31487 }
31488 #else
31489 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31490  float16x8_t __ret;
31491  __ret = (float16x8_t)(__p0);
31492  return __ret;
31493 }
31494 #endif
31495 
31496 #ifdef __LITTLE_ENDIAN__
31497 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31498  float16x8_t __ret;
31499  __ret = (float16x8_t)(__p0);
31500  return __ret;
31501 }
31502 #else
31503 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31504  float16x8_t __ret;
31505  __ret = (float16x8_t)(__p0);
31506  return __ret;
31507 }
31508 #endif
31509 
31510 #ifdef __LITTLE_ENDIAN__
31511 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31512  float16x8_t __ret;
31513  __ret = (float16x8_t)(__p0);
31514  return __ret;
31515 }
31516 #else
31517 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31518  float16x8_t __ret;
31519  __ret = (float16x8_t)(__p0);
31520  return __ret;
31521 }
31522 #endif
31523 
31524 #ifdef __LITTLE_ENDIAN__
31525 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31526  float16x8_t __ret;
31527  __ret = (float16x8_t)(__p0);
31528  return __ret;
31529 }
31530 #else
31531 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31532  float16x8_t __ret;
31533  __ret = (float16x8_t)(__p0);
31534  return __ret;
31535 }
31536 #endif
31537 
31538 #ifdef __LITTLE_ENDIAN__
31539 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31540  float16x8_t __ret;
31541  __ret = (float16x8_t)(__p0);
31542  return __ret;
31543 }
31544 #else
31545 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31546  float16x8_t __ret;
31547  __ret = (float16x8_t)(__p0);
31548  return __ret;
31549 }
31550 #endif
31551 
31552 #ifdef __LITTLE_ENDIAN__
31553 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31554  float16x8_t __ret;
31555  __ret = (float16x8_t)(__p0);
31556  return __ret;
31557 }
31558 #else
31559 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31560  float16x8_t __ret;
31561  __ret = (float16x8_t)(__p0);
31562  return __ret;
31563 }
31564 #endif
31565 
31566 #ifdef __LITTLE_ENDIAN__
31567 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31568  float16x8_t __ret;
31569  __ret = (float16x8_t)(__p0);
31570  return __ret;
31571 }
31572 #else
31573 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31574  float16x8_t __ret;
31575  __ret = (float16x8_t)(__p0);
31576  return __ret;
31577 }
31578 #endif
31579 
31580 #ifdef __LITTLE_ENDIAN__
31581 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31582  float16x8_t __ret;
31583  __ret = (float16x8_t)(__p0);
31584  return __ret;
31585 }
31586 #else
31587 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31588  float16x8_t __ret;
31589  __ret = (float16x8_t)(__p0);
31590  return __ret;
31591 }
31592 #endif
31593 
31594 #ifdef __LITTLE_ENDIAN__
31595 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31596  float16x8_t __ret;
31597  __ret = (float16x8_t)(__p0);
31598  return __ret;
31599 }
31600 #else
31601 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31602  float16x8_t __ret;
31603  __ret = (float16x8_t)(__p0);
31604  return __ret;
31605 }
31606 #endif
31607 
31608 #ifdef __LITTLE_ENDIAN__
31609 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31610  float16x8_t __ret;
31611  __ret = (float16x8_t)(__p0);
31612  return __ret;
31613 }
31614 #else
31615 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31616  float16x8_t __ret;
31617  __ret = (float16x8_t)(__p0);
31618  return __ret;
31619 }
31620 #endif
31621 
31622 #ifdef __LITTLE_ENDIAN__
31623 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31624  int32x4_t __ret;
31625  __ret = (int32x4_t)(__p0);
31626  return __ret;
31627 }
31628 #else
31629 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31630  int32x4_t __ret;
31631  __ret = (int32x4_t)(__p0);
31632  return __ret;
31633 }
31634 #endif
31635 
31636 #ifdef __LITTLE_ENDIAN__
31637 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31638  int32x4_t __ret;
31639  __ret = (int32x4_t)(__p0);
31640  return __ret;
31641 }
31642 #else
31643 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31644  int32x4_t __ret;
31645  __ret = (int32x4_t)(__p0);
31646  return __ret;
31647 }
31648 #endif
31649 
31650 #ifdef __LITTLE_ENDIAN__
31651 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31652  int32x4_t __ret;
31653  __ret = (int32x4_t)(__p0);
31654  return __ret;
31655 }
31656 #else
31657 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31658  int32x4_t __ret;
31659  __ret = (int32x4_t)(__p0);
31660  return __ret;
31661 }
31662 #endif
31663 
31664 #ifdef __LITTLE_ENDIAN__
31665 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31666  int32x4_t __ret;
31667  __ret = (int32x4_t)(__p0);
31668  return __ret;
31669 }
31670 #else
31671 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31672  int32x4_t __ret;
31673  __ret = (int32x4_t)(__p0);
31674  return __ret;
31675 }
31676 #endif
31677 
31678 #ifdef __LITTLE_ENDIAN__
31679 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31680  int32x4_t __ret;
31681  __ret = (int32x4_t)(__p0);
31682  return __ret;
31683 }
31684 #else
31685 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31686  int32x4_t __ret;
31687  __ret = (int32x4_t)(__p0);
31688  return __ret;
31689 }
31690 #endif
31691 
31692 #ifdef __LITTLE_ENDIAN__
31693 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31694  int32x4_t __ret;
31695  __ret = (int32x4_t)(__p0);
31696  return __ret;
31697 }
31698 #else
31699 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31700  int32x4_t __ret;
31701  __ret = (int32x4_t)(__p0);
31702  return __ret;
31703 }
31704 #endif
31705 
31706 #ifdef __LITTLE_ENDIAN__
31707 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31708  int32x4_t __ret;
31709  __ret = (int32x4_t)(__p0);
31710  return __ret;
31711 }
31712 #else
31713 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31714  int32x4_t __ret;
31715  __ret = (int32x4_t)(__p0);
31716  return __ret;
31717 }
31718 #endif
31719 
31720 #ifdef __LITTLE_ENDIAN__
31721 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31722  int32x4_t __ret;
31723  __ret = (int32x4_t)(__p0);
31724  return __ret;
31725 }
31726 #else
31727 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31728  int32x4_t __ret;
31729  __ret = (int32x4_t)(__p0);
31730  return __ret;
31731 }
31732 #endif
31733 
31734 #ifdef __LITTLE_ENDIAN__
31735 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31736  int32x4_t __ret;
31737  __ret = (int32x4_t)(__p0);
31738  return __ret;
31739 }
31740 #else
31741 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31742  int32x4_t __ret;
31743  __ret = (int32x4_t)(__p0);
31744  return __ret;
31745 }
31746 #endif
31747 
31748 #ifdef __LITTLE_ENDIAN__
31749 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31750  int32x4_t __ret;
31751  __ret = (int32x4_t)(__p0);
31752  return __ret;
31753 }
31754 #else
31755 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31756  int32x4_t __ret;
31757  __ret = (int32x4_t)(__p0);
31758  return __ret;
31759 }
31760 #endif
31761 
31762 #ifdef __LITTLE_ENDIAN__
31763 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31764  int32x4_t __ret;
31765  __ret = (int32x4_t)(__p0);
31766  return __ret;
31767 }
31768 #else
31769 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31770  int32x4_t __ret;
31771  __ret = (int32x4_t)(__p0);
31772  return __ret;
31773 }
31774 #endif
31775 
31776 #ifdef __LITTLE_ENDIAN__
31777 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31778  int64x2_t __ret;
31779  __ret = (int64x2_t)(__p0);
31780  return __ret;
31781 }
31782 #else
31783 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31784  int64x2_t __ret;
31785  __ret = (int64x2_t)(__p0);
31786  return __ret;
31787 }
31788 #endif
31789 
31790 #ifdef __LITTLE_ENDIAN__
31791 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31792  int64x2_t __ret;
31793  __ret = (int64x2_t)(__p0);
31794  return __ret;
31795 }
31796 #else
31797 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31798  int64x2_t __ret;
31799  __ret = (int64x2_t)(__p0);
31800  return __ret;
31801 }
31802 #endif
31803 
31804 #ifdef __LITTLE_ENDIAN__
31805 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31806  int64x2_t __ret;
31807  __ret = (int64x2_t)(__p0);
31808  return __ret;
31809 }
31810 #else
31811 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31812  int64x2_t __ret;
31813  __ret = (int64x2_t)(__p0);
31814  return __ret;
31815 }
31816 #endif
31817 
31818 #ifdef __LITTLE_ENDIAN__
31819 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31820  int64x2_t __ret;
31821  __ret = (int64x2_t)(__p0);
31822  return __ret;
31823 }
31824 #else
31825 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31826  int64x2_t __ret;
31827  __ret = (int64x2_t)(__p0);
31828  return __ret;
31829 }
31830 #endif
31831 
31832 #ifdef __LITTLE_ENDIAN__
31833 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31834  int64x2_t __ret;
31835  __ret = (int64x2_t)(__p0);
31836  return __ret;
31837 }
31838 #else
31839 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31840  int64x2_t __ret;
31841  __ret = (int64x2_t)(__p0);
31842  return __ret;
31843 }
31844 #endif
31845 
31846 #ifdef __LITTLE_ENDIAN__
31847 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31848  int64x2_t __ret;
31849  __ret = (int64x2_t)(__p0);
31850  return __ret;
31851 }
31852 #else
31853 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31854  int64x2_t __ret;
31855  __ret = (int64x2_t)(__p0);
31856  return __ret;
31857 }
31858 #endif
31859 
31860 #ifdef __LITTLE_ENDIAN__
31861 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31862  int64x2_t __ret;
31863  __ret = (int64x2_t)(__p0);
31864  return __ret;
31865 }
31866 #else
31867 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31868  int64x2_t __ret;
31869  __ret = (int64x2_t)(__p0);
31870  return __ret;
31871 }
31872 #endif
31873 
31874 #ifdef __LITTLE_ENDIAN__
31875 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31876  int64x2_t __ret;
31877  __ret = (int64x2_t)(__p0);
31878  return __ret;
31879 }
31880 #else
31881 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31882  int64x2_t __ret;
31883  __ret = (int64x2_t)(__p0);
31884  return __ret;
31885 }
31886 #endif
31887 
31888 #ifdef __LITTLE_ENDIAN__
31889 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31890  int64x2_t __ret;
31891  __ret = (int64x2_t)(__p0);
31892  return __ret;
31893 }
31894 #else
31895 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31896  int64x2_t __ret;
31897  __ret = (int64x2_t)(__p0);
31898  return __ret;
31899 }
31900 #endif
31901 
31902 #ifdef __LITTLE_ENDIAN__
31903 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31904  int64x2_t __ret;
31905  __ret = (int64x2_t)(__p0);
31906  return __ret;
31907 }
31908 #else
31909 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31910  int64x2_t __ret;
31911  __ret = (int64x2_t)(__p0);
31912  return __ret;
31913 }
31914 #endif
31915 
31916 #ifdef __LITTLE_ENDIAN__
31917 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31918  int64x2_t __ret;
31919  __ret = (int64x2_t)(__p0);
31920  return __ret;
31921 }
31922 #else
31923 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31924  int64x2_t __ret;
31925  __ret = (int64x2_t)(__p0);
31926  return __ret;
31927 }
31928 #endif
31929 
31930 #ifdef __LITTLE_ENDIAN__
31931 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31932  int16x8_t __ret;
31933  __ret = (int16x8_t)(__p0);
31934  return __ret;
31935 }
31936 #else
31937 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31938  int16x8_t __ret;
31939  __ret = (int16x8_t)(__p0);
31940  return __ret;
31941 }
31942 #endif
31943 
31944 #ifdef __LITTLE_ENDIAN__
31945 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31946  int16x8_t __ret;
31947  __ret = (int16x8_t)(__p0);
31948  return __ret;
31949 }
31950 #else
31951 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31952  int16x8_t __ret;
31953  __ret = (int16x8_t)(__p0);
31954  return __ret;
31955 }
31956 #endif
31957 
31958 #ifdef __LITTLE_ENDIAN__
31959 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31960  int16x8_t __ret;
31961  __ret = (int16x8_t)(__p0);
31962  return __ret;
31963 }
31964 #else
31965 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31966  int16x8_t __ret;
31967  __ret = (int16x8_t)(__p0);
31968  return __ret;
31969 }
31970 #endif
31971 
31972 #ifdef __LITTLE_ENDIAN__
31973 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31974  int16x8_t __ret;
31975  __ret = (int16x8_t)(__p0);
31976  return __ret;
31977 }
31978 #else
31979 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31980  int16x8_t __ret;
31981  __ret = (int16x8_t)(__p0);
31982  return __ret;
31983 }
31984 #endif
31985 
31986 #ifdef __LITTLE_ENDIAN__
31987 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31988  int16x8_t __ret;
31989  __ret = (int16x8_t)(__p0);
31990  return __ret;
31991 }
31992 #else
31993 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31994  int16x8_t __ret;
31995  __ret = (int16x8_t)(__p0);
31996  return __ret;
31997 }
31998 #endif
31999 
32000 #ifdef __LITTLE_ENDIAN__
32001 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
32002  int16x8_t __ret;
32003  __ret = (int16x8_t)(__p0);
32004  return __ret;
32005 }
32006 #else
32007 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
32008  int16x8_t __ret;
32009  __ret = (int16x8_t)(__p0);
32010  return __ret;
32011 }
32012 #endif
32013 
32014 #ifdef __LITTLE_ENDIAN__
32015 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
32016  int16x8_t __ret;
32017  __ret = (int16x8_t)(__p0);
32018  return __ret;
32019 }
32020 #else
32021 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
32022  int16x8_t __ret;
32023  __ret = (int16x8_t)(__p0);
32024  return __ret;
32025 }
32026 #endif
32027 
32028 #ifdef __LITTLE_ENDIAN__
32029 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
32030  int16x8_t __ret;
32031  __ret = (int16x8_t)(__p0);
32032  return __ret;
32033 }
32034 #else
32035 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
32036  int16x8_t __ret;
32037  __ret = (int16x8_t)(__p0);
32038  return __ret;
32039 }
32040 #endif
32041 
32042 #ifdef __LITTLE_ENDIAN__
32043 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32044  int16x8_t __ret;
32045  __ret = (int16x8_t)(__p0);
32046  return __ret;
32047 }
32048 #else
32049 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32050  int16x8_t __ret;
32051  __ret = (int16x8_t)(__p0);
32052  return __ret;
32053 }
32054 #endif
32055 
32056 #ifdef __LITTLE_ENDIAN__
32057 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32058  int16x8_t __ret;
32059  __ret = (int16x8_t)(__p0);
32060  return __ret;
32061 }
32062 #else
32063 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32064  int16x8_t __ret;
32065  __ret = (int16x8_t)(__p0);
32066  return __ret;
32067 }
32068 #endif
32069 
32070 #ifdef __LITTLE_ENDIAN__
32071 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32072  int16x8_t __ret;
32073  __ret = (int16x8_t)(__p0);
32074  return __ret;
32075 }
32076 #else
32077 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32078  int16x8_t __ret;
32079  __ret = (int16x8_t)(__p0);
32080  return __ret;
32081 }
32082 #endif
32083 
32084 #ifdef __LITTLE_ENDIAN__
32085 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32086  uint8x8_t __ret;
32087  __ret = (uint8x8_t)(__p0);
32088  return __ret;
32089 }
32090 #else
32091 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32092  uint8x8_t __ret;
32093  __ret = (uint8x8_t)(__p0);
32094  return __ret;
32095 }
32096 #endif
32097 
32098 #ifdef __LITTLE_ENDIAN__
32099 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32100  uint8x8_t __ret;
32101  __ret = (uint8x8_t)(__p0);
32102  return __ret;
32103 }
32104 #else
32105 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32106  uint8x8_t __ret;
32107  __ret = (uint8x8_t)(__p0);
32108  return __ret;
32109 }
32110 #endif
32111 
32112 #ifdef __LITTLE_ENDIAN__
32113 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32114  uint8x8_t __ret;
32115  __ret = (uint8x8_t)(__p0);
32116  return __ret;
32117 }
32118 #else
32119 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32120  uint8x8_t __ret;
32121  __ret = (uint8x8_t)(__p0);
32122  return __ret;
32123 }
32124 #endif
32125 
32126 #ifdef __LITTLE_ENDIAN__
32127 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32128  uint8x8_t __ret;
32129  __ret = (uint8x8_t)(__p0);
32130  return __ret;
32131 }
32132 #else
32133 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32134  uint8x8_t __ret;
32135  __ret = (uint8x8_t)(__p0);
32136  return __ret;
32137 }
32138 #endif
32139 
32140 #ifdef __LITTLE_ENDIAN__
32141 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32142  uint8x8_t __ret;
32143  __ret = (uint8x8_t)(__p0);
32144  return __ret;
32145 }
32146 #else
32147 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32148  uint8x8_t __ret;
32149  __ret = (uint8x8_t)(__p0);
32150  return __ret;
32151 }
32152 #endif
32153 
32154 #ifdef __LITTLE_ENDIAN__
32155 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32156  uint8x8_t __ret;
32157  __ret = (uint8x8_t)(__p0);
32158  return __ret;
32159 }
32160 #else
32161 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32162  uint8x8_t __ret;
32163  __ret = (uint8x8_t)(__p0);
32164  return __ret;
32165 }
32166 #endif
32167 
32168 #ifdef __LITTLE_ENDIAN__
32169 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32170  uint8x8_t __ret;
32171  __ret = (uint8x8_t)(__p0);
32172  return __ret;
32173 }
32174 #else
32175 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32176  uint8x8_t __ret;
32177  __ret = (uint8x8_t)(__p0);
32178  return __ret;
32179 }
32180 #endif
32181 
32182 #ifdef __LITTLE_ENDIAN__
32183 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32184  uint8x8_t __ret;
32185  __ret = (uint8x8_t)(__p0);
32186  return __ret;
32187 }
32188 #else
32189 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32190  uint8x8_t __ret;
32191  __ret = (uint8x8_t)(__p0);
32192  return __ret;
32193 }
32194 #endif
32195 
32196 #ifdef __LITTLE_ENDIAN__
32197 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32198  uint8x8_t __ret;
32199  __ret = (uint8x8_t)(__p0);
32200  return __ret;
32201 }
32202 #else
32203 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32204  uint8x8_t __ret;
32205  __ret = (uint8x8_t)(__p0);
32206  return __ret;
32207 }
32208 #endif
32209 
32210 #ifdef __LITTLE_ENDIAN__
32211 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32212  uint8x8_t __ret;
32213  __ret = (uint8x8_t)(__p0);
32214  return __ret;
32215 }
32216 #else
32217 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32218  uint8x8_t __ret;
32219  __ret = (uint8x8_t)(__p0);
32220  return __ret;
32221 }
32222 #endif
32223 
32224 #ifdef __LITTLE_ENDIAN__
32225 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32226  uint8x8_t __ret;
32227  __ret = (uint8x8_t)(__p0);
32228  return __ret;
32229 }
32230 #else
32231 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32232  uint8x8_t __ret;
32233  __ret = (uint8x8_t)(__p0);
32234  return __ret;
32235 }
32236 #endif
32237 
32238 #ifdef __LITTLE_ENDIAN__
32239 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32240  uint32x2_t __ret;
32241  __ret = (uint32x2_t)(__p0);
32242  return __ret;
32243 }
32244 #else
32245 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32246  uint32x2_t __ret;
32247  __ret = (uint32x2_t)(__p0);
32248  return __ret;
32249 }
32250 #endif
32251 
32252 #ifdef __LITTLE_ENDIAN__
32253 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32254  uint32x2_t __ret;
32255  __ret = (uint32x2_t)(__p0);
32256  return __ret;
32257 }
32258 #else
32259 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32260  uint32x2_t __ret;
32261  __ret = (uint32x2_t)(__p0);
32262  return __ret;
32263 }
32264 #endif
32265 
32266 #ifdef __LITTLE_ENDIAN__
32267 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32268  uint32x2_t __ret;
32269  __ret = (uint32x2_t)(__p0);
32270  return __ret;
32271 }
32272 #else
32273 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32274  uint32x2_t __ret;
32275  __ret = (uint32x2_t)(__p0);
32276  return __ret;
32277 }
32278 #endif
32279 
32280 #ifdef __LITTLE_ENDIAN__
32281 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32282  uint32x2_t __ret;
32283  __ret = (uint32x2_t)(__p0);
32284  return __ret;
32285 }
32286 #else
32287 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32288  uint32x2_t __ret;
32289  __ret = (uint32x2_t)(__p0);
32290  return __ret;
32291 }
32292 #endif
32293 
32294 #ifdef __LITTLE_ENDIAN__
32295 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32296  uint32x2_t __ret;
32297  __ret = (uint32x2_t)(__p0);
32298  return __ret;
32299 }
32300 #else
32301 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32302  uint32x2_t __ret;
32303  __ret = (uint32x2_t)(__p0);
32304  return __ret;
32305 }
32306 #endif
32307 
32308 #ifdef __LITTLE_ENDIAN__
32309 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32310  uint32x2_t __ret;
32311  __ret = (uint32x2_t)(__p0);
32312  return __ret;
32313 }
32314 #else
32315 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32316  uint32x2_t __ret;
32317  __ret = (uint32x2_t)(__p0);
32318  return __ret;
32319 }
32320 #endif
32321 
32322 #ifdef __LITTLE_ENDIAN__
32323 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32324  uint32x2_t __ret;
32325  __ret = (uint32x2_t)(__p0);
32326  return __ret;
32327 }
32328 #else
32329 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32330  uint32x2_t __ret;
32331  __ret = (uint32x2_t)(__p0);
32332  return __ret;
32333 }
32334 #endif
32335 
32336 #ifdef __LITTLE_ENDIAN__
32337 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32338  uint32x2_t __ret;
32339  __ret = (uint32x2_t)(__p0);
32340  return __ret;
32341 }
32342 #else
32343 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32344  uint32x2_t __ret;
32345  __ret = (uint32x2_t)(__p0);
32346  return __ret;
32347 }
32348 #endif
32349 
32350 #ifdef __LITTLE_ENDIAN__
32351 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32352  uint32x2_t __ret;
32353  __ret = (uint32x2_t)(__p0);
32354  return __ret;
32355 }
32356 #else
32357 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32358  uint32x2_t __ret;
32359  __ret = (uint32x2_t)(__p0);
32360  return __ret;
32361 }
32362 #endif
32363 
32364 #ifdef __LITTLE_ENDIAN__
32365 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32366  uint32x2_t __ret;
32367  __ret = (uint32x2_t)(__p0);
32368  return __ret;
32369 }
32370 #else
32371 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32372  uint32x2_t __ret;
32373  __ret = (uint32x2_t)(__p0);
32374  return __ret;
32375 }
32376 #endif
32377 
32378 #ifdef __LITTLE_ENDIAN__
32379 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32380  uint32x2_t __ret;
32381  __ret = (uint32x2_t)(__p0);
32382  return __ret;
32383 }
32384 #else
32385 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32386  uint32x2_t __ret;
32387  __ret = (uint32x2_t)(__p0);
32388  return __ret;
32389 }
32390 #endif
32391 
32392 #ifdef __LITTLE_ENDIAN__
32393 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32394  uint64x1_t __ret;
32395  __ret = (uint64x1_t)(__p0);
32396  return __ret;
32397 }
32398 #else
32399 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32400  uint64x1_t __ret;
32401  __ret = (uint64x1_t)(__p0);
32402  return __ret;
32403 }
32404 #endif
32405 
32406 #ifdef __LITTLE_ENDIAN__
32407 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32408  uint64x1_t __ret;
32409  __ret = (uint64x1_t)(__p0);
32410  return __ret;
32411 }
32412 #else
32413 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32414  uint64x1_t __ret;
32415  __ret = (uint64x1_t)(__p0);
32416  return __ret;
32417 }
32418 #endif
32419 
32420 #ifdef __LITTLE_ENDIAN__
32421 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32422  uint64x1_t __ret;
32423  __ret = (uint64x1_t)(__p0);
32424  return __ret;
32425 }
32426 #else
32427 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32428  uint64x1_t __ret;
32429  __ret = (uint64x1_t)(__p0);
32430  return __ret;
32431 }
32432 #endif
32433 
32434 #ifdef __LITTLE_ENDIAN__
32435 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32436  uint64x1_t __ret;
32437  __ret = (uint64x1_t)(__p0);
32438  return __ret;
32439 }
32440 #else
32441 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32442  uint64x1_t __ret;
32443  __ret = (uint64x1_t)(__p0);
32444  return __ret;
32445 }
32446 #endif
32447 
32448 #ifdef __LITTLE_ENDIAN__
32449 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32450  uint64x1_t __ret;
32451  __ret = (uint64x1_t)(__p0);
32452  return __ret;
32453 }
32454 #else
32455 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32456  uint64x1_t __ret;
32457  __ret = (uint64x1_t)(__p0);
32458  return __ret;
32459 }
32460 #endif
32461 
32462 #ifdef __LITTLE_ENDIAN__
32463 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32464  uint64x1_t __ret;
32465  __ret = (uint64x1_t)(__p0);
32466  return __ret;
32467 }
32468 #else
32469 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32470  uint64x1_t __ret;
32471  __ret = (uint64x1_t)(__p0);
32472  return __ret;
32473 }
32474 #endif
32475 
32476 #ifdef __LITTLE_ENDIAN__
32477 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32478  uint64x1_t __ret;
32479  __ret = (uint64x1_t)(__p0);
32480  return __ret;
32481 }
32482 #else
32483 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32484  uint64x1_t __ret;
32485  __ret = (uint64x1_t)(__p0);
32486  return __ret;
32487 }
32488 #endif
32489 
32490 #ifdef __LITTLE_ENDIAN__
32491 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32492  uint64x1_t __ret;
32493  __ret = (uint64x1_t)(__p0);
32494  return __ret;
32495 }
32496 #else
32497 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32498  uint64x1_t __ret;
32499  __ret = (uint64x1_t)(__p0);
32500  return __ret;
32501 }
32502 #endif
32503 
32504 #ifdef __LITTLE_ENDIAN__
32505 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32506  uint64x1_t __ret;
32507  __ret = (uint64x1_t)(__p0);
32508  return __ret;
32509 }
32510 #else
32511 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32512  uint64x1_t __ret;
32513  __ret = (uint64x1_t)(__p0);
32514  return __ret;
32515 }
32516 #endif
32517 
32518 #ifdef __LITTLE_ENDIAN__
32519 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32520  uint64x1_t __ret;
32521  __ret = (uint64x1_t)(__p0);
32522  return __ret;
32523 }
32524 #else
32525 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32526  uint64x1_t __ret;
32527  __ret = (uint64x1_t)(__p0);
32528  return __ret;
32529 }
32530 #endif
32531 
32532 #ifdef __LITTLE_ENDIAN__
32533 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32534  uint64x1_t __ret;
32535  __ret = (uint64x1_t)(__p0);
32536  return __ret;
32537 }
32538 #else
32539 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32540  uint64x1_t __ret;
32541  __ret = (uint64x1_t)(__p0);
32542  return __ret;
32543 }
32544 #endif
32545 
32546 #ifdef __LITTLE_ENDIAN__
32547 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32548  uint16x4_t __ret;
32549  __ret = (uint16x4_t)(__p0);
32550  return __ret;
32551 }
32552 #else
32553 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32554  uint16x4_t __ret;
32555  __ret = (uint16x4_t)(__p0);
32556  return __ret;
32557 }
32558 #endif
32559 
32560 #ifdef __LITTLE_ENDIAN__
32561 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32562  uint16x4_t __ret;
32563  __ret = (uint16x4_t)(__p0);
32564  return __ret;
32565 }
32566 #else
32567 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32568  uint16x4_t __ret;
32569  __ret = (uint16x4_t)(__p0);
32570  return __ret;
32571 }
32572 #endif
32573 
32574 #ifdef __LITTLE_ENDIAN__
32575 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32576  uint16x4_t __ret;
32577  __ret = (uint16x4_t)(__p0);
32578  return __ret;
32579 }
32580 #else
32581 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32582  uint16x4_t __ret;
32583  __ret = (uint16x4_t)(__p0);
32584  return __ret;
32585 }
32586 #endif
32587 
32588 #ifdef __LITTLE_ENDIAN__
32589 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32590  uint16x4_t __ret;
32591  __ret = (uint16x4_t)(__p0);
32592  return __ret;
32593 }
32594 #else
32595 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32596  uint16x4_t __ret;
32597  __ret = (uint16x4_t)(__p0);
32598  return __ret;
32599 }
32600 #endif
32601 
32602 #ifdef __LITTLE_ENDIAN__
32603 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32604  uint16x4_t __ret;
32605  __ret = (uint16x4_t)(__p0);
32606  return __ret;
32607 }
32608 #else
32609 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32610  uint16x4_t __ret;
32611  __ret = (uint16x4_t)(__p0);
32612  return __ret;
32613 }
32614 #endif
32615 
32616 #ifdef __LITTLE_ENDIAN__
32617 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32618  uint16x4_t __ret;
32619  __ret = (uint16x4_t)(__p0);
32620  return __ret;
32621 }
32622 #else
32623 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32624  uint16x4_t __ret;
32625  __ret = (uint16x4_t)(__p0);
32626  return __ret;
32627 }
32628 #endif
32629 
32630 #ifdef __LITTLE_ENDIAN__
32631 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32632  uint16x4_t __ret;
32633  __ret = (uint16x4_t)(__p0);
32634  return __ret;
32635 }
32636 #else
32637 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32638  uint16x4_t __ret;
32639  __ret = (uint16x4_t)(__p0);
32640  return __ret;
32641 }
32642 #endif
32643 
32644 #ifdef __LITTLE_ENDIAN__
32645 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32646  uint16x4_t __ret;
32647  __ret = (uint16x4_t)(__p0);
32648  return __ret;
32649 }
32650 #else
32651 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32652  uint16x4_t __ret;
32653  __ret = (uint16x4_t)(__p0);
32654  return __ret;
32655 }
32656 #endif
32657 
32658 #ifdef __LITTLE_ENDIAN__
32659 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32660  uint16x4_t __ret;
32661  __ret = (uint16x4_t)(__p0);
32662  return __ret;
32663 }
32664 #else
32665 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32666  uint16x4_t __ret;
32667  __ret = (uint16x4_t)(__p0);
32668  return __ret;
32669 }
32670 #endif
32671 
32672 #ifdef __LITTLE_ENDIAN__
32673 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32674  uint16x4_t __ret;
32675  __ret = (uint16x4_t)(__p0);
32676  return __ret;
32677 }
32678 #else
32679 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32680  uint16x4_t __ret;
32681  __ret = (uint16x4_t)(__p0);
32682  return __ret;
32683 }
32684 #endif
32685 
32686 #ifdef __LITTLE_ENDIAN__
32687 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32688  uint16x4_t __ret;
32689  __ret = (uint16x4_t)(__p0);
32690  return __ret;
32691 }
32692 #else
32693 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32694  uint16x4_t __ret;
32695  __ret = (uint16x4_t)(__p0);
32696  return __ret;
32697 }
32698 #endif
32699 
32700 #ifdef __LITTLE_ENDIAN__
32701 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32702  int8x8_t __ret;
32703  __ret = (int8x8_t)(__p0);
32704  return __ret;
32705 }
32706 #else
32707 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32708  int8x8_t __ret;
32709  __ret = (int8x8_t)(__p0);
32710  return __ret;
32711 }
32712 #endif
32713 
32714 #ifdef __LITTLE_ENDIAN__
32715 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32716  int8x8_t __ret;
32717  __ret = (int8x8_t)(__p0);
32718  return __ret;
32719 }
32720 #else
32721 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32722  int8x8_t __ret;
32723  __ret = (int8x8_t)(__p0);
32724  return __ret;
32725 }
32726 #endif
32727 
32728 #ifdef __LITTLE_ENDIAN__
32729 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32730  int8x8_t __ret;
32731  __ret = (int8x8_t)(__p0);
32732  return __ret;
32733 }
32734 #else
32735 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32736  int8x8_t __ret;
32737  __ret = (int8x8_t)(__p0);
32738  return __ret;
32739 }
32740 #endif
32741 
32742 #ifdef __LITTLE_ENDIAN__
32743 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32744  int8x8_t __ret;
32745  __ret = (int8x8_t)(__p0);
32746  return __ret;
32747 }
32748 #else
32749 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32750  int8x8_t __ret;
32751  __ret = (int8x8_t)(__p0);
32752  return __ret;
32753 }
32754 #endif
32755 
32756 #ifdef __LITTLE_ENDIAN__
32757 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32758  int8x8_t __ret;
32759  __ret = (int8x8_t)(__p0);
32760  return __ret;
32761 }
32762 #else
32763 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32764  int8x8_t __ret;
32765  __ret = (int8x8_t)(__p0);
32766  return __ret;
32767 }
32768 #endif
32769 
32770 #ifdef __LITTLE_ENDIAN__
32771 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32772  int8x8_t __ret;
32773  __ret = (int8x8_t)(__p0);
32774  return __ret;
32775 }
32776 #else
32777 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32778  int8x8_t __ret;
32779  __ret = (int8x8_t)(__p0);
32780  return __ret;
32781 }
32782 #endif
32783 
32784 #ifdef __LITTLE_ENDIAN__
32785 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32786  int8x8_t __ret;
32787  __ret = (int8x8_t)(__p0);
32788  return __ret;
32789 }
32790 #else
32791 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32792  int8x8_t __ret;
32793  __ret = (int8x8_t)(__p0);
32794  return __ret;
32795 }
32796 #endif
32797 
32798 #ifdef __LITTLE_ENDIAN__
32799 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32800  int8x8_t __ret;
32801  __ret = (int8x8_t)(__p0);
32802  return __ret;
32803 }
32804 #else
32805 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32806  int8x8_t __ret;
32807  __ret = (int8x8_t)(__p0);
32808  return __ret;
32809 }
32810 #endif
32811 
32812 #ifdef __LITTLE_ENDIAN__
32813 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32814  int8x8_t __ret;
32815  __ret = (int8x8_t)(__p0);
32816  return __ret;
32817 }
32818 #else
32819 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32820  int8x8_t __ret;
32821  __ret = (int8x8_t)(__p0);
32822  return __ret;
32823 }
32824 #endif
32825 
32826 #ifdef __LITTLE_ENDIAN__
32827 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32828  int8x8_t __ret;
32829  __ret = (int8x8_t)(__p0);
32830  return __ret;
32831 }
32832 #else
32833 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32834  int8x8_t __ret;
32835  __ret = (int8x8_t)(__p0);
32836  return __ret;
32837 }
32838 #endif
32839 
32840 #ifdef __LITTLE_ENDIAN__
32841 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32842  int8x8_t __ret;
32843  __ret = (int8x8_t)(__p0);
32844  return __ret;
32845 }
32846 #else
32847 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32848  int8x8_t __ret;
32849  __ret = (int8x8_t)(__p0);
32850  return __ret;
32851 }
32852 #endif
32853 
32854 #ifdef __LITTLE_ENDIAN__
32855 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32856  float32x2_t __ret;
32857  __ret = (float32x2_t)(__p0);
32858  return __ret;
32859 }
32860 #else
32861 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32862  float32x2_t __ret;
32863  __ret = (float32x2_t)(__p0);
32864  return __ret;
32865 }
32866 #endif
32867 
32868 #ifdef __LITTLE_ENDIAN__
32869 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32870  float32x2_t __ret;
32871  __ret = (float32x2_t)(__p0);
32872  return __ret;
32873 }
32874 #else
32875 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32876  float32x2_t __ret;
32877  __ret = (float32x2_t)(__p0);
32878  return __ret;
32879 }
32880 #endif
32881 
32882 #ifdef __LITTLE_ENDIAN__
32883 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32884  float32x2_t __ret;
32885  __ret = (float32x2_t)(__p0);
32886  return __ret;
32887 }
32888 #else
32889 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32890  float32x2_t __ret;
32891  __ret = (float32x2_t)(__p0);
32892  return __ret;
32893 }
32894 #endif
32895 
32896 #ifdef __LITTLE_ENDIAN__
32897 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32898  float32x2_t __ret;
32899  __ret = (float32x2_t)(__p0);
32900  return __ret;
32901 }
32902 #else
32903 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32904  float32x2_t __ret;
32905  __ret = (float32x2_t)(__p0);
32906  return __ret;
32907 }
32908 #endif
32909 
32910 #ifdef __LITTLE_ENDIAN__
32911 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32912  float32x2_t __ret;
32913  __ret = (float32x2_t)(__p0);
32914  return __ret;
32915 }
32916 #else
32917 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32918  float32x2_t __ret;
32919  __ret = (float32x2_t)(__p0);
32920  return __ret;
32921 }
32922 #endif
32923 
32924 #ifdef __LITTLE_ENDIAN__
32925 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32926  float32x2_t __ret;
32927  __ret = (float32x2_t)(__p0);
32928  return __ret;
32929 }
32930 #else
32931 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32932  float32x2_t __ret;
32933  __ret = (float32x2_t)(__p0);
32934  return __ret;
32935 }
32936 #endif
32937 
32938 #ifdef __LITTLE_ENDIAN__
32939 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32940  float32x2_t __ret;
32941  __ret = (float32x2_t)(__p0);
32942  return __ret;
32943 }
32944 #else
32945 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32946  float32x2_t __ret;
32947  __ret = (float32x2_t)(__p0);
32948  return __ret;
32949 }
32950 #endif
32951 
32952 #ifdef __LITTLE_ENDIAN__
32953 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32954  float32x2_t __ret;
32955  __ret = (float32x2_t)(__p0);
32956  return __ret;
32957 }
32958 #else
32959 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32960  float32x2_t __ret;
32961  __ret = (float32x2_t)(__p0);
32962  return __ret;
32963 }
32964 #endif
32965 
32966 #ifdef __LITTLE_ENDIAN__
32967 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32968  float32x2_t __ret;
32969  __ret = (float32x2_t)(__p0);
32970  return __ret;
32971 }
32972 #else
32973 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32974  float32x2_t __ret;
32975  __ret = (float32x2_t)(__p0);
32976  return __ret;
32977 }
32978 #endif
32979 
32980 #ifdef __LITTLE_ENDIAN__
32981 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32982  float32x2_t __ret;
32983  __ret = (float32x2_t)(__p0);
32984  return __ret;
32985 }
32986 #else
32987 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32988  float32x2_t __ret;
32989  __ret = (float32x2_t)(__p0);
32990  return __ret;
32991 }
32992 #endif
32993 
32994 #ifdef __LITTLE_ENDIAN__
32995 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32996  float32x2_t __ret;
32997  __ret = (float32x2_t)(__p0);
32998  return __ret;
32999 }
33000 #else
33001 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
33002  float32x2_t __ret;
33003  __ret = (float32x2_t)(__p0);
33004  return __ret;
33005 }
33006 #endif
33007 
33008 #ifdef __LITTLE_ENDIAN__
33009 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
33010  float16x4_t __ret;
33011  __ret = (float16x4_t)(__p0);
33012  return __ret;
33013 }
33014 #else
33015 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
33016  float16x4_t __ret;
33017  __ret = (float16x4_t)(__p0);
33018  return __ret;
33019 }
33020 #endif
33021 
33022 #ifdef __LITTLE_ENDIAN__
33023 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
33024  float16x4_t __ret;
33025  __ret = (float16x4_t)(__p0);
33026  return __ret;
33027 }
33028 #else
33029 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
33030  float16x4_t __ret;
33031  __ret = (float16x4_t)(__p0);
33032  return __ret;
33033 }
33034 #endif
33035 
33036 #ifdef __LITTLE_ENDIAN__
33037 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33038  float16x4_t __ret;
33039  __ret = (float16x4_t)(__p0);
33040  return __ret;
33041 }
33042 #else
33043 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33044  float16x4_t __ret;
33045  __ret = (float16x4_t)(__p0);
33046  return __ret;
33047 }
33048 #endif
33049 
33050 #ifdef __LITTLE_ENDIAN__
33051 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33052  float16x4_t __ret;
33053  __ret = (float16x4_t)(__p0);
33054  return __ret;
33055 }
33056 #else
33057 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33058  float16x4_t __ret;
33059  __ret = (float16x4_t)(__p0);
33060  return __ret;
33061 }
33062 #endif
33063 
33064 #ifdef __LITTLE_ENDIAN__
33065 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33066  float16x4_t __ret;
33067  __ret = (float16x4_t)(__p0);
33068  return __ret;
33069 }
33070 #else
33071 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33072  float16x4_t __ret;
33073  __ret = (float16x4_t)(__p0);
33074  return __ret;
33075 }
33076 #endif
33077 
33078 #ifdef __LITTLE_ENDIAN__
33079 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33080  float16x4_t __ret;
33081  __ret = (float16x4_t)(__p0);
33082  return __ret;
33083 }
33084 #else
33085 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33086  float16x4_t __ret;
33087  __ret = (float16x4_t)(__p0);
33088  return __ret;
33089 }
33090 #endif
33091 
33092 #ifdef __LITTLE_ENDIAN__
33093 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33094  float16x4_t __ret;
33095  __ret = (float16x4_t)(__p0);
33096  return __ret;
33097 }
33098 #else
33099 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33100  float16x4_t __ret;
33101  __ret = (float16x4_t)(__p0);
33102  return __ret;
33103 }
33104 #endif
33105 
33106 #ifdef __LITTLE_ENDIAN__
33107 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33108  float16x4_t __ret;
33109  __ret = (float16x4_t)(__p0);
33110  return __ret;
33111 }
33112 #else
33113 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33114  float16x4_t __ret;
33115  __ret = (float16x4_t)(__p0);
33116  return __ret;
33117 }
33118 #endif
33119 
33120 #ifdef __LITTLE_ENDIAN__
33121 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33122  float16x4_t __ret;
33123  __ret = (float16x4_t)(__p0);
33124  return __ret;
33125 }
33126 #else
33127 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33128  float16x4_t __ret;
33129  __ret = (float16x4_t)(__p0);
33130  return __ret;
33131 }
33132 #endif
33133 
33134 #ifdef __LITTLE_ENDIAN__
33135 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33136  float16x4_t __ret;
33137  __ret = (float16x4_t)(__p0);
33138  return __ret;
33139 }
33140 #else
33141 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33142  float16x4_t __ret;
33143  __ret = (float16x4_t)(__p0);
33144  return __ret;
33145 }
33146 #endif
33147 
33148 #ifdef __LITTLE_ENDIAN__
33149 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33150  float16x4_t __ret;
33151  __ret = (float16x4_t)(__p0);
33152  return __ret;
33153 }
33154 #else
33155 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33156  float16x4_t __ret;
33157  __ret = (float16x4_t)(__p0);
33158  return __ret;
33159 }
33160 #endif
33161 
33162 #ifdef __LITTLE_ENDIAN__
33163 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33164  int32x2_t __ret;
33165  __ret = (int32x2_t)(__p0);
33166  return __ret;
33167 }
33168 #else
33169 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33170  int32x2_t __ret;
33171  __ret = (int32x2_t)(__p0);
33172  return __ret;
33173 }
33174 #endif
33175 
33176 #ifdef __LITTLE_ENDIAN__
33177 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33178  int32x2_t __ret;
33179  __ret = (int32x2_t)(__p0);
33180  return __ret;
33181 }
33182 #else
33183 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33184  int32x2_t __ret;
33185  __ret = (int32x2_t)(__p0);
33186  return __ret;
33187 }
33188 #endif
33189 
33190 #ifdef __LITTLE_ENDIAN__
33191 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33192  int32x2_t __ret;
33193  __ret = (int32x2_t)(__p0);
33194  return __ret;
33195 }
33196 #else
33197 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33198  int32x2_t __ret;
33199  __ret = (int32x2_t)(__p0);
33200  return __ret;
33201 }
33202 #endif
33203 
33204 #ifdef __LITTLE_ENDIAN__
33205 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33206  int32x2_t __ret;
33207  __ret = (int32x2_t)(__p0);
33208  return __ret;
33209 }
33210 #else
33211 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33212  int32x2_t __ret;
33213  __ret = (int32x2_t)(__p0);
33214  return __ret;
33215 }
33216 #endif
33217 
33218 #ifdef __LITTLE_ENDIAN__
33219 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33220  int32x2_t __ret;
33221  __ret = (int32x2_t)(__p0);
33222  return __ret;
33223 }
33224 #else
33225 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33226  int32x2_t __ret;
33227  __ret = (int32x2_t)(__p0);
33228  return __ret;
33229 }
33230 #endif
33231 
33232 #ifdef __LITTLE_ENDIAN__
33233 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33234  int32x2_t __ret;
33235  __ret = (int32x2_t)(__p0);
33236  return __ret;
33237 }
33238 #else
33239 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33240  int32x2_t __ret;
33241  __ret = (int32x2_t)(__p0);
33242  return __ret;
33243 }
33244 #endif
33245 
33246 #ifdef __LITTLE_ENDIAN__
33247 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33248  int32x2_t __ret;
33249  __ret = (int32x2_t)(__p0);
33250  return __ret;
33251 }
33252 #else
33253 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33254  int32x2_t __ret;
33255  __ret = (int32x2_t)(__p0);
33256  return __ret;
33257 }
33258 #endif
33259 
33260 #ifdef __LITTLE_ENDIAN__
33261 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33262  int32x2_t __ret;
33263  __ret = (int32x2_t)(__p0);
33264  return __ret;
33265 }
33266 #else
33267 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33268  int32x2_t __ret;
33269  __ret = (int32x2_t)(__p0);
33270  return __ret;
33271 }
33272 #endif
33273 
33274 #ifdef __LITTLE_ENDIAN__
33275 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33276  int32x2_t __ret;
33277  __ret = (int32x2_t)(__p0);
33278  return __ret;
33279 }
33280 #else
33281 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33282  int32x2_t __ret;
33283  __ret = (int32x2_t)(__p0);
33284  return __ret;
33285 }
33286 #endif
33287 
33288 #ifdef __LITTLE_ENDIAN__
33289 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33290  int32x2_t __ret;
33291  __ret = (int32x2_t)(__p0);
33292  return __ret;
33293 }
33294 #else
33295 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33296  int32x2_t __ret;
33297  __ret = (int32x2_t)(__p0);
33298  return __ret;
33299 }
33300 #endif
33301 
33302 #ifdef __LITTLE_ENDIAN__
33303 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33304  int32x2_t __ret;
33305  __ret = (int32x2_t)(__p0);
33306  return __ret;
33307 }
33308 #else
33309 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33310  int32x2_t __ret;
33311  __ret = (int32x2_t)(__p0);
33312  return __ret;
33313 }
33314 #endif
33315 
33316 #ifdef __LITTLE_ENDIAN__
33317 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33318  int64x1_t __ret;
33319  __ret = (int64x1_t)(__p0);
33320  return __ret;
33321 }
33322 #else
33323 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33324  int64x1_t __ret;
33325  __ret = (int64x1_t)(__p0);
33326  return __ret;
33327 }
33328 #endif
33329 
33330 #ifdef __LITTLE_ENDIAN__
33331 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33332  int64x1_t __ret;
33333  __ret = (int64x1_t)(__p0);
33334  return __ret;
33335 }
33336 #else
33337 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33338  int64x1_t __ret;
33339  __ret = (int64x1_t)(__p0);
33340  return __ret;
33341 }
33342 #endif
33343 
33344 #ifdef __LITTLE_ENDIAN__
33345 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33346  int64x1_t __ret;
33347  __ret = (int64x1_t)(__p0);
33348  return __ret;
33349 }
33350 #else
33351 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33352  int64x1_t __ret;
33353  __ret = (int64x1_t)(__p0);
33354  return __ret;
33355 }
33356 #endif
33357 
33358 #ifdef __LITTLE_ENDIAN__
33359 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33360  int64x1_t __ret;
33361  __ret = (int64x1_t)(__p0);
33362  return __ret;
33363 }
33364 #else
33365 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33366  int64x1_t __ret;
33367  __ret = (int64x1_t)(__p0);
33368  return __ret;
33369 }
33370 #endif
33371 
33372 #ifdef __LITTLE_ENDIAN__
33373 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33374  int64x1_t __ret;
33375  __ret = (int64x1_t)(__p0);
33376  return __ret;
33377 }
33378 #else
33379 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33380  int64x1_t __ret;
33381  __ret = (int64x1_t)(__p0);
33382  return __ret;
33383 }
33384 #endif
33385 
33386 #ifdef __LITTLE_ENDIAN__
33387 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33388  int64x1_t __ret;
33389  __ret = (int64x1_t)(__p0);
33390  return __ret;
33391 }
33392 #else
33393 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33394  int64x1_t __ret;
33395  __ret = (int64x1_t)(__p0);
33396  return __ret;
33397 }
33398 #endif
33399 
33400 #ifdef __LITTLE_ENDIAN__
33401 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33402  int64x1_t __ret;
33403  __ret = (int64x1_t)(__p0);
33404  return __ret;
33405 }
33406 #else
33407 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33408  int64x1_t __ret;
33409  __ret = (int64x1_t)(__p0);
33410  return __ret;
33411 }
33412 #endif
33413 
33414 #ifdef __LITTLE_ENDIAN__
33415 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33416  int64x1_t __ret;
33417  __ret = (int64x1_t)(__p0);
33418  return __ret;
33419 }
33420 #else
33421 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33422  int64x1_t __ret;
33423  __ret = (int64x1_t)(__p0);
33424  return __ret;
33425 }
33426 #endif
33427 
33428 #ifdef __LITTLE_ENDIAN__
33429 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33430  int64x1_t __ret;
33431  __ret = (int64x1_t)(__p0);
33432  return __ret;
33433 }
33434 #else
33435 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33436  int64x1_t __ret;
33437  __ret = (int64x1_t)(__p0);
33438  return __ret;
33439 }
33440 #endif
33441 
33442 #ifdef __LITTLE_ENDIAN__
33443 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33444  int64x1_t __ret;
33445  __ret = (int64x1_t)(__p0);
33446  return __ret;
33447 }
33448 #else
33449 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33450  int64x1_t __ret;
33451  __ret = (int64x1_t)(__p0);
33452  return __ret;
33453 }
33454 #endif
33455 
33456 #ifdef __LITTLE_ENDIAN__
33457 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33458  int64x1_t __ret;
33459  __ret = (int64x1_t)(__p0);
33460  return __ret;
33461 }
33462 #else
33463 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33464  int64x1_t __ret;
33465  __ret = (int64x1_t)(__p0);
33466  return __ret;
33467 }
33468 #endif
33469 
33470 #ifdef __LITTLE_ENDIAN__
33471 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33472  int16x4_t __ret;
33473  __ret = (int16x4_t)(__p0);
33474  return __ret;
33475 }
33476 #else
33477 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33478  int16x4_t __ret;
33479  __ret = (int16x4_t)(__p0);
33480  return __ret;
33481 }
33482 #endif
33483 
33484 #ifdef __LITTLE_ENDIAN__
33485 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33486  int16x4_t __ret;
33487  __ret = (int16x4_t)(__p0);
33488  return __ret;
33489 }
33490 #else
33491 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33492  int16x4_t __ret;
33493  __ret = (int16x4_t)(__p0);
33494  return __ret;
33495 }
33496 #endif
33497 
33498 #ifdef __LITTLE_ENDIAN__
33499 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33500  int16x4_t __ret;
33501  __ret = (int16x4_t)(__p0);
33502  return __ret;
33503 }
33504 #else
33505 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33506  int16x4_t __ret;
33507  __ret = (int16x4_t)(__p0);
33508  return __ret;
33509 }
33510 #endif
33511 
33512 #ifdef __LITTLE_ENDIAN__
33513 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33514  int16x4_t __ret;
33515  __ret = (int16x4_t)(__p0);
33516  return __ret;
33517 }
33518 #else
33519 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33520  int16x4_t __ret;
33521  __ret = (int16x4_t)(__p0);
33522  return __ret;
33523 }
33524 #endif
33525 
33526 #ifdef __LITTLE_ENDIAN__
33527 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33528  int16x4_t __ret;
33529  __ret = (int16x4_t)(__p0);
33530  return __ret;
33531 }
33532 #else
33533 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33534  int16x4_t __ret;
33535  __ret = (int16x4_t)(__p0);
33536  return __ret;
33537 }
33538 #endif
33539 
33540 #ifdef __LITTLE_ENDIAN__
33541 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33542  int16x4_t __ret;
33543  __ret = (int16x4_t)(__p0);
33544  return __ret;
33545 }
33546 #else
33547 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33548  int16x4_t __ret;
33549  __ret = (int16x4_t)(__p0);
33550  return __ret;
33551 }
33552 #endif
33553 
33554 #ifdef __LITTLE_ENDIAN__
33555 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33556  int16x4_t __ret;
33557  __ret = (int16x4_t)(__p0);
33558  return __ret;
33559 }
33560 #else
33561 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33562  int16x4_t __ret;
33563  __ret = (int16x4_t)(__p0);
33564  return __ret;
33565 }
33566 #endif
33567 
33568 #ifdef __LITTLE_ENDIAN__
33569 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33570  int16x4_t __ret;
33571  __ret = (int16x4_t)(__p0);
33572  return __ret;
33573 }
33574 #else
33575 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33576  int16x4_t __ret;
33577  __ret = (int16x4_t)(__p0);
33578  return __ret;
33579 }
33580 #endif
33581 
33582 #ifdef __LITTLE_ENDIAN__
33583 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33584  int16x4_t __ret;
33585  __ret = (int16x4_t)(__p0);
33586  return __ret;
33587 }
33588 #else
33589 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33590  int16x4_t __ret;
33591  __ret = (int16x4_t)(__p0);
33592  return __ret;
33593 }
33594 #endif
33595 
33596 #ifdef __LITTLE_ENDIAN__
33597 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33598  int16x4_t __ret;
33599  __ret = (int16x4_t)(__p0);
33600  return __ret;
33601 }
33602 #else
33603 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33604  int16x4_t __ret;
33605  __ret = (int16x4_t)(__p0);
33606  return __ret;
33607 }
33608 #endif
33609 
33610 #ifdef __LITTLE_ENDIAN__
33611 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33612  int16x4_t __ret;
33613  __ret = (int16x4_t)(__p0);
33614  return __ret;
33615 }
33616 #else
33617 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33618  int16x4_t __ret;
33619  __ret = (int16x4_t)(__p0);
33620  return __ret;
33621 }
33622 #endif
33623 
33624 #endif
33625 #if __ARM_ARCH >= 8
33626 #ifdef __LITTLE_ENDIAN__
33627 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33628  int32x4_t __ret;
33629  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
33630  return __ret;
33631 }
33632 #else
33633 __ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33634  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33635  int32x4_t __ret;
33636  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
33637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33638  return __ret;
33639 }
33640 #endif
33641 
33642 #ifdef __LITTLE_ENDIAN__
33643 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33644  int32x2_t __ret;
33645  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
33646  return __ret;
33647 }
33648 #else
33649 __ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33650  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33651  int32x2_t __ret;
33652  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
33653  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33654  return __ret;
33655 }
33656 #endif
33657 
33658 #ifdef __LITTLE_ENDIAN__
33659 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33660  uint32x4_t __ret;
33661  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
33662  return __ret;
33663 }
33664 #else
33665 __ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33666  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33667  uint32x4_t __ret;
33668  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
33669  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33670  return __ret;
33671 }
33672 #endif
33673 
33674 #ifdef __LITTLE_ENDIAN__
33675 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33676  uint32x2_t __ret;
33677  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
33678  return __ret;
33679 }
33680 #else
33681 __ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33682  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33683  uint32x2_t __ret;
33684  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
33685  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33686  return __ret;
33687 }
33688 #endif
33689 
33690 #ifdef __LITTLE_ENDIAN__
33691 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33692  int32x4_t __ret;
33693  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
33694  return __ret;
33695 }
33696 #else
33697 __ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33698  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33699  int32x4_t __ret;
33700  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
33701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33702  return __ret;
33703 }
33704 #endif
33705 
33706 #ifdef __LITTLE_ENDIAN__
33707 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33708  int32x2_t __ret;
33709  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
33710  return __ret;
33711 }
33712 #else
33713 __ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33714  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33715  int32x2_t __ret;
33716  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
33717  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33718  return __ret;
33719 }
33720 #endif
33721 
33722 #ifdef __LITTLE_ENDIAN__
33723 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33724  uint32x4_t __ret;
33725  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
33726  return __ret;
33727 }
33728 #else
33729 __ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33730  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33731  uint32x4_t __ret;
33732  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
33733  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33734  return __ret;
33735 }
33736 #endif
33737 
33738 #ifdef __LITTLE_ENDIAN__
33739 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33740  uint32x2_t __ret;
33741  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
33742  return __ret;
33743 }
33744 #else
33745 __ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33746  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33747  uint32x2_t __ret;
33748  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
33749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33750  return __ret;
33751 }
33752 #endif
33753 
33754 #ifdef __LITTLE_ENDIAN__
33755 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33756  int32x4_t __ret;
33757  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
33758  return __ret;
33759 }
33760 #else
33761 __ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33762  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33763  int32x4_t __ret;
33764  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
33765  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33766  return __ret;
33767 }
33768 #endif
33769 
33770 #ifdef __LITTLE_ENDIAN__
33771 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33772  int32x2_t __ret;
33773  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
33774  return __ret;
33775 }
33776 #else
33777 __ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33778  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33779  int32x2_t __ret;
33780  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
33781  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33782  return __ret;
33783 }
33784 #endif
33785 
33786 #ifdef __LITTLE_ENDIAN__
33787 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33788  uint32x4_t __ret;
33789  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
33790  return __ret;
33791 }
33792 #else
33793 __ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33794  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33795  uint32x4_t __ret;
33796  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
33797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33798  return __ret;
33799 }
33800 #endif
33801 
33802 #ifdef __LITTLE_ENDIAN__
33803 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33804  uint32x2_t __ret;
33805  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
33806  return __ret;
33807 }
33808 #else
33809 __ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33810  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33811  uint32x2_t __ret;
33812  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
33813  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33814  return __ret;
33815 }
33816 #endif
33817 
33818 #ifdef __LITTLE_ENDIAN__
33819 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33820  int32x4_t __ret;
33821  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
33822  return __ret;
33823 }
33824 #else
33825 __ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33826  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33827  int32x4_t __ret;
33828  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
33829  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33830  return __ret;
33831 }
33832 #endif
33833 
33834 #ifdef __LITTLE_ENDIAN__
33835 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33836  int32x2_t __ret;
33837  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
33838  return __ret;
33839 }
33840 #else
33841 __ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33842  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33843  int32x2_t __ret;
33844  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
33845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33846  return __ret;
33847 }
33848 #endif
33849 
33850 #ifdef __LITTLE_ENDIAN__
33851 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33852  uint32x4_t __ret;
33853  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
33854  return __ret;
33855 }
33856 #else
33857 __ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33858  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33859  uint32x4_t __ret;
33860  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
33861  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33862  return __ret;
33863 }
33864 #endif
33865 
33866 #ifdef __LITTLE_ENDIAN__
33867 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33868  uint32x2_t __ret;
33869  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
33870  return __ret;
33871 }
33872 #else
33873 __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33874  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33875  uint32x2_t __ret;
33876  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
33877  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33878  return __ret;
33879 }
33880 #endif
33881 
33882 #endif
33883 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
33884 #ifdef __LITTLE_ENDIAN__
33885 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
33886  float32x4_t __ret;
33887  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
33888  return __ret;
33889 }
33890 #else
33891 __ai float32x4_t vrndq_f32(float32x4_t __p0) {
33892  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33893  float32x4_t __ret;
33894  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
33895  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33896  return __ret;
33897 }
33898 #endif
33899 
33900 #ifdef __LITTLE_ENDIAN__
33901 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
33902  float32x2_t __ret;
33903  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
33904  return __ret;
33905 }
33906 #else
33907 __ai float32x2_t vrnd_f32(float32x2_t __p0) {
33908  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33909  float32x2_t __ret;
33910  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
33911  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33912  return __ret;
33913 }
33914 #endif
33915 
33916 #ifdef __LITTLE_ENDIAN__
33917 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33918  float32x4_t __ret;
33919  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
33920  return __ret;
33921 }
33922 #else
33923 __ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33924  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33925  float32x4_t __ret;
33926  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
33927  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33928  return __ret;
33929 }
33930 #endif
33931 
33932 #ifdef __LITTLE_ENDIAN__
33933 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
33934  float32x2_t __ret;
33935  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
33936  return __ret;
33937 }
33938 #else
33939 __ai float32x2_t vrnda_f32(float32x2_t __p0) {
33940  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33941  float32x2_t __ret;
33942  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
33943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33944  return __ret;
33945 }
33946 #endif
33947 
33948 #ifdef __LITTLE_ENDIAN__
33949 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33950  float32x4_t __ret;
33951  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
33952  return __ret;
33953 }
33954 #else
33955 __ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33956  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33957  float32x4_t __ret;
33958  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
33959  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33960  return __ret;
33961 }
33962 #endif
33963 
33964 #ifdef __LITTLE_ENDIAN__
33965 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
33966  float32x2_t __ret;
33967  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
33968  return __ret;
33969 }
33970 #else
33971 __ai float32x2_t vrndm_f32(float32x2_t __p0) {
33972  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33973  float32x2_t __ret;
33974  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
33975  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33976  return __ret;
33977 }
33978 #endif
33979 
33980 #ifdef __LITTLE_ENDIAN__
33981 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33982  float32x4_t __ret;
33983  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
33984  return __ret;
33985 }
33986 #else
33987 __ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33988  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33989  float32x4_t __ret;
33990  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
33991  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33992  return __ret;
33993 }
33994 #endif
33995 
33996 #ifdef __LITTLE_ENDIAN__
33997 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
33998  float32x2_t __ret;
33999  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
34000  return __ret;
34001 }
34002 #else
34003 __ai float32x2_t vrndn_f32(float32x2_t __p0) {
34004  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34005  float32x2_t __ret;
34006  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
34007  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34008  return __ret;
34009 }
34010 #endif
34011 
34012 #ifdef __LITTLE_ENDIAN__
34013 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
34014  float32x4_t __ret;
34015  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
34016  return __ret;
34017 }
34018 #else
34019 __ai float32x4_t vrndpq_f32(float32x4_t __p0) {
34020  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34021  float32x4_t __ret;
34022  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
34023  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34024  return __ret;
34025 }
34026 #endif
34027 
34028 #ifdef __LITTLE_ENDIAN__
34029 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
34030  float32x2_t __ret;
34031  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
34032  return __ret;
34033 }
34034 #else
34035 __ai float32x2_t vrndp_f32(float32x2_t __p0) {
34036  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34037  float32x2_t __ret;
34038  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
34039  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34040  return __ret;
34041 }
34042 #endif
34043 
34044 #ifdef __LITTLE_ENDIAN__
34045 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34046  float32x4_t __ret;
34047  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
34048  return __ret;
34049 }
34050 #else
34051 __ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34052  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34053  float32x4_t __ret;
34054  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
34055  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34056  return __ret;
34057 }
34058 #endif
34059 
34060 #ifdef __LITTLE_ENDIAN__
34061 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
34062  float32x2_t __ret;
34063  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
34064  return __ret;
34065 }
34066 #else
34067 __ai float32x2_t vrndx_f32(float32x2_t __p0) {
34068  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34069  float32x2_t __ret;
34070  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
34071  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34072  return __ret;
34073 }
34074 #endif
34075 
34076 #endif
34077 #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
34078 #ifdef __LITTLE_ENDIAN__
34079 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34080  float32x4_t __ret;
34081  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34082  return __ret;
34083 }
34084 #else
34085 __ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34086  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34087  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34088  float32x4_t __ret;
34089  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34090  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34091  return __ret;
34092 }
34093 #endif
34094 
34095 #ifdef __LITTLE_ENDIAN__
34096 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34097  float32x2_t __ret;
34098  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34099  return __ret;
34100 }
34101 #else
34102 __ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34103  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34104  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34105  float32x2_t __ret;
34106  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34108  return __ret;
34109 }
34110 #endif
34111 
34112 #ifdef __LITTLE_ENDIAN__
34113 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34114  float32x4_t __ret;
34115  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34116  return __ret;
34117 }
34118 #else
34119 __ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34120  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34121  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34122  float32x4_t __ret;
34123  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34124  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34125  return __ret;
34126 }
34127 #endif
34128 
34129 #ifdef __LITTLE_ENDIAN__
34130 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34131  float32x2_t __ret;
34132  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34133  return __ret;
34134 }
34135 #else
34136 __ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34137  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34138  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34139  float32x2_t __ret;
34140  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34141  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34142  return __ret;
34143 }
34144 #endif
34145 
34146 #endif
34147 #if __ARM_ARCH >= 8 && defined(__aarch64__)
34148 #ifdef __LITTLE_ENDIAN__
34149 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34150  int64x2_t __ret;
34151  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
34152  return __ret;
34153 }
34154 #else
34155 __ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34156  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34157  int64x2_t __ret;
34158  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
34159  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34160  return __ret;
34161 }
34162 #endif
34163 
34164 #ifdef __LITTLE_ENDIAN__
34165 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34166  int64x1_t __ret;
34167  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34168  return __ret;
34169 }
34170 #else
34171 __ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34172  int64x1_t __ret;
34173  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34174  return __ret;
34175 }
34176 #endif
34177 
34178 #ifdef __LITTLE_ENDIAN__
34179 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34180  uint64x2_t __ret;
34181  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
34182  return __ret;
34183 }
34184 #else
34185 __ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34186  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34187  uint64x2_t __ret;
34188  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
34189  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34190  return __ret;
34191 }
34192 #endif
34193 
34194 #ifdef __LITTLE_ENDIAN__
34195 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34196  uint64x1_t __ret;
34197  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34198  return __ret;
34199 }
34200 #else
34201 __ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34202  uint64x1_t __ret;
34203  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34204  return __ret;
34205 }
34206 #endif
34207 
34208 #ifdef __LITTLE_ENDIAN__
34209 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34210  int64x2_t __ret;
34211  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
34212  return __ret;
34213 }
34214 #else
34215 __ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34216  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34217  int64x2_t __ret;
34218  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
34219  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34220  return __ret;
34221 }
34222 #endif
34223 
34224 #ifdef __LITTLE_ENDIAN__
34225 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34226  int64x1_t __ret;
34227  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34228  return __ret;
34229 }
34230 #else
34231 __ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34232  int64x1_t __ret;
34233  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34234  return __ret;
34235 }
34236 #endif
34237 
34238 #ifdef __LITTLE_ENDIAN__
34239 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34240  uint64x2_t __ret;
34241  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
34242  return __ret;
34243 }
34244 #else
34245 __ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34246  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34247  uint64x2_t __ret;
34248  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
34249  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34250  return __ret;
34251 }
34252 #endif
34253 
34254 #ifdef __LITTLE_ENDIAN__
34255 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34256  uint64x1_t __ret;
34257  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34258  return __ret;
34259 }
34260 #else
34261 __ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34262  uint64x1_t __ret;
34263  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34264  return __ret;
34265 }
34266 #endif
34267 
34268 #ifdef __LITTLE_ENDIAN__
34269 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34270  int64x2_t __ret;
34271  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
34272  return __ret;
34273 }
34274 #else
34275 __ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34276  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34277  int64x2_t __ret;
34278  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
34279  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34280  return __ret;
34281 }
34282 #endif
34283 
34284 #ifdef __LITTLE_ENDIAN__
34285 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34286  int64x1_t __ret;
34287  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34288  return __ret;
34289 }
34290 #else
34291 __ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34292  int64x1_t __ret;
34293  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34294  return __ret;
34295 }
34296 #endif
34297 
34298 #ifdef __LITTLE_ENDIAN__
34299 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34300  uint64x2_t __ret;
34301  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
34302  return __ret;
34303 }
34304 #else
34305 __ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34306  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34307  uint64x2_t __ret;
34308  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
34309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34310  return __ret;
34311 }
34312 #endif
34313 
34314 #ifdef __LITTLE_ENDIAN__
34315 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34316  uint64x1_t __ret;
34317  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34318  return __ret;
34319 }
34320 #else
34321 __ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34322  uint64x1_t __ret;
34323  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34324  return __ret;
34325 }
34326 #endif
34327 
34328 #ifdef __LITTLE_ENDIAN__
34329 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34330  int64x2_t __ret;
34331  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
34332  return __ret;
34333 }
34334 #else
34335 __ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34336  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34337  int64x2_t __ret;
34338  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
34339  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34340  return __ret;
34341 }
34342 #endif
34343 
34344 #ifdef __LITTLE_ENDIAN__
34345 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34346  int64x1_t __ret;
34347  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34348  return __ret;
34349 }
34350 #else
34351 __ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34352  int64x1_t __ret;
34353  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34354  return __ret;
34355 }
34356 #endif
34357 
34358 #ifdef __LITTLE_ENDIAN__
34359 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34360  uint64x2_t __ret;
34361  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
34362  return __ret;
34363 }
34364 #else
34365 __ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34366  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34367  uint64x2_t __ret;
34368  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
34369  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34370  return __ret;
34371 }
34372 #endif
34373 
34374 #ifdef __LITTLE_ENDIAN__
34375 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34376  uint64x1_t __ret;
34377  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34378  return __ret;
34379 }
34380 #else
34381 __ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34382  uint64x1_t __ret;
34383  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34384  return __ret;
34385 }
34386 #endif
34387 
34388 #ifdef __LITTLE_ENDIAN__
34389 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34390  poly8x8_t __ret;
34391  __ret = (poly8x8_t)(__p0);
34392  return __ret;
34393 }
34394 #else
34395 __ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34396  poly8x8_t __ret;
34397  __ret = (poly8x8_t)(__p0);
34398  return __ret;
34399 }
34400 #endif
34401 
34402 #ifdef __LITTLE_ENDIAN__
34403 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34404  poly8x8_t __ret;
34405  __ret = (poly8x8_t)(__p0);
34406  return __ret;
34407 }
34408 #else
34409 __ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34410  poly8x8_t __ret;
34411  __ret = (poly8x8_t)(__p0);
34412  return __ret;
34413 }
34414 #endif
34415 
34416 #ifdef __LITTLE_ENDIAN__
34417 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34418  poly8x8_t __ret;
34419  __ret = (poly8x8_t)(__p0);
34420  return __ret;
34421 }
34422 #else
34423 __ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34424  poly8x8_t __ret;
34425  __ret = (poly8x8_t)(__p0);
34426  return __ret;
34427 }
34428 #endif
34429 
34430 #ifdef __LITTLE_ENDIAN__
34431 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34432  poly8x8_t __ret;
34433  __ret = (poly8x8_t)(__p0);
34434  return __ret;
34435 }
34436 #else
34437 __ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34438  poly8x8_t __ret;
34439  __ret = (poly8x8_t)(__p0);
34440  return __ret;
34441 }
34442 #endif
34443 
34444 #ifdef __LITTLE_ENDIAN__
34445 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34446  poly8x8_t __ret;
34447  __ret = (poly8x8_t)(__p0);
34448  return __ret;
34449 }
34450 #else
34451 __ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34452  poly8x8_t __ret;
34453  __ret = (poly8x8_t)(__p0);
34454  return __ret;
34455 }
34456 #endif
34457 
34458 #ifdef __LITTLE_ENDIAN__
34459 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34460  poly8x8_t __ret;
34461  __ret = (poly8x8_t)(__p0);
34462  return __ret;
34463 }
34464 #else
34465 __ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34466  poly8x8_t __ret;
34467  __ret = (poly8x8_t)(__p0);
34468  return __ret;
34469 }
34470 #endif
34471 
34472 #ifdef __LITTLE_ENDIAN__
34473 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34474  poly8x8_t __ret;
34475  __ret = (poly8x8_t)(__p0);
34476  return __ret;
34477 }
34478 #else
34479 __ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34480  poly8x8_t __ret;
34481  __ret = (poly8x8_t)(__p0);
34482  return __ret;
34483 }
34484 #endif
34485 
34486 #ifdef __LITTLE_ENDIAN__
34487 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34488  poly8x8_t __ret;
34489  __ret = (poly8x8_t)(__p0);
34490  return __ret;
34491 }
34492 #else
34493 __ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34494  poly8x8_t __ret;
34495  __ret = (poly8x8_t)(__p0);
34496  return __ret;
34497 }
34498 #endif
34499 
34500 #ifdef __LITTLE_ENDIAN__
34501 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34502  poly8x8_t __ret;
34503  __ret = (poly8x8_t)(__p0);
34504  return __ret;
34505 }
34506 #else
34507 __ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34508  poly8x8_t __ret;
34509  __ret = (poly8x8_t)(__p0);
34510  return __ret;
34511 }
34512 #endif
34513 
34514 #ifdef __LITTLE_ENDIAN__
34515 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34516  poly8x8_t __ret;
34517  __ret = (poly8x8_t)(__p0);
34518  return __ret;
34519 }
34520 #else
34521 __ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34522  poly8x8_t __ret;
34523  __ret = (poly8x8_t)(__p0);
34524  return __ret;
34525 }
34526 #endif
34527 
34528 #ifdef __LITTLE_ENDIAN__
34529 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34530  poly8x8_t __ret;
34531  __ret = (poly8x8_t)(__p0);
34532  return __ret;
34533 }
34534 #else
34535 __ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34536  poly8x8_t __ret;
34537  __ret = (poly8x8_t)(__p0);
34538  return __ret;
34539 }
34540 #endif
34541 
34542 #ifdef __LITTLE_ENDIAN__
34543 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34544  poly8x8_t __ret;
34545  __ret = (poly8x8_t)(__p0);
34546  return __ret;
34547 }
34548 #else
34549 __ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34550  poly8x8_t __ret;
34551  __ret = (poly8x8_t)(__p0);
34552  return __ret;
34553 }
34554 #endif
34555 
34556 #ifdef __LITTLE_ENDIAN__
34557 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34558  poly8x8_t __ret;
34559  __ret = (poly8x8_t)(__p0);
34560  return __ret;
34561 }
34562 #else
34563 __ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34564  poly8x8_t __ret;
34565  __ret = (poly8x8_t)(__p0);
34566  return __ret;
34567 }
34568 #endif
34569 
34570 #ifdef __LITTLE_ENDIAN__
34571 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34572  poly64x1_t __ret;
34573  __ret = (poly64x1_t)(__p0);
34574  return __ret;
34575 }
34576 #else
34577 __ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34578  poly64x1_t __ret;
34579  __ret = (poly64x1_t)(__p0);
34580  return __ret;
34581 }
34582 #endif
34583 
34584 #ifdef __LITTLE_ENDIAN__
34585 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34586  poly64x1_t __ret;
34587  __ret = (poly64x1_t)(__p0);
34588  return __ret;
34589 }
34590 #else
34591 __ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34592  poly64x1_t __ret;
34593  __ret = (poly64x1_t)(__p0);
34594  return __ret;
34595 }
34596 #endif
34597 
34598 #ifdef __LITTLE_ENDIAN__
34599 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34600  poly64x1_t __ret;
34601  __ret = (poly64x1_t)(__p0);
34602  return __ret;
34603 }
34604 #else
34605 __ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34606  poly64x1_t __ret;
34607  __ret = (poly64x1_t)(__p0);
34608  return __ret;
34609 }
34610 #endif
34611 
34612 #ifdef __LITTLE_ENDIAN__
34613 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34614  poly64x1_t __ret;
34615  __ret = (poly64x1_t)(__p0);
34616  return __ret;
34617 }
34618 #else
34619 __ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34620  poly64x1_t __ret;
34621  __ret = (poly64x1_t)(__p0);
34622  return __ret;
34623 }
34624 #endif
34625 
34626 #ifdef __LITTLE_ENDIAN__
34627 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34628  poly64x1_t __ret;
34629  __ret = (poly64x1_t)(__p0);
34630  return __ret;
34631 }
34632 #else
34633 __ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34634  poly64x1_t __ret;
34635  __ret = (poly64x1_t)(__p0);
34636  return __ret;
34637 }
34638 #endif
34639 
34640 #ifdef __LITTLE_ENDIAN__
34641 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34642  poly64x1_t __ret;
34643  __ret = (poly64x1_t)(__p0);
34644  return __ret;
34645 }
34646 #else
34647 __ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34648  poly64x1_t __ret;
34649  __ret = (poly64x1_t)(__p0);
34650  return __ret;
34651 }
34652 #endif
34653 
34654 #ifdef __LITTLE_ENDIAN__
34655 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34656  poly64x1_t __ret;
34657  __ret = (poly64x1_t)(__p0);
34658  return __ret;
34659 }
34660 #else
34661 __ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34662  poly64x1_t __ret;
34663  __ret = (poly64x1_t)(__p0);
34664  return __ret;
34665 }
34666 #endif
34667 
34668 #ifdef __LITTLE_ENDIAN__
34669 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34670  poly64x1_t __ret;
34671  __ret = (poly64x1_t)(__p0);
34672  return __ret;
34673 }
34674 #else
34675 __ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34676  poly64x1_t __ret;
34677  __ret = (poly64x1_t)(__p0);
34678  return __ret;
34679 }
34680 #endif
34681 
34682 #ifdef __LITTLE_ENDIAN__
34683 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34684  poly64x1_t __ret;
34685  __ret = (poly64x1_t)(__p0);
34686  return __ret;
34687 }
34688 #else
34689 __ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34690  poly64x1_t __ret;
34691  __ret = (poly64x1_t)(__p0);
34692  return __ret;
34693 }
34694 #endif
34695 
34696 #ifdef __LITTLE_ENDIAN__
34697 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34698  poly64x1_t __ret;
34699  __ret = (poly64x1_t)(__p0);
34700  return __ret;
34701 }
34702 #else
34703 __ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34704  poly64x1_t __ret;
34705  __ret = (poly64x1_t)(__p0);
34706  return __ret;
34707 }
34708 #endif
34709 
34710 #ifdef __LITTLE_ENDIAN__
34711 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34712  poly64x1_t __ret;
34713  __ret = (poly64x1_t)(__p0);
34714  return __ret;
34715 }
34716 #else
34717 __ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34718  poly64x1_t __ret;
34719  __ret = (poly64x1_t)(__p0);
34720  return __ret;
34721 }
34722 #endif
34723 
34724 #ifdef __LITTLE_ENDIAN__
34725 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34726  poly64x1_t __ret;
34727  __ret = (poly64x1_t)(__p0);
34728  return __ret;
34729 }
34730 #else
34731 __ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34732  poly64x1_t __ret;
34733  __ret = (poly64x1_t)(__p0);
34734  return __ret;
34735 }
34736 #endif
34737 
34738 #ifdef __LITTLE_ENDIAN__
34739 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34740  poly64x1_t __ret;
34741  __ret = (poly64x1_t)(__p0);
34742  return __ret;
34743 }
34744 #else
34745 __ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34746  poly64x1_t __ret;
34747  __ret = (poly64x1_t)(__p0);
34748  return __ret;
34749 }
34750 #endif
34751 
34752 #ifdef __LITTLE_ENDIAN__
34753 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34754  poly16x4_t __ret;
34755  __ret = (poly16x4_t)(__p0);
34756  return __ret;
34757 }
34758 #else
34759 __ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34760  poly16x4_t __ret;
34761  __ret = (poly16x4_t)(__p0);
34762  return __ret;
34763 }
34764 #endif
34765 
34766 #ifdef __LITTLE_ENDIAN__
34767 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34768  poly16x4_t __ret;
34769  __ret = (poly16x4_t)(__p0);
34770  return __ret;
34771 }
34772 #else
34773 __ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34774  poly16x4_t __ret;
34775  __ret = (poly16x4_t)(__p0);
34776  return __ret;
34777 }
34778 #endif
34779 
34780 #ifdef __LITTLE_ENDIAN__
34781 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34782  poly16x4_t __ret;
34783  __ret = (poly16x4_t)(__p0);
34784  return __ret;
34785 }
34786 #else
34787 __ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34788  poly16x4_t __ret;
34789  __ret = (poly16x4_t)(__p0);
34790  return __ret;
34791 }
34792 #endif
34793 
34794 #ifdef __LITTLE_ENDIAN__
34795 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34796  poly16x4_t __ret;
34797  __ret = (poly16x4_t)(__p0);
34798  return __ret;
34799 }
34800 #else
34801 __ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34802  poly16x4_t __ret;
34803  __ret = (poly16x4_t)(__p0);
34804  return __ret;
34805 }
34806 #endif
34807 
34808 #ifdef __LITTLE_ENDIAN__
34809 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34810  poly16x4_t __ret;
34811  __ret = (poly16x4_t)(__p0);
34812  return __ret;
34813 }
34814 #else
34815 __ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34816  poly16x4_t __ret;
34817  __ret = (poly16x4_t)(__p0);
34818  return __ret;
34819 }
34820 #endif
34821 
34822 #ifdef __LITTLE_ENDIAN__
34823 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34824  poly16x4_t __ret;
34825  __ret = (poly16x4_t)(__p0);
34826  return __ret;
34827 }
34828 #else
34829 __ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34830  poly16x4_t __ret;
34831  __ret = (poly16x4_t)(__p0);
34832  return __ret;
34833 }
34834 #endif
34835 
34836 #ifdef __LITTLE_ENDIAN__
34837 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34838  poly16x4_t __ret;
34839  __ret = (poly16x4_t)(__p0);
34840  return __ret;
34841 }
34842 #else
34843 __ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34844  poly16x4_t __ret;
34845  __ret = (poly16x4_t)(__p0);
34846  return __ret;
34847 }
34848 #endif
34849 
34850 #ifdef __LITTLE_ENDIAN__
34851 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34852  poly16x4_t __ret;
34853  __ret = (poly16x4_t)(__p0);
34854  return __ret;
34855 }
34856 #else
34857 __ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34858  poly16x4_t __ret;
34859  __ret = (poly16x4_t)(__p0);
34860  return __ret;
34861 }
34862 #endif
34863 
34864 #ifdef __LITTLE_ENDIAN__
34865 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34866  poly16x4_t __ret;
34867  __ret = (poly16x4_t)(__p0);
34868  return __ret;
34869 }
34870 #else
34871 __ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34872  poly16x4_t __ret;
34873  __ret = (poly16x4_t)(__p0);
34874  return __ret;
34875 }
34876 #endif
34877 
34878 #ifdef __LITTLE_ENDIAN__
34879 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34880  poly16x4_t __ret;
34881  __ret = (poly16x4_t)(__p0);
34882  return __ret;
34883 }
34884 #else
34885 __ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34886  poly16x4_t __ret;
34887  __ret = (poly16x4_t)(__p0);
34888  return __ret;
34889 }
34890 #endif
34891 
34892 #ifdef __LITTLE_ENDIAN__
34893 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34894  poly16x4_t __ret;
34895  __ret = (poly16x4_t)(__p0);
34896  return __ret;
34897 }
34898 #else
34899 __ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34900  poly16x4_t __ret;
34901  __ret = (poly16x4_t)(__p0);
34902  return __ret;
34903 }
34904 #endif
34905 
34906 #ifdef __LITTLE_ENDIAN__
34907 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34908  poly16x4_t __ret;
34909  __ret = (poly16x4_t)(__p0);
34910  return __ret;
34911 }
34912 #else
34913 __ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34914  poly16x4_t __ret;
34915  __ret = (poly16x4_t)(__p0);
34916  return __ret;
34917 }
34918 #endif
34919 
34920 #ifdef __LITTLE_ENDIAN__
34921 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34922  poly16x4_t __ret;
34923  __ret = (poly16x4_t)(__p0);
34924  return __ret;
34925 }
34926 #else
34927 __ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34928  poly16x4_t __ret;
34929  __ret = (poly16x4_t)(__p0);
34930  return __ret;
34931 }
34932 #endif
34933 
34934 #ifdef __LITTLE_ENDIAN__
34935 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34936  poly8x16_t __ret;
34937  __ret = (poly8x16_t)(__p0);
34938  return __ret;
34939 }
34940 #else
34941 __ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34942  poly8x16_t __ret;
34943  __ret = (poly8x16_t)(__p0);
34944  return __ret;
34945 }
34946 #endif
34947 
34948 #ifdef __LITTLE_ENDIAN__
34949 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34950  poly8x16_t __ret;
34951  __ret = (poly8x16_t)(__p0);
34952  return __ret;
34953 }
34954 #else
34955 __ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34956  poly8x16_t __ret;
34957  __ret = (poly8x16_t)(__p0);
34958  return __ret;
34959 }
34960 #endif
34961 
34962 #ifdef __LITTLE_ENDIAN__
34963 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34964  poly8x16_t __ret;
34965  __ret = (poly8x16_t)(__p0);
34966  return __ret;
34967 }
34968 #else
34969 __ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34970  poly8x16_t __ret;
34971  __ret = (poly8x16_t)(__p0);
34972  return __ret;
34973 }
34974 #endif
34975 
34976 #ifdef __LITTLE_ENDIAN__
34977 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34978  poly8x16_t __ret;
34979  __ret = (poly8x16_t)(__p0);
34980  return __ret;
34981 }
34982 #else
34983 __ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34984  poly8x16_t __ret;
34985  __ret = (poly8x16_t)(__p0);
34986  return __ret;
34987 }
34988 #endif
34989 
34990 #ifdef __LITTLE_ENDIAN__
34991 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34992  poly8x16_t __ret;
34993  __ret = (poly8x16_t)(__p0);
34994  return __ret;
34995 }
34996 #else
34997 __ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34998  poly8x16_t __ret;
34999  __ret = (poly8x16_t)(__p0);
35000  return __ret;
35001 }
35002 #endif
35003 
35004 #ifdef __LITTLE_ENDIAN__
35005 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
35006  poly8x16_t __ret;
35007  __ret = (poly8x16_t)(__p0);
35008  return __ret;
35009 }
35010 #else
35011 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
35012  poly8x16_t __ret;
35013  __ret = (poly8x16_t)(__p0);
35014  return __ret;
35015 }
35016 #endif
35017 
35018 #ifdef __LITTLE_ENDIAN__
35019 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
35020  poly8x16_t __ret;
35021  __ret = (poly8x16_t)(__p0);
35022  return __ret;
35023 }
35024 #else
35025 __ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
35026  poly8x16_t __ret;
35027  __ret = (poly8x16_t)(__p0);
35028  return __ret;
35029 }
35030 #endif
35031 
35032 #ifdef __LITTLE_ENDIAN__
35033 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
35034  poly8x16_t __ret;
35035  __ret = (poly8x16_t)(__p0);
35036  return __ret;
35037 }
35038 #else
35039 __ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
35040  poly8x16_t __ret;
35041  __ret = (poly8x16_t)(__p0);
35042  return __ret;
35043 }
35044 #endif
35045 
35046 #ifdef __LITTLE_ENDIAN__
35047 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35048  poly8x16_t __ret;
35049  __ret = (poly8x16_t)(__p0);
35050  return __ret;
35051 }
35052 #else
35053 __ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35054  poly8x16_t __ret;
35055  __ret = (poly8x16_t)(__p0);
35056  return __ret;
35057 }
35058 #endif
35059 
35060 #ifdef __LITTLE_ENDIAN__
35061 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35062  poly8x16_t __ret;
35063  __ret = (poly8x16_t)(__p0);
35064  return __ret;
35065 }
35066 #else
35067 __ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35068  poly8x16_t __ret;
35069  __ret = (poly8x16_t)(__p0);
35070  return __ret;
35071 }
35072 #endif
35073 
35074 #ifdef __LITTLE_ENDIAN__
35075 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35076  poly8x16_t __ret;
35077  __ret = (poly8x16_t)(__p0);
35078  return __ret;
35079 }
35080 #else
35081 __ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35082  poly8x16_t __ret;
35083  __ret = (poly8x16_t)(__p0);
35084  return __ret;
35085 }
35086 #endif
35087 
35088 #ifdef __LITTLE_ENDIAN__
35089 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35090  poly8x16_t __ret;
35091  __ret = (poly8x16_t)(__p0);
35092  return __ret;
35093 }
35094 #else
35095 __ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35096  poly8x16_t __ret;
35097  __ret = (poly8x16_t)(__p0);
35098  return __ret;
35099 }
35100 #endif
35101 
35102 #ifdef __LITTLE_ENDIAN__
35103 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35104  poly8x16_t __ret;
35105  __ret = (poly8x16_t)(__p0);
35106  return __ret;
35107 }
35108 #else
35109 __ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35110  poly8x16_t __ret;
35111  __ret = (poly8x16_t)(__p0);
35112  return __ret;
35113 }
35114 #endif
35115 
35116 #ifdef __LITTLE_ENDIAN__
35117 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35118  poly8x16_t __ret;
35119  __ret = (poly8x16_t)(__p0);
35120  return __ret;
35121 }
35122 #else
35123 __ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35124  poly8x16_t __ret;
35125  __ret = (poly8x16_t)(__p0);
35126  return __ret;
35127 }
35128 #endif
35129 
35130 #ifdef __LITTLE_ENDIAN__
35131 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35132  poly128_t __ret;
35133  __ret = (poly128_t)(__p0);
35134  return __ret;
35135 }
35136 #else
35137 __ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35138  poly128_t __ret;
35139  __ret = (poly128_t)(__p0);
35140  return __ret;
35141 }
35142 #endif
35143 
35144 #ifdef __LITTLE_ENDIAN__
35145 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35146  poly128_t __ret;
35147  __ret = (poly128_t)(__p0);
35148  return __ret;
35149 }
35150 #else
35151 __ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35152  poly128_t __ret;
35153  __ret = (poly128_t)(__p0);
35154  return __ret;
35155 }
35156 #endif
35157 
35158 #ifdef __LITTLE_ENDIAN__
35159 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35160  poly128_t __ret;
35161  __ret = (poly128_t)(__p0);
35162  return __ret;
35163 }
35164 #else
35165 __ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35166  poly128_t __ret;
35167  __ret = (poly128_t)(__p0);
35168  return __ret;
35169 }
35170 #endif
35171 
35172 #ifdef __LITTLE_ENDIAN__
35173 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35174  poly128_t __ret;
35175  __ret = (poly128_t)(__p0);
35176  return __ret;
35177 }
35178 #else
35179 __ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35180  poly128_t __ret;
35181  __ret = (poly128_t)(__p0);
35182  return __ret;
35183 }
35184 #endif
35185 
35186 #ifdef __LITTLE_ENDIAN__
35187 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35188  poly128_t __ret;
35189  __ret = (poly128_t)(__p0);
35190  return __ret;
35191 }
35192 #else
35193 __ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35194  poly128_t __ret;
35195  __ret = (poly128_t)(__p0);
35196  return __ret;
35197 }
35198 #endif
35199 
35200 #ifdef __LITTLE_ENDIAN__
35201 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35202  poly128_t __ret;
35203  __ret = (poly128_t)(__p0);
35204  return __ret;
35205 }
35206 #else
35207 __ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35208  poly128_t __ret;
35209  __ret = (poly128_t)(__p0);
35210  return __ret;
35211 }
35212 #endif
35213 
35214 #ifdef __LITTLE_ENDIAN__
35215 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35216  poly128_t __ret;
35217  __ret = (poly128_t)(__p0);
35218  return __ret;
35219 }
35220 #else
35221 __ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35222  poly128_t __ret;
35223  __ret = (poly128_t)(__p0);
35224  return __ret;
35225 }
35226 #endif
35227 
35228 #ifdef __LITTLE_ENDIAN__
35229 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35230  poly128_t __ret;
35231  __ret = (poly128_t)(__p0);
35232  return __ret;
35233 }
35234 #else
35235 __ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35236  poly128_t __ret;
35237  __ret = (poly128_t)(__p0);
35238  return __ret;
35239 }
35240 #endif
35241 
35242 #ifdef __LITTLE_ENDIAN__
35243 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35244  poly128_t __ret;
35245  __ret = (poly128_t)(__p0);
35246  return __ret;
35247 }
35248 #else
35249 __ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35250  poly128_t __ret;
35251  __ret = (poly128_t)(__p0);
35252  return __ret;
35253 }
35254 #endif
35255 
35256 #ifdef __LITTLE_ENDIAN__
35257 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35258  poly128_t __ret;
35259  __ret = (poly128_t)(__p0);
35260  return __ret;
35261 }
35262 #else
35263 __ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35264  poly128_t __ret;
35265  __ret = (poly128_t)(__p0);
35266  return __ret;
35267 }
35268 #endif
35269 
35270 #ifdef __LITTLE_ENDIAN__
35271 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35272  poly128_t __ret;
35273  __ret = (poly128_t)(__p0);
35274  return __ret;
35275 }
35276 #else
35277 __ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35278  poly128_t __ret;
35279  __ret = (poly128_t)(__p0);
35280  return __ret;
35281 }
35282 #endif
35283 
35284 #ifdef __LITTLE_ENDIAN__
35285 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35286  poly128_t __ret;
35287  __ret = (poly128_t)(__p0);
35288  return __ret;
35289 }
35290 #else
35291 __ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35292  poly128_t __ret;
35293  __ret = (poly128_t)(__p0);
35294  return __ret;
35295 }
35296 #endif
35297 
35298 #ifdef __LITTLE_ENDIAN__
35299 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35300  poly128_t __ret;
35301  __ret = (poly128_t)(__p0);
35302  return __ret;
35303 }
35304 #else
35305 __ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35306  poly128_t __ret;
35307  __ret = (poly128_t)(__p0);
35308  return __ret;
35309 }
35310 #endif
35311 
35312 #ifdef __LITTLE_ENDIAN__
35313 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35314  poly128_t __ret;
35315  __ret = (poly128_t)(__p0);
35316  return __ret;
35317 }
35318 #else
35319 __ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35320  poly128_t __ret;
35321  __ret = (poly128_t)(__p0);
35322  return __ret;
35323 }
35324 #endif
35325 
35326 #ifdef __LITTLE_ENDIAN__
35327 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35328  poly64x2_t __ret;
35329  __ret = (poly64x2_t)(__p0);
35330  return __ret;
35331 }
35332 #else
35333 __ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35334  poly64x2_t __ret;
35335  __ret = (poly64x2_t)(__p0);
35336  return __ret;
35337 }
35338 #endif
35339 
35340 #ifdef __LITTLE_ENDIAN__
35341 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35342  poly64x2_t __ret;
35343  __ret = (poly64x2_t)(__p0);
35344  return __ret;
35345 }
35346 #else
35347 __ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35348  poly64x2_t __ret;
35349  __ret = (poly64x2_t)(__p0);
35350  return __ret;
35351 }
35352 #endif
35353 
35354 #ifdef __LITTLE_ENDIAN__
35355 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35356  poly64x2_t __ret;
35357  __ret = (poly64x2_t)(__p0);
35358  return __ret;
35359 }
35360 #else
35361 __ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35362  poly64x2_t __ret;
35363  __ret = (poly64x2_t)(__p0);
35364  return __ret;
35365 }
35366 #endif
35367 
35368 #ifdef __LITTLE_ENDIAN__
35369 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35370  poly64x2_t __ret;
35371  __ret = (poly64x2_t)(__p0);
35372  return __ret;
35373 }
35374 #else
35375 __ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35376  poly64x2_t __ret;
35377  __ret = (poly64x2_t)(__p0);
35378  return __ret;
35379 }
35380 #endif
35381 
35382 #ifdef __LITTLE_ENDIAN__
35383 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35384  poly64x2_t __ret;
35385  __ret = (poly64x2_t)(__p0);
35386  return __ret;
35387 }
35388 #else
35389 __ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35390  poly64x2_t __ret;
35391  __ret = (poly64x2_t)(__p0);
35392  return __ret;
35393 }
35394 #endif
35395 
35396 #ifdef __LITTLE_ENDIAN__
35397 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35398  poly64x2_t __ret;
35399  __ret = (poly64x2_t)(__p0);
35400  return __ret;
35401 }
35402 #else
35403 __ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35404  poly64x2_t __ret;
35405  __ret = (poly64x2_t)(__p0);
35406  return __ret;
35407 }
35408 #endif
35409 
35410 #ifdef __LITTLE_ENDIAN__
35411 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35412  poly64x2_t __ret;
35413  __ret = (poly64x2_t)(__p0);
35414  return __ret;
35415 }
35416 #else
35417 __ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35418  poly64x2_t __ret;
35419  __ret = (poly64x2_t)(__p0);
35420  return __ret;
35421 }
35422 #endif
35423 
35424 #ifdef __LITTLE_ENDIAN__
35425 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35426  poly64x2_t __ret;
35427  __ret = (poly64x2_t)(__p0);
35428  return __ret;
35429 }
35430 #else
35431 __ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35432  poly64x2_t __ret;
35433  __ret = (poly64x2_t)(__p0);
35434  return __ret;
35435 }
35436 #endif
35437 
35438 #ifdef __LITTLE_ENDIAN__
35439 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35440  poly64x2_t __ret;
35441  __ret = (poly64x2_t)(__p0);
35442  return __ret;
35443 }
35444 #else
35445 __ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35446  poly64x2_t __ret;
35447  __ret = (poly64x2_t)(__p0);
35448  return __ret;
35449 }
35450 #endif
35451 
35452 #ifdef __LITTLE_ENDIAN__
35453 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35454  poly64x2_t __ret;
35455  __ret = (poly64x2_t)(__p0);
35456  return __ret;
35457 }
35458 #else
35459 __ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35460  poly64x2_t __ret;
35461  __ret = (poly64x2_t)(__p0);
35462  return __ret;
35463 }
35464 #endif
35465 
35466 #ifdef __LITTLE_ENDIAN__
35467 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35468  poly64x2_t __ret;
35469  __ret = (poly64x2_t)(__p0);
35470  return __ret;
35471 }
35472 #else
35473 __ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35474  poly64x2_t __ret;
35475  __ret = (poly64x2_t)(__p0);
35476  return __ret;
35477 }
35478 #endif
35479 
35480 #ifdef __LITTLE_ENDIAN__
35481 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35482  poly64x2_t __ret;
35483  __ret = (poly64x2_t)(__p0);
35484  return __ret;
35485 }
35486 #else
35487 __ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35488  poly64x2_t __ret;
35489  __ret = (poly64x2_t)(__p0);
35490  return __ret;
35491 }
35492 #endif
35493 
35494 #ifdef __LITTLE_ENDIAN__
35495 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35496  poly64x2_t __ret;
35497  __ret = (poly64x2_t)(__p0);
35498  return __ret;
35499 }
35500 #else
35501 __ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35502  poly64x2_t __ret;
35503  __ret = (poly64x2_t)(__p0);
35504  return __ret;
35505 }
35506 #endif
35507 
35508 #ifdef __LITTLE_ENDIAN__
35509 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35510  poly64x2_t __ret;
35511  __ret = (poly64x2_t)(__p0);
35512  return __ret;
35513 }
35514 #else
35515 __ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35516  poly64x2_t __ret;
35517  __ret = (poly64x2_t)(__p0);
35518  return __ret;
35519 }
35520 #endif
35521 
35522 #ifdef __LITTLE_ENDIAN__
35523 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35524  poly16x8_t __ret;
35525  __ret = (poly16x8_t)(__p0);
35526  return __ret;
35527 }
35528 #else
35529 __ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35530  poly16x8_t __ret;
35531  __ret = (poly16x8_t)(__p0);
35532  return __ret;
35533 }
35534 #endif
35535 
35536 #ifdef __LITTLE_ENDIAN__
35537 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35538  poly16x8_t __ret;
35539  __ret = (poly16x8_t)(__p0);
35540  return __ret;
35541 }
35542 #else
35543 __ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35544  poly16x8_t __ret;
35545  __ret = (poly16x8_t)(__p0);
35546  return __ret;
35547 }
35548 #endif
35549 
35550 #ifdef __LITTLE_ENDIAN__
35551 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35552  poly16x8_t __ret;
35553  __ret = (poly16x8_t)(__p0);
35554  return __ret;
35555 }
35556 #else
35557 __ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35558  poly16x8_t __ret;
35559  __ret = (poly16x8_t)(__p0);
35560  return __ret;
35561 }
35562 #endif
35563 
35564 #ifdef __LITTLE_ENDIAN__
35565 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35566  poly16x8_t __ret;
35567  __ret = (poly16x8_t)(__p0);
35568  return __ret;
35569 }
35570 #else
35571 __ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35572  poly16x8_t __ret;
35573  __ret = (poly16x8_t)(__p0);
35574  return __ret;
35575 }
35576 #endif
35577 
35578 #ifdef __LITTLE_ENDIAN__
35579 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35580  poly16x8_t __ret;
35581  __ret = (poly16x8_t)(__p0);
35582  return __ret;
35583 }
35584 #else
35585 __ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35586  poly16x8_t __ret;
35587  __ret = (poly16x8_t)(__p0);
35588  return __ret;
35589 }
35590 #endif
35591 
35592 #ifdef __LITTLE_ENDIAN__
35593 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35594  poly16x8_t __ret;
35595  __ret = (poly16x8_t)(__p0);
35596  return __ret;
35597 }
35598 #else
35599 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35600  poly16x8_t __ret;
35601  __ret = (poly16x8_t)(__p0);
35602  return __ret;
35603 }
35604 #endif
35605 
35606 #ifdef __LITTLE_ENDIAN__
35607 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35608  poly16x8_t __ret;
35609  __ret = (poly16x8_t)(__p0);
35610  return __ret;
35611 }
35612 #else
35613 __ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35614  poly16x8_t __ret;
35615  __ret = (poly16x8_t)(__p0);
35616  return __ret;
35617 }
35618 #endif
35619 
35620 #ifdef __LITTLE_ENDIAN__
35621 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35622  poly16x8_t __ret;
35623  __ret = (poly16x8_t)(__p0);
35624  return __ret;
35625 }
35626 #else
35627 __ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35628  poly16x8_t __ret;
35629  __ret = (poly16x8_t)(__p0);
35630  return __ret;
35631 }
35632 #endif
35633 
35634 #ifdef __LITTLE_ENDIAN__
35635 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35636  poly16x8_t __ret;
35637  __ret = (poly16x8_t)(__p0);
35638  return __ret;
35639 }
35640 #else
35641 __ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35642  poly16x8_t __ret;
35643  __ret = (poly16x8_t)(__p0);
35644  return __ret;
35645 }
35646 #endif
35647 
35648 #ifdef __LITTLE_ENDIAN__
35649 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35650  poly16x8_t __ret;
35651  __ret = (poly16x8_t)(__p0);
35652  return __ret;
35653 }
35654 #else
35655 __ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35656  poly16x8_t __ret;
35657  __ret = (poly16x8_t)(__p0);
35658  return __ret;
35659 }
35660 #endif
35661 
35662 #ifdef __LITTLE_ENDIAN__
35663 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35664  poly16x8_t __ret;
35665  __ret = (poly16x8_t)(__p0);
35666  return __ret;
35667 }
35668 #else
35669 __ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35670  poly16x8_t __ret;
35671  __ret = (poly16x8_t)(__p0);
35672  return __ret;
35673 }
35674 #endif
35675 
35676 #ifdef __LITTLE_ENDIAN__
35677 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35678  poly16x8_t __ret;
35679  __ret = (poly16x8_t)(__p0);
35680  return __ret;
35681 }
35682 #else
35683 __ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35684  poly16x8_t __ret;
35685  __ret = (poly16x8_t)(__p0);
35686  return __ret;
35687 }
35688 #endif
35689 
35690 #ifdef __LITTLE_ENDIAN__
35691 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35692  poly16x8_t __ret;
35693  __ret = (poly16x8_t)(__p0);
35694  return __ret;
35695 }
35696 #else
35697 __ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35698  poly16x8_t __ret;
35699  __ret = (poly16x8_t)(__p0);
35700  return __ret;
35701 }
35702 #endif
35703 
35704 #ifdef __LITTLE_ENDIAN__
35705 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35706  poly16x8_t __ret;
35707  __ret = (poly16x8_t)(__p0);
35708  return __ret;
35709 }
35710 #else
35711 __ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35712  poly16x8_t __ret;
35713  __ret = (poly16x8_t)(__p0);
35714  return __ret;
35715 }
35716 #endif
35717 
35718 #ifdef __LITTLE_ENDIAN__
35719 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35720  uint8x16_t __ret;
35721  __ret = (uint8x16_t)(__p0);
35722  return __ret;
35723 }
35724 #else
35725 __ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35726  uint8x16_t __ret;
35727  __ret = (uint8x16_t)(__p0);
35728  return __ret;
35729 }
35730 #endif
35731 
35732 #ifdef __LITTLE_ENDIAN__
35733 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35734  uint8x16_t __ret;
35735  __ret = (uint8x16_t)(__p0);
35736  return __ret;
35737 }
35738 #else
35739 __ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35740  uint8x16_t __ret;
35741  __ret = (uint8x16_t)(__p0);
35742  return __ret;
35743 }
35744 #endif
35745 
35746 #ifdef __LITTLE_ENDIAN__
35747 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35748  uint8x16_t __ret;
35749  __ret = (uint8x16_t)(__p0);
35750  return __ret;
35751 }
35752 #else
35753 __ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35754  uint8x16_t __ret;
35755  __ret = (uint8x16_t)(__p0);
35756  return __ret;
35757 }
35758 #endif
35759 
35760 #ifdef __LITTLE_ENDIAN__
35761 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35762  uint8x16_t __ret;
35763  __ret = (uint8x16_t)(__p0);
35764  return __ret;
35765 }
35766 #else
35767 __ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35768  uint8x16_t __ret;
35769  __ret = (uint8x16_t)(__p0);
35770  return __ret;
35771 }
35772 #endif
35773 
35774 #ifdef __LITTLE_ENDIAN__
35775 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35776  uint8x16_t __ret;
35777  __ret = (uint8x16_t)(__p0);
35778  return __ret;
35779 }
35780 #else
35781 __ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35782  uint8x16_t __ret;
35783  __ret = (uint8x16_t)(__p0);
35784  return __ret;
35785 }
35786 #endif
35787 
35788 #ifdef __LITTLE_ENDIAN__
35789 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35790  uint8x16_t __ret;
35791  __ret = (uint8x16_t)(__p0);
35792  return __ret;
35793 }
35794 #else
35795 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35796  uint8x16_t __ret;
35797  __ret = (uint8x16_t)(__p0);
35798  return __ret;
35799 }
35800 #endif
35801 
35802 #ifdef __LITTLE_ENDIAN__
35803 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35804  uint8x16_t __ret;
35805  __ret = (uint8x16_t)(__p0);
35806  return __ret;
35807 }
35808 #else
35809 __ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35810  uint8x16_t __ret;
35811  __ret = (uint8x16_t)(__p0);
35812  return __ret;
35813 }
35814 #endif
35815 
35816 #ifdef __LITTLE_ENDIAN__
35817 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35818  uint8x16_t __ret;
35819  __ret = (uint8x16_t)(__p0);
35820  return __ret;
35821 }
35822 #else
35823 __ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35824  uint8x16_t __ret;
35825  __ret = (uint8x16_t)(__p0);
35826  return __ret;
35827 }
35828 #endif
35829 
35830 #ifdef __LITTLE_ENDIAN__
35831 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35832  uint8x16_t __ret;
35833  __ret = (uint8x16_t)(__p0);
35834  return __ret;
35835 }
35836 #else
35837 __ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35838  uint8x16_t __ret;
35839  __ret = (uint8x16_t)(__p0);
35840  return __ret;
35841 }
35842 #endif
35843 
35844 #ifdef __LITTLE_ENDIAN__
35845 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35846  uint8x16_t __ret;
35847  __ret = (uint8x16_t)(__p0);
35848  return __ret;
35849 }
35850 #else
35851 __ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35852  uint8x16_t __ret;
35853  __ret = (uint8x16_t)(__p0);
35854  return __ret;
35855 }
35856 #endif
35857 
35858 #ifdef __LITTLE_ENDIAN__
35859 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35860  uint8x16_t __ret;
35861  __ret = (uint8x16_t)(__p0);
35862  return __ret;
35863 }
35864 #else
35865 __ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35866  uint8x16_t __ret;
35867  __ret = (uint8x16_t)(__p0);
35868  return __ret;
35869 }
35870 #endif
35871 
35872 #ifdef __LITTLE_ENDIAN__
35873 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35874  uint8x16_t __ret;
35875  __ret = (uint8x16_t)(__p0);
35876  return __ret;
35877 }
35878 #else
35879 __ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35880  uint8x16_t __ret;
35881  __ret = (uint8x16_t)(__p0);
35882  return __ret;
35883 }
35884 #endif
35885 
35886 #ifdef __LITTLE_ENDIAN__
35887 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35888  uint8x16_t __ret;
35889  __ret = (uint8x16_t)(__p0);
35890  return __ret;
35891 }
35892 #else
35893 __ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35894  uint8x16_t __ret;
35895  __ret = (uint8x16_t)(__p0);
35896  return __ret;
35897 }
35898 #endif
35899 
35900 #ifdef __LITTLE_ENDIAN__
35901 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35902  uint8x16_t __ret;
35903  __ret = (uint8x16_t)(__p0);
35904  return __ret;
35905 }
35906 #else
35907 __ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35908  uint8x16_t __ret;
35909  __ret = (uint8x16_t)(__p0);
35910  return __ret;
35911 }
35912 #endif
35913 
35914 #ifdef __LITTLE_ENDIAN__
35915 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35916  uint32x4_t __ret;
35917  __ret = (uint32x4_t)(__p0);
35918  return __ret;
35919 }
35920 #else
35921 __ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35922  uint32x4_t __ret;
35923  __ret = (uint32x4_t)(__p0);
35924  return __ret;
35925 }
35926 #endif
35927 
35928 #ifdef __LITTLE_ENDIAN__
35929 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35930  uint32x4_t __ret;
35931  __ret = (uint32x4_t)(__p0);
35932  return __ret;
35933 }
35934 #else
35935 __ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35936  uint32x4_t __ret;
35937  __ret = (uint32x4_t)(__p0);
35938  return __ret;
35939 }
35940 #endif
35941 
35942 #ifdef __LITTLE_ENDIAN__
35943 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35944  uint32x4_t __ret;
35945  __ret = (uint32x4_t)(__p0);
35946  return __ret;
35947 }
35948 #else
35949 __ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35950  uint32x4_t __ret;
35951  __ret = (uint32x4_t)(__p0);
35952  return __ret;
35953 }
35954 #endif
35955 
35956 #ifdef __LITTLE_ENDIAN__
35957 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35958  uint32x4_t __ret;
35959  __ret = (uint32x4_t)(__p0);
35960  return __ret;
35961 }
35962 #else
35963 __ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35964  uint32x4_t __ret;
35965  __ret = (uint32x4_t)(__p0);
35966  return __ret;
35967 }
35968 #endif
35969 
35970 #ifdef __LITTLE_ENDIAN__
35971 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35972  uint32x4_t __ret;
35973  __ret = (uint32x4_t)(__p0);
35974  return __ret;
35975 }
35976 #else
35977 __ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35978  uint32x4_t __ret;
35979  __ret = (uint32x4_t)(__p0);
35980  return __ret;
35981 }
35982 #endif
35983 
35984 #ifdef __LITTLE_ENDIAN__
35985 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35986  uint32x4_t __ret;
35987  __ret = (uint32x4_t)(__p0);
35988  return __ret;
35989 }
35990 #else
35991 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35992  uint32x4_t __ret;
35993  __ret = (uint32x4_t)(__p0);
35994  return __ret;
35995 }
35996 #endif
35997 
35998 #ifdef __LITTLE_ENDIAN__
35999 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
36000  uint32x4_t __ret;
36001  __ret = (uint32x4_t)(__p0);
36002  return __ret;
36003 }
36004 #else
36005 __ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
36006  uint32x4_t __ret;
36007  __ret = (uint32x4_t)(__p0);
36008  return __ret;
36009 }
36010 #endif
36011 
36012 #ifdef __LITTLE_ENDIAN__
36013 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
36014  uint32x4_t __ret;
36015  __ret = (uint32x4_t)(__p0);
36016  return __ret;
36017 }
36018 #else
36019 __ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
36020  uint32x4_t __ret;
36021  __ret = (uint32x4_t)(__p0);
36022  return __ret;
36023 }
36024 #endif
36025 
36026 #ifdef __LITTLE_ENDIAN__
36027 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
36028  uint32x4_t __ret;
36029  __ret = (uint32x4_t)(__p0);
36030  return __ret;
36031 }
36032 #else
36033 __ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
36034  uint32x4_t __ret;
36035  __ret = (uint32x4_t)(__p0);
36036  return __ret;
36037 }
36038 #endif
36039 
36040 #ifdef __LITTLE_ENDIAN__
36041 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36042  uint32x4_t __ret;
36043  __ret = (uint32x4_t)(__p0);
36044  return __ret;
36045 }
36046 #else
36047 __ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36048  uint32x4_t __ret;
36049  __ret = (uint32x4_t)(__p0);
36050  return __ret;
36051 }
36052 #endif
36053 
36054 #ifdef __LITTLE_ENDIAN__
36055 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36056  uint32x4_t __ret;
36057  __ret = (uint32x4_t)(__p0);
36058  return __ret;
36059 }
36060 #else
36061 __ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36062  uint32x4_t __ret;
36063  __ret = (uint32x4_t)(__p0);
36064  return __ret;
36065 }
36066 #endif
36067 
36068 #ifdef __LITTLE_ENDIAN__
36069 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36070  uint32x4_t __ret;
36071  __ret = (uint32x4_t)(__p0);
36072  return __ret;
36073 }
36074 #else
36075 __ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36076  uint32x4_t __ret;
36077  __ret = (uint32x4_t)(__p0);
36078  return __ret;
36079 }
36080 #endif
36081 
36082 #ifdef __LITTLE_ENDIAN__
36083 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36084  uint32x4_t __ret;
36085  __ret = (uint32x4_t)(__p0);
36086  return __ret;
36087 }
36088 #else
36089 __ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36090  uint32x4_t __ret;
36091  __ret = (uint32x4_t)(__p0);
36092  return __ret;
36093 }
36094 #endif
36095 
36096 #ifdef __LITTLE_ENDIAN__
36097 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36098  uint32x4_t __ret;
36099  __ret = (uint32x4_t)(__p0);
36100  return __ret;
36101 }
36102 #else
36103 __ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36104  uint32x4_t __ret;
36105  __ret = (uint32x4_t)(__p0);
36106  return __ret;
36107 }
36108 #endif
36109 
36110 #ifdef __LITTLE_ENDIAN__
36111 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36112  uint64x2_t __ret;
36113  __ret = (uint64x2_t)(__p0);
36114  return __ret;
36115 }
36116 #else
36117 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36118  uint64x2_t __ret;
36119  __ret = (uint64x2_t)(__p0);
36120  return __ret;
36121 }
36122 #endif
36123 
36124 #ifdef __LITTLE_ENDIAN__
36125 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36126  uint64x2_t __ret;
36127  __ret = (uint64x2_t)(__p0);
36128  return __ret;
36129 }
36130 #else
36131 __ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36132  uint64x2_t __ret;
36133  __ret = (uint64x2_t)(__p0);
36134  return __ret;
36135 }
36136 #endif
36137 
36138 #ifdef __LITTLE_ENDIAN__
36139 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36140  uint64x2_t __ret;
36141  __ret = (uint64x2_t)(__p0);
36142  return __ret;
36143 }
36144 #else
36145 __ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36146  uint64x2_t __ret;
36147  __ret = (uint64x2_t)(__p0);
36148  return __ret;
36149 }
36150 #endif
36151 
36152 #ifdef __LITTLE_ENDIAN__
36153 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36154  uint64x2_t __ret;
36155  __ret = (uint64x2_t)(__p0);
36156  return __ret;
36157 }
36158 #else
36159 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36160  uint64x2_t __ret;
36161  __ret = (uint64x2_t)(__p0);
36162  return __ret;
36163 }
36164 #endif
36165 
36166 #ifdef __LITTLE_ENDIAN__
36167 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36168  uint64x2_t __ret;
36169  __ret = (uint64x2_t)(__p0);
36170  return __ret;
36171 }
36172 #else
36173 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36174  uint64x2_t __ret;
36175  __ret = (uint64x2_t)(__p0);
36176  return __ret;
36177 }
36178 #endif
36179 
36180 #ifdef __LITTLE_ENDIAN__
36181 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36182  uint64x2_t __ret;
36183  __ret = (uint64x2_t)(__p0);
36184  return __ret;
36185 }
36186 #else
36187 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36188  uint64x2_t __ret;
36189  __ret = (uint64x2_t)(__p0);
36190  return __ret;
36191 }
36192 #endif
36193 
36194 #ifdef __LITTLE_ENDIAN__
36195 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36196  uint64x2_t __ret;
36197  __ret = (uint64x2_t)(__p0);
36198  return __ret;
36199 }
36200 #else
36201 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36202  uint64x2_t __ret;
36203  __ret = (uint64x2_t)(__p0);
36204  return __ret;
36205 }
36206 #endif
36207 
36208 #ifdef __LITTLE_ENDIAN__
36209 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36210  uint64x2_t __ret;
36211  __ret = (uint64x2_t)(__p0);
36212  return __ret;
36213 }
36214 #else
36215 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36216  uint64x2_t __ret;
36217  __ret = (uint64x2_t)(__p0);
36218  return __ret;
36219 }
36220 #endif
36221 
36222 #ifdef __LITTLE_ENDIAN__
36223 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36224  uint64x2_t __ret;
36225  __ret = (uint64x2_t)(__p0);
36226  return __ret;
36227 }
36228 #else
36229 __ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36230  uint64x2_t __ret;
36231  __ret = (uint64x2_t)(__p0);
36232  return __ret;
36233 }
36234 #endif
36235 
36236 #ifdef __LITTLE_ENDIAN__
36237 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36238  uint64x2_t __ret;
36239  __ret = (uint64x2_t)(__p0);
36240  return __ret;
36241 }
36242 #else
36243 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36244  uint64x2_t __ret;
36245  __ret = (uint64x2_t)(__p0);
36246  return __ret;
36247 }
36248 #endif
36249 
36250 #ifdef __LITTLE_ENDIAN__
36251 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36252  uint64x2_t __ret;
36253  __ret = (uint64x2_t)(__p0);
36254  return __ret;
36255 }
36256 #else
36257 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36258  uint64x2_t __ret;
36259  __ret = (uint64x2_t)(__p0);
36260  return __ret;
36261 }
36262 #endif
36263 
36264 #ifdef __LITTLE_ENDIAN__
36265 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36266  uint64x2_t __ret;
36267  __ret = (uint64x2_t)(__p0);
36268  return __ret;
36269 }
36270 #else
36271 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36272  uint64x2_t __ret;
36273  __ret = (uint64x2_t)(__p0);
36274  return __ret;
36275 }
36276 #endif
36277 
36278 #ifdef __LITTLE_ENDIAN__
36279 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36280  uint64x2_t __ret;
36281  __ret = (uint64x2_t)(__p0);
36282  return __ret;
36283 }
36284 #else
36285 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36286  uint64x2_t __ret;
36287  __ret = (uint64x2_t)(__p0);
36288  return __ret;
36289 }
36290 #endif
36291 
36292 #ifdef __LITTLE_ENDIAN__
36293 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36294  uint64x2_t __ret;
36295  __ret = (uint64x2_t)(__p0);
36296  return __ret;
36297 }
36298 #else
36299 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36300  uint64x2_t __ret;
36301  __ret = (uint64x2_t)(__p0);
36302  return __ret;
36303 }
36304 #endif
36305 
36306 #ifdef __LITTLE_ENDIAN__
36307 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36308  uint16x8_t __ret;
36309  __ret = (uint16x8_t)(__p0);
36310  return __ret;
36311 }
36312 #else
36313 __ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36314  uint16x8_t __ret;
36315  __ret = (uint16x8_t)(__p0);
36316  return __ret;
36317 }
36318 #endif
36319 
36320 #ifdef __LITTLE_ENDIAN__
36321 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36322  uint16x8_t __ret;
36323  __ret = (uint16x8_t)(__p0);
36324  return __ret;
36325 }
36326 #else
36327 __ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36328  uint16x8_t __ret;
36329  __ret = (uint16x8_t)(__p0);
36330  return __ret;
36331 }
36332 #endif
36333 
36334 #ifdef __LITTLE_ENDIAN__
36335 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36336  uint16x8_t __ret;
36337  __ret = (uint16x8_t)(__p0);
36338  return __ret;
36339 }
36340 #else
36341 __ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36342  uint16x8_t __ret;
36343  __ret = (uint16x8_t)(__p0);
36344  return __ret;
36345 }
36346 #endif
36347 
36348 #ifdef __LITTLE_ENDIAN__
36349 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36350  uint16x8_t __ret;
36351  __ret = (uint16x8_t)(__p0);
36352  return __ret;
36353 }
36354 #else
36355 __ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36356  uint16x8_t __ret;
36357  __ret = (uint16x8_t)(__p0);
36358  return __ret;
36359 }
36360 #endif
36361 
36362 #ifdef __LITTLE_ENDIAN__
36363 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36364  uint16x8_t __ret;
36365  __ret = (uint16x8_t)(__p0);
36366  return __ret;
36367 }
36368 #else
36369 __ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36370  uint16x8_t __ret;
36371  __ret = (uint16x8_t)(__p0);
36372  return __ret;
36373 }
36374 #endif
36375 
36376 #ifdef __LITTLE_ENDIAN__
36377 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36378  uint16x8_t __ret;
36379  __ret = (uint16x8_t)(__p0);
36380  return __ret;
36381 }
36382 #else
36383 __ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36384  uint16x8_t __ret;
36385  __ret = (uint16x8_t)(__p0);
36386  return __ret;
36387 }
36388 #endif
36389 
36390 #ifdef __LITTLE_ENDIAN__
36391 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36392  uint16x8_t __ret;
36393  __ret = (uint16x8_t)(__p0);
36394  return __ret;
36395 }
36396 #else
36397 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36398  uint16x8_t __ret;
36399  __ret = (uint16x8_t)(__p0);
36400  return __ret;
36401 }
36402 #endif
36403 
36404 #ifdef __LITTLE_ENDIAN__
36405 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36406  uint16x8_t __ret;
36407  __ret = (uint16x8_t)(__p0);
36408  return __ret;
36409 }
36410 #else
36411 __ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36412  uint16x8_t __ret;
36413  __ret = (uint16x8_t)(__p0);
36414  return __ret;
36415 }
36416 #endif
36417 
36418 #ifdef __LITTLE_ENDIAN__
36419 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36420  uint16x8_t __ret;
36421  __ret = (uint16x8_t)(__p0);
36422  return __ret;
36423 }
36424 #else
36425 __ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36426  uint16x8_t __ret;
36427  __ret = (uint16x8_t)(__p0);
36428  return __ret;
36429 }
36430 #endif
36431 
36432 #ifdef __LITTLE_ENDIAN__
36433 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36434  uint16x8_t __ret;
36435  __ret = (uint16x8_t)(__p0);
36436  return __ret;
36437 }
36438 #else
36439 __ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36440  uint16x8_t __ret;
36441  __ret = (uint16x8_t)(__p0);
36442  return __ret;
36443 }
36444 #endif
36445 
36446 #ifdef __LITTLE_ENDIAN__
36447 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36448  uint16x8_t __ret;
36449  __ret = (uint16x8_t)(__p0);
36450  return __ret;
36451 }
36452 #else
36453 __ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36454  uint16x8_t __ret;
36455  __ret = (uint16x8_t)(__p0);
36456  return __ret;
36457 }
36458 #endif
36459 
36460 #ifdef __LITTLE_ENDIAN__
36461 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36462  uint16x8_t __ret;
36463  __ret = (uint16x8_t)(__p0);
36464  return __ret;
36465 }
36466 #else
36467 __ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36468  uint16x8_t __ret;
36469  __ret = (uint16x8_t)(__p0);
36470  return __ret;
36471 }
36472 #endif
36473 
36474 #ifdef __LITTLE_ENDIAN__
36475 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36476  uint16x8_t __ret;
36477  __ret = (uint16x8_t)(__p0);
36478  return __ret;
36479 }
36480 #else
36481 __ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36482  uint16x8_t __ret;
36483  __ret = (uint16x8_t)(__p0);
36484  return __ret;
36485 }
36486 #endif
36487 
36488 #ifdef __LITTLE_ENDIAN__
36489 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36490  uint16x8_t __ret;
36491  __ret = (uint16x8_t)(__p0);
36492  return __ret;
36493 }
36494 #else
36495 __ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36496  uint16x8_t __ret;
36497  __ret = (uint16x8_t)(__p0);
36498  return __ret;
36499 }
36500 #endif
36501 
36502 #ifdef __LITTLE_ENDIAN__
36503 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36504  int8x16_t __ret;
36505  __ret = (int8x16_t)(__p0);
36506  return __ret;
36507 }
36508 #else
36509 __ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36510  int8x16_t __ret;
36511  __ret = (int8x16_t)(__p0);
36512  return __ret;
36513 }
36514 #endif
36515 
36516 #ifdef __LITTLE_ENDIAN__
36517 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36518  int8x16_t __ret;
36519  __ret = (int8x16_t)(__p0);
36520  return __ret;
36521 }
36522 #else
36523 __ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36524  int8x16_t __ret;
36525  __ret = (int8x16_t)(__p0);
36526  return __ret;
36527 }
36528 #endif
36529 
36530 #ifdef __LITTLE_ENDIAN__
36531 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36532  int8x16_t __ret;
36533  __ret = (int8x16_t)(__p0);
36534  return __ret;
36535 }
36536 #else
36537 __ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36538  int8x16_t __ret;
36539  __ret = (int8x16_t)(__p0);
36540  return __ret;
36541 }
36542 #endif
36543 
36544 #ifdef __LITTLE_ENDIAN__
36545 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36546  int8x16_t __ret;
36547  __ret = (int8x16_t)(__p0);
36548  return __ret;
36549 }
36550 #else
36551 __ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36552  int8x16_t __ret;
36553  __ret = (int8x16_t)(__p0);
36554  return __ret;
36555 }
36556 #endif
36557 
36558 #ifdef __LITTLE_ENDIAN__
36559 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36560  int8x16_t __ret;
36561  __ret = (int8x16_t)(__p0);
36562  return __ret;
36563 }
36564 #else
36565 __ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36566  int8x16_t __ret;
36567  __ret = (int8x16_t)(__p0);
36568  return __ret;
36569 }
36570 #endif
36571 
36572 #ifdef __LITTLE_ENDIAN__
36573 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36574  int8x16_t __ret;
36575  __ret = (int8x16_t)(__p0);
36576  return __ret;
36577 }
36578 #else
36579 __ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36580  int8x16_t __ret;
36581  __ret = (int8x16_t)(__p0);
36582  return __ret;
36583 }
36584 #endif
36585 
36586 #ifdef __LITTLE_ENDIAN__
36587 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36588  int8x16_t __ret;
36589  __ret = (int8x16_t)(__p0);
36590  return __ret;
36591 }
36592 #else
36593 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36594  int8x16_t __ret;
36595  __ret = (int8x16_t)(__p0);
36596  return __ret;
36597 }
36598 #endif
36599 
36600 #ifdef __LITTLE_ENDIAN__
36601 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36602  int8x16_t __ret;
36603  __ret = (int8x16_t)(__p0);
36604  return __ret;
36605 }
36606 #else
36607 __ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36608  int8x16_t __ret;
36609  __ret = (int8x16_t)(__p0);
36610  return __ret;
36611 }
36612 #endif
36613 
36614 #ifdef __LITTLE_ENDIAN__
36615 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36616  int8x16_t __ret;
36617  __ret = (int8x16_t)(__p0);
36618  return __ret;
36619 }
36620 #else
36621 __ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36622  int8x16_t __ret;
36623  __ret = (int8x16_t)(__p0);
36624  return __ret;
36625 }
36626 #endif
36627 
36628 #ifdef __LITTLE_ENDIAN__
36629 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36630  int8x16_t __ret;
36631  __ret = (int8x16_t)(__p0);
36632  return __ret;
36633 }
36634 #else
36635 __ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36636  int8x16_t __ret;
36637  __ret = (int8x16_t)(__p0);
36638  return __ret;
36639 }
36640 #endif
36641 
36642 #ifdef __LITTLE_ENDIAN__
36643 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36644  int8x16_t __ret;
36645  __ret = (int8x16_t)(__p0);
36646  return __ret;
36647 }
36648 #else
36649 __ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36650  int8x16_t __ret;
36651  __ret = (int8x16_t)(__p0);
36652  return __ret;
36653 }
36654 #endif
36655 
36656 #ifdef __LITTLE_ENDIAN__
36657 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36658  int8x16_t __ret;
36659  __ret = (int8x16_t)(__p0);
36660  return __ret;
36661 }
36662 #else
36663 __ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36664  int8x16_t __ret;
36665  __ret = (int8x16_t)(__p0);
36666  return __ret;
36667 }
36668 #endif
36669 
36670 #ifdef __LITTLE_ENDIAN__
36671 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36672  int8x16_t __ret;
36673  __ret = (int8x16_t)(__p0);
36674  return __ret;
36675 }
36676 #else
36677 __ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36678  int8x16_t __ret;
36679  __ret = (int8x16_t)(__p0);
36680  return __ret;
36681 }
36682 #endif
36683 
36684 #ifdef __LITTLE_ENDIAN__
36685 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36686  int8x16_t __ret;
36687  __ret = (int8x16_t)(__p0);
36688  return __ret;
36689 }
36690 #else
36691 __ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36692  int8x16_t __ret;
36693  __ret = (int8x16_t)(__p0);
36694  return __ret;
36695 }
36696 #endif
36697 
36698 #ifdef __LITTLE_ENDIAN__
36699 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36700  float64x2_t __ret;
36701  __ret = (float64x2_t)(__p0);
36702  return __ret;
36703 }
36704 #else
36705 __ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36706  float64x2_t __ret;
36707  __ret = (float64x2_t)(__p0);
36708  return __ret;
36709 }
36710 #endif
36711 
36712 #ifdef __LITTLE_ENDIAN__
36713 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36714  float64x2_t __ret;
36715  __ret = (float64x2_t)(__p0);
36716  return __ret;
36717 }
36718 #else
36719 __ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36720  float64x2_t __ret;
36721  __ret = (float64x2_t)(__p0);
36722  return __ret;
36723 }
36724 #endif
36725 
36726 #ifdef __LITTLE_ENDIAN__
36727 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36728  float64x2_t __ret;
36729  __ret = (float64x2_t)(__p0);
36730  return __ret;
36731 }
36732 #else
36733 __ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36734  float64x2_t __ret;
36735  __ret = (float64x2_t)(__p0);
36736  return __ret;
36737 }
36738 #endif
36739 
36740 #ifdef __LITTLE_ENDIAN__
36741 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36742  float64x2_t __ret;
36743  __ret = (float64x2_t)(__p0);
36744  return __ret;
36745 }
36746 #else
36747 __ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36748  float64x2_t __ret;
36749  __ret = (float64x2_t)(__p0);
36750  return __ret;
36751 }
36752 #endif
36753 
36754 #ifdef __LITTLE_ENDIAN__
36755 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36756  float64x2_t __ret;
36757  __ret = (float64x2_t)(__p0);
36758  return __ret;
36759 }
36760 #else
36761 __ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36762  float64x2_t __ret;
36763  __ret = (float64x2_t)(__p0);
36764  return __ret;
36765 }
36766 #endif
36767 
36768 #ifdef __LITTLE_ENDIAN__
36769 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36770  float64x2_t __ret;
36771  __ret = (float64x2_t)(__p0);
36772  return __ret;
36773 }
36774 #else
36775 __ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36776  float64x2_t __ret;
36777  __ret = (float64x2_t)(__p0);
36778  return __ret;
36779 }
36780 #endif
36781 
36782 #ifdef __LITTLE_ENDIAN__
36783 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36784  float64x2_t __ret;
36785  __ret = (float64x2_t)(__p0);
36786  return __ret;
36787 }
36788 #else
36789 __ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36790  float64x2_t __ret;
36791  __ret = (float64x2_t)(__p0);
36792  return __ret;
36793 }
36794 #endif
36795 
36796 #ifdef __LITTLE_ENDIAN__
36797 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36798  float64x2_t __ret;
36799  __ret = (float64x2_t)(__p0);
36800  return __ret;
36801 }
36802 #else
36803 __ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36804  float64x2_t __ret;
36805  __ret = (float64x2_t)(__p0);
36806  return __ret;
36807 }
36808 #endif
36809 
36810 #ifdef __LITTLE_ENDIAN__
36811 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36812  float64x2_t __ret;
36813  __ret = (float64x2_t)(__p0);
36814  return __ret;
36815 }
36816 #else
36817 __ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36818  float64x2_t __ret;
36819  __ret = (float64x2_t)(__p0);
36820  return __ret;
36821 }
36822 #endif
36823 
36824 #ifdef __LITTLE_ENDIAN__
36825 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36826  float64x2_t __ret;
36827  __ret = (float64x2_t)(__p0);
36828  return __ret;
36829 }
36830 #else
36831 __ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36832  float64x2_t __ret;
36833  __ret = (float64x2_t)(__p0);
36834  return __ret;
36835 }
36836 #endif
36837 
36838 #ifdef __LITTLE_ENDIAN__
36839 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36840  float64x2_t __ret;
36841  __ret = (float64x2_t)(__p0);
36842  return __ret;
36843 }
36844 #else
36845 __ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36846  float64x2_t __ret;
36847  __ret = (float64x2_t)(__p0);
36848  return __ret;
36849 }
36850 #endif
36851 
36852 #ifdef __LITTLE_ENDIAN__
36853 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36854  float64x2_t __ret;
36855  __ret = (float64x2_t)(__p0);
36856  return __ret;
36857 }
36858 #else
36859 __ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36860  float64x2_t __ret;
36861  __ret = (float64x2_t)(__p0);
36862  return __ret;
36863 }
36864 #endif
36865 
36866 #ifdef __LITTLE_ENDIAN__
36867 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36868  float64x2_t __ret;
36869  __ret = (float64x2_t)(__p0);
36870  return __ret;
36871 }
36872 #else
36873 __ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36874  float64x2_t __ret;
36875  __ret = (float64x2_t)(__p0);
36876  return __ret;
36877 }
36878 #endif
36879 
36880 #ifdef __LITTLE_ENDIAN__
36881 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36882  float64x2_t __ret;
36883  __ret = (float64x2_t)(__p0);
36884  return __ret;
36885 }
36886 #else
36887 __ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36888  float64x2_t __ret;
36889  __ret = (float64x2_t)(__p0);
36890  return __ret;
36891 }
36892 #endif
36893 
36894 #ifdef __LITTLE_ENDIAN__
36895 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36896  float32x4_t __ret;
36897  __ret = (float32x4_t)(__p0);
36898  return __ret;
36899 }
36900 #else
36901 __ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36902  float32x4_t __ret;
36903  __ret = (float32x4_t)(__p0);
36904  return __ret;
36905 }
36906 #endif
36907 
36908 #ifdef __LITTLE_ENDIAN__
36909 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36910  float32x4_t __ret;
36911  __ret = (float32x4_t)(__p0);
36912  return __ret;
36913 }
36914 #else
36915 __ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36916  float32x4_t __ret;
36917  __ret = (float32x4_t)(__p0);
36918  return __ret;
36919 }
36920 #endif
36921 
36922 #ifdef __LITTLE_ENDIAN__
36923 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36924  float32x4_t __ret;
36925  __ret = (float32x4_t)(__p0);
36926  return __ret;
36927 }
36928 #else
36929 __ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36930  float32x4_t __ret;
36931  __ret = (float32x4_t)(__p0);
36932  return __ret;
36933 }
36934 #endif
36935 
36936 #ifdef __LITTLE_ENDIAN__
36937 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36938  float32x4_t __ret;
36939  __ret = (float32x4_t)(__p0);
36940  return __ret;
36941 }
36942 #else
36943 __ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36944  float32x4_t __ret;
36945  __ret = (float32x4_t)(__p0);
36946  return __ret;
36947 }
36948 #endif
36949 
36950 #ifdef __LITTLE_ENDIAN__
36951 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36952  float32x4_t __ret;
36953  __ret = (float32x4_t)(__p0);
36954  return __ret;
36955 }
36956 #else
36957 __ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36958  float32x4_t __ret;
36959  __ret = (float32x4_t)(__p0);
36960  return __ret;
36961 }
36962 #endif
36963 
36964 #ifdef __LITTLE_ENDIAN__
36965 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36966  float32x4_t __ret;
36967  __ret = (float32x4_t)(__p0);
36968  return __ret;
36969 }
36970 #else
36971 __ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36972  float32x4_t __ret;
36973  __ret = (float32x4_t)(__p0);
36974  return __ret;
36975 }
36976 #endif
36977 
36978 #ifdef __LITTLE_ENDIAN__
36979 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36980  float32x4_t __ret;
36981  __ret = (float32x4_t)(__p0);
36982  return __ret;
36983 }
36984 #else
36985 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36986  float32x4_t __ret;
36987  __ret = (float32x4_t)(__p0);
36988  return __ret;
36989 }
36990 #endif
36991 
36992 #ifdef __LITTLE_ENDIAN__
36993 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36994  float32x4_t __ret;
36995  __ret = (float32x4_t)(__p0);
36996  return __ret;
36997 }
36998 #else
36999 __ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
37000  float32x4_t __ret;
37001  __ret = (float32x4_t)(__p0);
37002  return __ret;
37003 }
37004 #endif
37005 
37006 #ifdef __LITTLE_ENDIAN__
37007 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
37008  float32x4_t __ret;
37009  __ret = (float32x4_t)(__p0);
37010  return __ret;
37011 }
37012 #else
37013 __ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
37014  float32x4_t __ret;
37015  __ret = (float32x4_t)(__p0);
37016  return __ret;
37017 }
37018 #endif
37019 
37020 #ifdef __LITTLE_ENDIAN__
37021 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
37022  float32x4_t __ret;
37023  __ret = (float32x4_t)(__p0);
37024  return __ret;
37025 }
37026 #else
37027 __ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
37028  float32x4_t __ret;
37029  __ret = (float32x4_t)(__p0);
37030  return __ret;
37031 }
37032 #endif
37033 
37034 #ifdef __LITTLE_ENDIAN__
37035 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
37036  float32x4_t __ret;
37037  __ret = (float32x4_t)(__p0);
37038  return __ret;
37039 }
37040 #else
37041 __ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
37042  float32x4_t __ret;
37043  __ret = (float32x4_t)(__p0);
37044  return __ret;
37045 }
37046 #endif
37047 
37048 #ifdef __LITTLE_ENDIAN__
37049 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37050  float32x4_t __ret;
37051  __ret = (float32x4_t)(__p0);
37052  return __ret;
37053 }
37054 #else
37055 __ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37056  float32x4_t __ret;
37057  __ret = (float32x4_t)(__p0);
37058  return __ret;
37059 }
37060 #endif
37061 
37062 #ifdef __LITTLE_ENDIAN__
37063 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37064  float32x4_t __ret;
37065  __ret = (float32x4_t)(__p0);
37066  return __ret;
37067 }
37068 #else
37069 __ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37070  float32x4_t __ret;
37071  __ret = (float32x4_t)(__p0);
37072  return __ret;
37073 }
37074 #endif
37075 
37076 #ifdef __LITTLE_ENDIAN__
37077 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37078  float32x4_t __ret;
37079  __ret = (float32x4_t)(__p0);
37080  return __ret;
37081 }
37082 #else
37083 __ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37084  float32x4_t __ret;
37085  __ret = (float32x4_t)(__p0);
37086  return __ret;
37087 }
37088 #endif
37089 
37090 #ifdef __LITTLE_ENDIAN__
37091 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37092  float16x8_t __ret;
37093  __ret = (float16x8_t)(__p0);
37094  return __ret;
37095 }
37096 #else
37097 __ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37098  float16x8_t __ret;
37099  __ret = (float16x8_t)(__p0);
37100  return __ret;
37101 }
37102 #endif
37103 
37104 #ifdef __LITTLE_ENDIAN__
37105 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37106  float16x8_t __ret;
37107  __ret = (float16x8_t)(__p0);
37108  return __ret;
37109 }
37110 #else
37111 __ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37112  float16x8_t __ret;
37113  __ret = (float16x8_t)(__p0);
37114  return __ret;
37115 }
37116 #endif
37117 
37118 #ifdef __LITTLE_ENDIAN__
37119 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37120  float16x8_t __ret;
37121  __ret = (float16x8_t)(__p0);
37122  return __ret;
37123 }
37124 #else
37125 __ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37126  float16x8_t __ret;
37127  __ret = (float16x8_t)(__p0);
37128  return __ret;
37129 }
37130 #endif
37131 
37132 #ifdef __LITTLE_ENDIAN__
37133 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37134  float16x8_t __ret;
37135  __ret = (float16x8_t)(__p0);
37136  return __ret;
37137 }
37138 #else
37139 __ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37140  float16x8_t __ret;
37141  __ret = (float16x8_t)(__p0);
37142  return __ret;
37143 }
37144 #endif
37145 
37146 #ifdef __LITTLE_ENDIAN__
37147 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37148  float16x8_t __ret;
37149  __ret = (float16x8_t)(__p0);
37150  return __ret;
37151 }
37152 #else
37153 __ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37154  float16x8_t __ret;
37155  __ret = (float16x8_t)(__p0);
37156  return __ret;
37157 }
37158 #endif
37159 
37160 #ifdef __LITTLE_ENDIAN__
37161 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37162  float16x8_t __ret;
37163  __ret = (float16x8_t)(__p0);
37164  return __ret;
37165 }
37166 #else
37167 __ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37168  float16x8_t __ret;
37169  __ret = (float16x8_t)(__p0);
37170  return __ret;
37171 }
37172 #endif
37173 
37174 #ifdef __LITTLE_ENDIAN__
37175 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37176  float16x8_t __ret;
37177  __ret = (float16x8_t)(__p0);
37178  return __ret;
37179 }
37180 #else
37181 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37182  float16x8_t __ret;
37183  __ret = (float16x8_t)(__p0);
37184  return __ret;
37185 }
37186 #endif
37187 
37188 #ifdef __LITTLE_ENDIAN__
37189 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37190  float16x8_t __ret;
37191  __ret = (float16x8_t)(__p0);
37192  return __ret;
37193 }
37194 #else
37195 __ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37196  float16x8_t __ret;
37197  __ret = (float16x8_t)(__p0);
37198  return __ret;
37199 }
37200 #endif
37201 
37202 #ifdef __LITTLE_ENDIAN__
37203 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37204  float16x8_t __ret;
37205  __ret = (float16x8_t)(__p0);
37206  return __ret;
37207 }
37208 #else
37209 __ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37210  float16x8_t __ret;
37211  __ret = (float16x8_t)(__p0);
37212  return __ret;
37213 }
37214 #endif
37215 
37216 #ifdef __LITTLE_ENDIAN__
37217 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37218  float16x8_t __ret;
37219  __ret = (float16x8_t)(__p0);
37220  return __ret;
37221 }
37222 #else
37223 __ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37224  float16x8_t __ret;
37225  __ret = (float16x8_t)(__p0);
37226  return __ret;
37227 }
37228 #endif
37229 
37230 #ifdef __LITTLE_ENDIAN__
37231 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37232  float16x8_t __ret;
37233  __ret = (float16x8_t)(__p0);
37234  return __ret;
37235 }
37236 #else
37237 __ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37238  float16x8_t __ret;
37239  __ret = (float16x8_t)(__p0);
37240  return __ret;
37241 }
37242 #endif
37243 
37244 #ifdef __LITTLE_ENDIAN__
37245 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37246  float16x8_t __ret;
37247  __ret = (float16x8_t)(__p0);
37248  return __ret;
37249 }
37250 #else
37251 __ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37252  float16x8_t __ret;
37253  __ret = (float16x8_t)(__p0);
37254  return __ret;
37255 }
37256 #endif
37257 
37258 #ifdef __LITTLE_ENDIAN__
37259 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37260  float16x8_t __ret;
37261  __ret = (float16x8_t)(__p0);
37262  return __ret;
37263 }
37264 #else
37265 __ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37266  float16x8_t __ret;
37267  __ret = (float16x8_t)(__p0);
37268  return __ret;
37269 }
37270 #endif
37271 
37272 #ifdef __LITTLE_ENDIAN__
37273 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37274  float16x8_t __ret;
37275  __ret = (float16x8_t)(__p0);
37276  return __ret;
37277 }
37278 #else
37279 __ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37280  float16x8_t __ret;
37281  __ret = (float16x8_t)(__p0);
37282  return __ret;
37283 }
37284 #endif
37285 
37286 #ifdef __LITTLE_ENDIAN__
37287 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37288  int32x4_t __ret;
37289  __ret = (int32x4_t)(__p0);
37290  return __ret;
37291 }
37292 #else
37293 __ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37294  int32x4_t __ret;
37295  __ret = (int32x4_t)(__p0);
37296  return __ret;
37297 }
37298 #endif
37299 
37300 #ifdef __LITTLE_ENDIAN__
37301 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37302  int32x4_t __ret;
37303  __ret = (int32x4_t)(__p0);
37304  return __ret;
37305 }
37306 #else
37307 __ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37308  int32x4_t __ret;
37309  __ret = (int32x4_t)(__p0);
37310  return __ret;
37311 }
37312 #endif
37313 
37314 #ifdef __LITTLE_ENDIAN__
37315 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37316  int32x4_t __ret;
37317  __ret = (int32x4_t)(__p0);
37318  return __ret;
37319 }
37320 #else
37321 __ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37322  int32x4_t __ret;
37323  __ret = (int32x4_t)(__p0);
37324  return __ret;
37325 }
37326 #endif
37327 
37328 #ifdef __LITTLE_ENDIAN__
37329 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37330  int32x4_t __ret;
37331  __ret = (int32x4_t)(__p0);
37332  return __ret;
37333 }
37334 #else
37335 __ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37336  int32x4_t __ret;
37337  __ret = (int32x4_t)(__p0);
37338  return __ret;
37339 }
37340 #endif
37341 
37342 #ifdef __LITTLE_ENDIAN__
37343 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37344  int32x4_t __ret;
37345  __ret = (int32x4_t)(__p0);
37346  return __ret;
37347 }
37348 #else
37349 __ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37350  int32x4_t __ret;
37351  __ret = (int32x4_t)(__p0);
37352  return __ret;
37353 }
37354 #endif
37355 
37356 #ifdef __LITTLE_ENDIAN__
37357 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37358  int32x4_t __ret;
37359  __ret = (int32x4_t)(__p0);
37360  return __ret;
37361 }
37362 #else
37363 __ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37364  int32x4_t __ret;
37365  __ret = (int32x4_t)(__p0);
37366  return __ret;
37367 }
37368 #endif
37369 
37370 #ifdef __LITTLE_ENDIAN__
37371 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37372  int32x4_t __ret;
37373  __ret = (int32x4_t)(__p0);
37374  return __ret;
37375 }
37376 #else
37377 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37378  int32x4_t __ret;
37379  __ret = (int32x4_t)(__p0);
37380  return __ret;
37381 }
37382 #endif
37383 
37384 #ifdef __LITTLE_ENDIAN__
37385 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37386  int32x4_t __ret;
37387  __ret = (int32x4_t)(__p0);
37388  return __ret;
37389 }
37390 #else
37391 __ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37392  int32x4_t __ret;
37393  __ret = (int32x4_t)(__p0);
37394  return __ret;
37395 }
37396 #endif
37397 
37398 #ifdef __LITTLE_ENDIAN__
37399 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37400  int32x4_t __ret;
37401  __ret = (int32x4_t)(__p0);
37402  return __ret;
37403 }
37404 #else
37405 __ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37406  int32x4_t __ret;
37407  __ret = (int32x4_t)(__p0);
37408  return __ret;
37409 }
37410 #endif
37411 
37412 #ifdef __LITTLE_ENDIAN__
37413 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37414  int32x4_t __ret;
37415  __ret = (int32x4_t)(__p0);
37416  return __ret;
37417 }
37418 #else
37419 __ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37420  int32x4_t __ret;
37421  __ret = (int32x4_t)(__p0);
37422  return __ret;
37423 }
37424 #endif
37425 
37426 #ifdef __LITTLE_ENDIAN__
37427 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37428  int32x4_t __ret;
37429  __ret = (int32x4_t)(__p0);
37430  return __ret;
37431 }
37432 #else
37433 __ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37434  int32x4_t __ret;
37435  __ret = (int32x4_t)(__p0);
37436  return __ret;
37437 }
37438 #endif
37439 
37440 #ifdef __LITTLE_ENDIAN__
37441 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37442  int32x4_t __ret;
37443  __ret = (int32x4_t)(__p0);
37444  return __ret;
37445 }
37446 #else
37447 __ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37448  int32x4_t __ret;
37449  __ret = (int32x4_t)(__p0);
37450  return __ret;
37451 }
37452 #endif
37453 
37454 #ifdef __LITTLE_ENDIAN__
37455 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37456  int32x4_t __ret;
37457  __ret = (int32x4_t)(__p0);
37458  return __ret;
37459 }
37460 #else
37461 __ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37462  int32x4_t __ret;
37463  __ret = (int32x4_t)(__p0);
37464  return __ret;
37465 }
37466 #endif
37467 
37468 #ifdef __LITTLE_ENDIAN__
37469 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37470  int32x4_t __ret;
37471  __ret = (int32x4_t)(__p0);
37472  return __ret;
37473 }
37474 #else
37475 __ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37476  int32x4_t __ret;
37477  __ret = (int32x4_t)(__p0);
37478  return __ret;
37479 }
37480 #endif
37481 
37482 #ifdef __LITTLE_ENDIAN__
37483 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37484  int64x2_t __ret;
37485  __ret = (int64x2_t)(__p0);
37486  return __ret;
37487 }
37488 #else
37489 __ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37490  int64x2_t __ret;
37491  __ret = (int64x2_t)(__p0);
37492  return __ret;
37493 }
37494 #endif
37495 
37496 #ifdef __LITTLE_ENDIAN__
37497 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37498  int64x2_t __ret;
37499  __ret = (int64x2_t)(__p0);
37500  return __ret;
37501 }
37502 #else
37503 __ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37504  int64x2_t __ret;
37505  __ret = (int64x2_t)(__p0);
37506  return __ret;
37507 }
37508 #endif
37509 
37510 #ifdef __LITTLE_ENDIAN__
37511 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37512  int64x2_t __ret;
37513  __ret = (int64x2_t)(__p0);
37514  return __ret;
37515 }
37516 #else
37517 __ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37518  int64x2_t __ret;
37519  __ret = (int64x2_t)(__p0);
37520  return __ret;
37521 }
37522 #endif
37523 
37524 #ifdef __LITTLE_ENDIAN__
37525 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37526  int64x2_t __ret;
37527  __ret = (int64x2_t)(__p0);
37528  return __ret;
37529 }
37530 #else
37531 __ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37532  int64x2_t __ret;
37533  __ret = (int64x2_t)(__p0);
37534  return __ret;
37535 }
37536 #endif
37537 
37538 #ifdef __LITTLE_ENDIAN__
37539 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37540  int64x2_t __ret;
37541  __ret = (int64x2_t)(__p0);
37542  return __ret;
37543 }
37544 #else
37545 __ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37546  int64x2_t __ret;
37547  __ret = (int64x2_t)(__p0);
37548  return __ret;
37549 }
37550 #endif
37551 
37552 #ifdef __LITTLE_ENDIAN__
37553 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37554  int64x2_t __ret;
37555  __ret = (int64x2_t)(__p0);
37556  return __ret;
37557 }
37558 #else
37559 __ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37560  int64x2_t __ret;
37561  __ret = (int64x2_t)(__p0);
37562  return __ret;
37563 }
37564 #endif
37565 
37566 #ifdef __LITTLE_ENDIAN__
37567 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37568  int64x2_t __ret;
37569  __ret = (int64x2_t)(__p0);
37570  return __ret;
37571 }
37572 #else
37573 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37574  int64x2_t __ret;
37575  __ret = (int64x2_t)(__p0);
37576  return __ret;
37577 }
37578 #endif
37579 
37580 #ifdef __LITTLE_ENDIAN__
37581 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37582  int64x2_t __ret;
37583  __ret = (int64x2_t)(__p0);
37584  return __ret;
37585 }
37586 #else
37587 __ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37588  int64x2_t __ret;
37589  __ret = (int64x2_t)(__p0);
37590  return __ret;
37591 }
37592 #endif
37593 
37594 #ifdef __LITTLE_ENDIAN__
37595 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37596  int64x2_t __ret;
37597  __ret = (int64x2_t)(__p0);
37598  return __ret;
37599 }
37600 #else
37601 __ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37602  int64x2_t __ret;
37603  __ret = (int64x2_t)(__p0);
37604  return __ret;
37605 }
37606 #endif
37607 
37608 #ifdef __LITTLE_ENDIAN__
37609 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37610  int64x2_t __ret;
37611  __ret = (int64x2_t)(__p0);
37612  return __ret;
37613 }
37614 #else
37615 __ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37616  int64x2_t __ret;
37617  __ret = (int64x2_t)(__p0);
37618  return __ret;
37619 }
37620 #endif
37621 
37622 #ifdef __LITTLE_ENDIAN__
37623 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37624  int64x2_t __ret;
37625  __ret = (int64x2_t)(__p0);
37626  return __ret;
37627 }
37628 #else
37629 __ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37630  int64x2_t __ret;
37631  __ret = (int64x2_t)(__p0);
37632  return __ret;
37633 }
37634 #endif
37635 
37636 #ifdef __LITTLE_ENDIAN__
37637 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37638  int64x2_t __ret;
37639  __ret = (int64x2_t)(__p0);
37640  return __ret;
37641 }
37642 #else
37643 __ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37644  int64x2_t __ret;
37645  __ret = (int64x2_t)(__p0);
37646  return __ret;
37647 }
37648 #endif
37649 
37650 #ifdef __LITTLE_ENDIAN__
37651 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37652  int64x2_t __ret;
37653  __ret = (int64x2_t)(__p0);
37654  return __ret;
37655 }
37656 #else
37657 __ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37658  int64x2_t __ret;
37659  __ret = (int64x2_t)(__p0);
37660  return __ret;
37661 }
37662 #endif
37663 
37664 #ifdef __LITTLE_ENDIAN__
37665 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37666  int64x2_t __ret;
37667  __ret = (int64x2_t)(__p0);
37668  return __ret;
37669 }
37670 #else
37671 __ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37672  int64x2_t __ret;
37673  __ret = (int64x2_t)(__p0);
37674  return __ret;
37675 }
37676 #endif
37677 
37678 #ifdef __LITTLE_ENDIAN__
37679 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37680  int16x8_t __ret;
37681  __ret = (int16x8_t)(__p0);
37682  return __ret;
37683 }
37684 #else
37685 __ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37686  int16x8_t __ret;
37687  __ret = (int16x8_t)(__p0);
37688  return __ret;
37689 }
37690 #endif
37691 
37692 #ifdef __LITTLE_ENDIAN__
37693 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37694  int16x8_t __ret;
37695  __ret = (int16x8_t)(__p0);
37696  return __ret;
37697 }
37698 #else
37699 __ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37700  int16x8_t __ret;
37701  __ret = (int16x8_t)(__p0);
37702  return __ret;
37703 }
37704 #endif
37705 
37706 #ifdef __LITTLE_ENDIAN__
37707 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37708  int16x8_t __ret;
37709  __ret = (int16x8_t)(__p0);
37710  return __ret;
37711 }
37712 #else
37713 __ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37714  int16x8_t __ret;
37715  __ret = (int16x8_t)(__p0);
37716  return __ret;
37717 }
37718 #endif
37719 
37720 #ifdef __LITTLE_ENDIAN__
37721 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37722  int16x8_t __ret;
37723  __ret = (int16x8_t)(__p0);
37724  return __ret;
37725 }
37726 #else
37727 __ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37728  int16x8_t __ret;
37729  __ret = (int16x8_t)(__p0);
37730  return __ret;
37731 }
37732 #endif
37733 
37734 #ifdef __LITTLE_ENDIAN__
37735 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37736  int16x8_t __ret;
37737  __ret = (int16x8_t)(__p0);
37738  return __ret;
37739 }
37740 #else
37741 __ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37742  int16x8_t __ret;
37743  __ret = (int16x8_t)(__p0);
37744  return __ret;
37745 }
37746 #endif
37747 
37748 #ifdef __LITTLE_ENDIAN__
37749 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37750  int16x8_t __ret;
37751  __ret = (int16x8_t)(__p0);
37752  return __ret;
37753 }
37754 #else
37755 __ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37756  int16x8_t __ret;
37757  __ret = (int16x8_t)(__p0);
37758  return __ret;
37759 }
37760 #endif
37761 
37762 #ifdef __LITTLE_ENDIAN__
37763 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37764  int16x8_t __ret;
37765  __ret = (int16x8_t)(__p0);
37766  return __ret;
37767 }
37768 #else
37769 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37770  int16x8_t __ret;
37771  __ret = (int16x8_t)(__p0);
37772  return __ret;
37773 }
37774 #endif
37775 
37776 #ifdef __LITTLE_ENDIAN__
37777 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37778  int16x8_t __ret;
37779  __ret = (int16x8_t)(__p0);
37780  return __ret;
37781 }
37782 #else
37783 __ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37784  int16x8_t __ret;
37785  __ret = (int16x8_t)(__p0);
37786  return __ret;
37787 }
37788 #endif
37789 
37790 #ifdef __LITTLE_ENDIAN__
37791 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37792  int16x8_t __ret;
37793  __ret = (int16x8_t)(__p0);
37794  return __ret;
37795 }
37796 #else
37797 __ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37798  int16x8_t __ret;
37799  __ret = (int16x8_t)(__p0);
37800  return __ret;
37801 }
37802 #endif
37803 
37804 #ifdef __LITTLE_ENDIAN__
37805 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37806  int16x8_t __ret;
37807  __ret = (int16x8_t)(__p0);
37808  return __ret;
37809 }
37810 #else
37811 __ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37812  int16x8_t __ret;
37813  __ret = (int16x8_t)(__p0);
37814  return __ret;
37815 }
37816 #endif
37817 
37818 #ifdef __LITTLE_ENDIAN__
37819 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37820  int16x8_t __ret;
37821  __ret = (int16x8_t)(__p0);
37822  return __ret;
37823 }
37824 #else
37825 __ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37826  int16x8_t __ret;
37827  __ret = (int16x8_t)(__p0);
37828  return __ret;
37829 }
37830 #endif
37831 
37832 #ifdef __LITTLE_ENDIAN__
37833 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37834  int16x8_t __ret;
37835  __ret = (int16x8_t)(__p0);
37836  return __ret;
37837 }
37838 #else
37839 __ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37840  int16x8_t __ret;
37841  __ret = (int16x8_t)(__p0);
37842  return __ret;
37843 }
37844 #endif
37845 
37846 #ifdef __LITTLE_ENDIAN__
37847 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37848  int16x8_t __ret;
37849  __ret = (int16x8_t)(__p0);
37850  return __ret;
37851 }
37852 #else
37853 __ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37854  int16x8_t __ret;
37855  __ret = (int16x8_t)(__p0);
37856  return __ret;
37857 }
37858 #endif
37859 
37860 #ifdef __LITTLE_ENDIAN__
37861 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37862  int16x8_t __ret;
37863  __ret = (int16x8_t)(__p0);
37864  return __ret;
37865 }
37866 #else
37867 __ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37868  int16x8_t __ret;
37869  __ret = (int16x8_t)(__p0);
37870  return __ret;
37871 }
37872 #endif
37873 
37874 #ifdef __LITTLE_ENDIAN__
37875 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37876  uint8x8_t __ret;
37877  __ret = (uint8x8_t)(__p0);
37878  return __ret;
37879 }
37880 #else
37881 __ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37882  uint8x8_t __ret;
37883  __ret = (uint8x8_t)(__p0);
37884  return __ret;
37885 }
37886 #endif
37887 
37888 #ifdef __LITTLE_ENDIAN__
37889 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37890  uint8x8_t __ret;
37891  __ret = (uint8x8_t)(__p0);
37892  return __ret;
37893 }
37894 #else
37895 __ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37896  uint8x8_t __ret;
37897  __ret = (uint8x8_t)(__p0);
37898  return __ret;
37899 }
37900 #endif
37901 
37902 #ifdef __LITTLE_ENDIAN__
37903 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37904  uint8x8_t __ret;
37905  __ret = (uint8x8_t)(__p0);
37906  return __ret;
37907 }
37908 #else
37909 __ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37910  uint8x8_t __ret;
37911  __ret = (uint8x8_t)(__p0);
37912  return __ret;
37913 }
37914 #endif
37915 
37916 #ifdef __LITTLE_ENDIAN__
37917 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37918  uint8x8_t __ret;
37919  __ret = (uint8x8_t)(__p0);
37920  return __ret;
37921 }
37922 #else
37923 __ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37924  uint8x8_t __ret;
37925  __ret = (uint8x8_t)(__p0);
37926  return __ret;
37927 }
37928 #endif
37929 
37930 #ifdef __LITTLE_ENDIAN__
37931 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37932  uint8x8_t __ret;
37933  __ret = (uint8x8_t)(__p0);
37934  return __ret;
37935 }
37936 #else
37937 __ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37938  uint8x8_t __ret;
37939  __ret = (uint8x8_t)(__p0);
37940  return __ret;
37941 }
37942 #endif
37943 
37944 #ifdef __LITTLE_ENDIAN__
37945 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37946  uint8x8_t __ret;
37947  __ret = (uint8x8_t)(__p0);
37948  return __ret;
37949 }
37950 #else
37951 __ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37952  uint8x8_t __ret;
37953  __ret = (uint8x8_t)(__p0);
37954  return __ret;
37955 }
37956 #endif
37957 
37958 #ifdef __LITTLE_ENDIAN__
37959 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37960  uint8x8_t __ret;
37961  __ret = (uint8x8_t)(__p0);
37962  return __ret;
37963 }
37964 #else
37965 __ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37966  uint8x8_t __ret;
37967  __ret = (uint8x8_t)(__p0);
37968  return __ret;
37969 }
37970 #endif
37971 
37972 #ifdef __LITTLE_ENDIAN__
37973 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37974  uint8x8_t __ret;
37975  __ret = (uint8x8_t)(__p0);
37976  return __ret;
37977 }
37978 #else
37979 __ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37980  uint8x8_t __ret;
37981  __ret = (uint8x8_t)(__p0);
37982  return __ret;
37983 }
37984 #endif
37985 
37986 #ifdef __LITTLE_ENDIAN__
37987 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37988  uint8x8_t __ret;
37989  __ret = (uint8x8_t)(__p0);
37990  return __ret;
37991 }
37992 #else
37993 __ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37994  uint8x8_t __ret;
37995  __ret = (uint8x8_t)(__p0);
37996  return __ret;
37997 }
37998 #endif
37999 
38000 #ifdef __LITTLE_ENDIAN__
38001 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
38002  uint8x8_t __ret;
38003  __ret = (uint8x8_t)(__p0);
38004  return __ret;
38005 }
38006 #else
38007 __ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
38008  uint8x8_t __ret;
38009  __ret = (uint8x8_t)(__p0);
38010  return __ret;
38011 }
38012 #endif
38013 
38014 #ifdef __LITTLE_ENDIAN__
38015 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
38016  uint8x8_t __ret;
38017  __ret = (uint8x8_t)(__p0);
38018  return __ret;
38019 }
38020 #else
38021 __ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
38022  uint8x8_t __ret;
38023  __ret = (uint8x8_t)(__p0);
38024  return __ret;
38025 }
38026 #endif
38027 
38028 #ifdef __LITTLE_ENDIAN__
38029 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
38030  uint8x8_t __ret;
38031  __ret = (uint8x8_t)(__p0);
38032  return __ret;
38033 }
38034 #else
38035 __ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
38036  uint8x8_t __ret;
38037  __ret = (uint8x8_t)(__p0);
38038  return __ret;
38039 }
38040 #endif
38041 
38042 #ifdef __LITTLE_ENDIAN__
38043 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38044  uint8x8_t __ret;
38045  __ret = (uint8x8_t)(__p0);
38046  return __ret;
38047 }
38048 #else
38049 __ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38050  uint8x8_t __ret;
38051  __ret = (uint8x8_t)(__p0);
38052  return __ret;
38053 }
38054 #endif
38055 
38056 #ifdef __LITTLE_ENDIAN__
38057 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38058  uint32x2_t __ret;
38059  __ret = (uint32x2_t)(__p0);
38060  return __ret;
38061 }
38062 #else
38063 __ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38064  uint32x2_t __ret;
38065  __ret = (uint32x2_t)(__p0);
38066  return __ret;
38067 }
38068 #endif
38069 
38070 #ifdef __LITTLE_ENDIAN__
38071 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38072  uint32x2_t __ret;
38073  __ret = (uint32x2_t)(__p0);
38074  return __ret;
38075 }
38076 #else
38077 __ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38078  uint32x2_t __ret;
38079  __ret = (uint32x2_t)(__p0);
38080  return __ret;
38081 }
38082 #endif
38083 
38084 #ifdef __LITTLE_ENDIAN__
38085 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38086  uint32x2_t __ret;
38087  __ret = (uint32x2_t)(__p0);
38088  return __ret;
38089 }
38090 #else
38091 __ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38092  uint32x2_t __ret;
38093  __ret = (uint32x2_t)(__p0);
38094  return __ret;
38095 }
38096 #endif
38097 
38098 #ifdef __LITTLE_ENDIAN__
38099 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38100  uint32x2_t __ret;
38101  __ret = (uint32x2_t)(__p0);
38102  return __ret;
38103 }
38104 #else
38105 __ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38106  uint32x2_t __ret;
38107  __ret = (uint32x2_t)(__p0);
38108  return __ret;
38109 }
38110 #endif
38111 
38112 #ifdef __LITTLE_ENDIAN__
38113 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38114  uint32x2_t __ret;
38115  __ret = (uint32x2_t)(__p0);
38116  return __ret;
38117 }
38118 #else
38119 __ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38120  uint32x2_t __ret;
38121  __ret = (uint32x2_t)(__p0);
38122  return __ret;
38123 }
38124 #endif
38125 
38126 #ifdef __LITTLE_ENDIAN__
38127 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38128  uint32x2_t __ret;
38129  __ret = (uint32x2_t)(__p0);
38130  return __ret;
38131 }
38132 #else
38133 __ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38134  uint32x2_t __ret;
38135  __ret = (uint32x2_t)(__p0);
38136  return __ret;
38137 }
38138 #endif
38139 
38140 #ifdef __LITTLE_ENDIAN__
38141 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38142  uint32x2_t __ret;
38143  __ret = (uint32x2_t)(__p0);
38144  return __ret;
38145 }
38146 #else
38147 __ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38148  uint32x2_t __ret;
38149  __ret = (uint32x2_t)(__p0);
38150  return __ret;
38151 }
38152 #endif
38153 
38154 #ifdef __LITTLE_ENDIAN__
38155 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38156  uint32x2_t __ret;
38157  __ret = (uint32x2_t)(__p0);
38158  return __ret;
38159 }
38160 #else
38161 __ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38162  uint32x2_t __ret;
38163  __ret = (uint32x2_t)(__p0);
38164  return __ret;
38165 }
38166 #endif
38167 
38168 #ifdef __LITTLE_ENDIAN__
38169 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38170  uint32x2_t __ret;
38171  __ret = (uint32x2_t)(__p0);
38172  return __ret;
38173 }
38174 #else
38175 __ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38176  uint32x2_t __ret;
38177  __ret = (uint32x2_t)(__p0);
38178  return __ret;
38179 }
38180 #endif
38181 
38182 #ifdef __LITTLE_ENDIAN__
38183 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38184  uint32x2_t __ret;
38185  __ret = (uint32x2_t)(__p0);
38186  return __ret;
38187 }
38188 #else
38189 __ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38190  uint32x2_t __ret;
38191  __ret = (uint32x2_t)(__p0);
38192  return __ret;
38193 }
38194 #endif
38195 
38196 #ifdef __LITTLE_ENDIAN__
38197 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38198  uint32x2_t __ret;
38199  __ret = (uint32x2_t)(__p0);
38200  return __ret;
38201 }
38202 #else
38203 __ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38204  uint32x2_t __ret;
38205  __ret = (uint32x2_t)(__p0);
38206  return __ret;
38207 }
38208 #endif
38209 
38210 #ifdef __LITTLE_ENDIAN__
38211 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38212  uint32x2_t __ret;
38213  __ret = (uint32x2_t)(__p0);
38214  return __ret;
38215 }
38216 #else
38217 __ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38218  uint32x2_t __ret;
38219  __ret = (uint32x2_t)(__p0);
38220  return __ret;
38221 }
38222 #endif
38223 
38224 #ifdef __LITTLE_ENDIAN__
38225 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38226  uint32x2_t __ret;
38227  __ret = (uint32x2_t)(__p0);
38228  return __ret;
38229 }
38230 #else
38231 __ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38232  uint32x2_t __ret;
38233  __ret = (uint32x2_t)(__p0);
38234  return __ret;
38235 }
38236 #endif
38237 
38238 #ifdef __LITTLE_ENDIAN__
38239 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38240  uint64x1_t __ret;
38241  __ret = (uint64x1_t)(__p0);
38242  return __ret;
38243 }
38244 #else
38245 __ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38246  uint64x1_t __ret;
38247  __ret = (uint64x1_t)(__p0);
38248  return __ret;
38249 }
38250 #endif
38251 
38252 #ifdef __LITTLE_ENDIAN__
38253 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38254  uint64x1_t __ret;
38255  __ret = (uint64x1_t)(__p0);
38256  return __ret;
38257 }
38258 #else
38259 __ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38260  uint64x1_t __ret;
38261  __ret = (uint64x1_t)(__p0);
38262  return __ret;
38263 }
38264 #endif
38265 
38266 #ifdef __LITTLE_ENDIAN__
38267 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38268  uint64x1_t __ret;
38269  __ret = (uint64x1_t)(__p0);
38270  return __ret;
38271 }
38272 #else
38273 __ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38274  uint64x1_t __ret;
38275  __ret = (uint64x1_t)(__p0);
38276  return __ret;
38277 }
38278 #endif
38279 
38280 #ifdef __LITTLE_ENDIAN__
38281 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38282  uint64x1_t __ret;
38283  __ret = (uint64x1_t)(__p0);
38284  return __ret;
38285 }
38286 #else
38287 __ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38288  uint64x1_t __ret;
38289  __ret = (uint64x1_t)(__p0);
38290  return __ret;
38291 }
38292 #endif
38293 
38294 #ifdef __LITTLE_ENDIAN__
38295 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38296  uint64x1_t __ret;
38297  __ret = (uint64x1_t)(__p0);
38298  return __ret;
38299 }
38300 #else
38301 __ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38302  uint64x1_t __ret;
38303  __ret = (uint64x1_t)(__p0);
38304  return __ret;
38305 }
38306 #endif
38307 
38308 #ifdef __LITTLE_ENDIAN__
38309 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38310  uint64x1_t __ret;
38311  __ret = (uint64x1_t)(__p0);
38312  return __ret;
38313 }
38314 #else
38315 __ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38316  uint64x1_t __ret;
38317  __ret = (uint64x1_t)(__p0);
38318  return __ret;
38319 }
38320 #endif
38321 
38322 #ifdef __LITTLE_ENDIAN__
38323 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38324  uint64x1_t __ret;
38325  __ret = (uint64x1_t)(__p0);
38326  return __ret;
38327 }
38328 #else
38329 __ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38330  uint64x1_t __ret;
38331  __ret = (uint64x1_t)(__p0);
38332  return __ret;
38333 }
38334 #endif
38335 
38336 #ifdef __LITTLE_ENDIAN__
38337 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38338  uint64x1_t __ret;
38339  __ret = (uint64x1_t)(__p0);
38340  return __ret;
38341 }
38342 #else
38343 __ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38344  uint64x1_t __ret;
38345  __ret = (uint64x1_t)(__p0);
38346  return __ret;
38347 }
38348 #endif
38349 
38350 #ifdef __LITTLE_ENDIAN__
38351 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38352  uint64x1_t __ret;
38353  __ret = (uint64x1_t)(__p0);
38354  return __ret;
38355 }
38356 #else
38357 __ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38358  uint64x1_t __ret;
38359  __ret = (uint64x1_t)(__p0);
38360  return __ret;
38361 }
38362 #endif
38363 
38364 #ifdef __LITTLE_ENDIAN__
38365 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38366  uint64x1_t __ret;
38367  __ret = (uint64x1_t)(__p0);
38368  return __ret;
38369 }
38370 #else
38371 __ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38372  uint64x1_t __ret;
38373  __ret = (uint64x1_t)(__p0);
38374  return __ret;
38375 }
38376 #endif
38377 
38378 #ifdef __LITTLE_ENDIAN__
38379 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38380  uint64x1_t __ret;
38381  __ret = (uint64x1_t)(__p0);
38382  return __ret;
38383 }
38384 #else
38385 __ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38386  uint64x1_t __ret;
38387  __ret = (uint64x1_t)(__p0);
38388  return __ret;
38389 }
38390 #endif
38391 
38392 #ifdef __LITTLE_ENDIAN__
38393 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38394  uint64x1_t __ret;
38395  __ret = (uint64x1_t)(__p0);
38396  return __ret;
38397 }
38398 #else
38399 __ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38400  uint64x1_t __ret;
38401  __ret = (uint64x1_t)(__p0);
38402  return __ret;
38403 }
38404 #endif
38405 
38406 #ifdef __LITTLE_ENDIAN__
38407 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38408  uint64x1_t __ret;
38409  __ret = (uint64x1_t)(__p0);
38410  return __ret;
38411 }
38412 #else
38413 __ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38414  uint64x1_t __ret;
38415  __ret = (uint64x1_t)(__p0);
38416  return __ret;
38417 }
38418 #endif
38419 
38420 #ifdef __LITTLE_ENDIAN__
38421 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38422  uint16x4_t __ret;
38423  __ret = (uint16x4_t)(__p0);
38424  return __ret;
38425 }
38426 #else
38427 __ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38428  uint16x4_t __ret;
38429  __ret = (uint16x4_t)(__p0);
38430  return __ret;
38431 }
38432 #endif
38433 
38434 #ifdef __LITTLE_ENDIAN__
38435 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38436  uint16x4_t __ret;
38437  __ret = (uint16x4_t)(__p0);
38438  return __ret;
38439 }
38440 #else
38441 __ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38442  uint16x4_t __ret;
38443  __ret = (uint16x4_t)(__p0);
38444  return __ret;
38445 }
38446 #endif
38447 
38448 #ifdef __LITTLE_ENDIAN__
38449 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38450  uint16x4_t __ret;
38451  __ret = (uint16x4_t)(__p0);
38452  return __ret;
38453 }
38454 #else
38455 __ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38456  uint16x4_t __ret;
38457  __ret = (uint16x4_t)(__p0);
38458  return __ret;
38459 }
38460 #endif
38461 
38462 #ifdef __LITTLE_ENDIAN__
38463 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38464  uint16x4_t __ret;
38465  __ret = (uint16x4_t)(__p0);
38466  return __ret;
38467 }
38468 #else
38469 __ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38470  uint16x4_t __ret;
38471  __ret = (uint16x4_t)(__p0);
38472  return __ret;
38473 }
38474 #endif
38475 
38476 #ifdef __LITTLE_ENDIAN__
38477 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38478  uint16x4_t __ret;
38479  __ret = (uint16x4_t)(__p0);
38480  return __ret;
38481 }
38482 #else
38483 __ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38484  uint16x4_t __ret;
38485  __ret = (uint16x4_t)(__p0);
38486  return __ret;
38487 }
38488 #endif
38489 
38490 #ifdef __LITTLE_ENDIAN__
38491 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38492  uint16x4_t __ret;
38493  __ret = (uint16x4_t)(__p0);
38494  return __ret;
38495 }
38496 #else
38497 __ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38498  uint16x4_t __ret;
38499  __ret = (uint16x4_t)(__p0);
38500  return __ret;
38501 }
38502 #endif
38503 
38504 #ifdef __LITTLE_ENDIAN__
38505 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38506  uint16x4_t __ret;
38507  __ret = (uint16x4_t)(__p0);
38508  return __ret;
38509 }
38510 #else
38511 __ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38512  uint16x4_t __ret;
38513  __ret = (uint16x4_t)(__p0);
38514  return __ret;
38515 }
38516 #endif
38517 
38518 #ifdef __LITTLE_ENDIAN__
38519 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38520  uint16x4_t __ret;
38521  __ret = (uint16x4_t)(__p0);
38522  return __ret;
38523 }
38524 #else
38525 __ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38526  uint16x4_t __ret;
38527  __ret = (uint16x4_t)(__p0);
38528  return __ret;
38529 }
38530 #endif
38531 
38532 #ifdef __LITTLE_ENDIAN__
38533 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38534  uint16x4_t __ret;
38535  __ret = (uint16x4_t)(__p0);
38536  return __ret;
38537 }
38538 #else
38539 __ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38540  uint16x4_t __ret;
38541  __ret = (uint16x4_t)(__p0);
38542  return __ret;
38543 }
38544 #endif
38545 
38546 #ifdef __LITTLE_ENDIAN__
38547 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38548  uint16x4_t __ret;
38549  __ret = (uint16x4_t)(__p0);
38550  return __ret;
38551 }
38552 #else
38553 __ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38554  uint16x4_t __ret;
38555  __ret = (uint16x4_t)(__p0);
38556  return __ret;
38557 }
38558 #endif
38559 
38560 #ifdef __LITTLE_ENDIAN__
38561 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38562  uint16x4_t __ret;
38563  __ret = (uint16x4_t)(__p0);
38564  return __ret;
38565 }
38566 #else
38567 __ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38568  uint16x4_t __ret;
38569  __ret = (uint16x4_t)(__p0);
38570  return __ret;
38571 }
38572 #endif
38573 
38574 #ifdef __LITTLE_ENDIAN__
38575 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38576  uint16x4_t __ret;
38577  __ret = (uint16x4_t)(__p0);
38578  return __ret;
38579 }
38580 #else
38581 __ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38582  uint16x4_t __ret;
38583  __ret = (uint16x4_t)(__p0);
38584  return __ret;
38585 }
38586 #endif
38587 
38588 #ifdef __LITTLE_ENDIAN__
38589 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38590  uint16x4_t __ret;
38591  __ret = (uint16x4_t)(__p0);
38592  return __ret;
38593 }
38594 #else
38595 __ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38596  uint16x4_t __ret;
38597  __ret = (uint16x4_t)(__p0);
38598  return __ret;
38599 }
38600 #endif
38601 
38602 #ifdef __LITTLE_ENDIAN__
38603 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38604  int8x8_t __ret;
38605  __ret = (int8x8_t)(__p0);
38606  return __ret;
38607 }
38608 #else
38609 __ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38610  int8x8_t __ret;
38611  __ret = (int8x8_t)(__p0);
38612  return __ret;
38613 }
38614 #endif
38615 
38616 #ifdef __LITTLE_ENDIAN__
38617 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38618  int8x8_t __ret;
38619  __ret = (int8x8_t)(__p0);
38620  return __ret;
38621 }
38622 #else
38623 __ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38624  int8x8_t __ret;
38625  __ret = (int8x8_t)(__p0);
38626  return __ret;
38627 }
38628 #endif
38629 
38630 #ifdef __LITTLE_ENDIAN__
38631 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38632  int8x8_t __ret;
38633  __ret = (int8x8_t)(__p0);
38634  return __ret;
38635 }
38636 #else
38637 __ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38638  int8x8_t __ret;
38639  __ret = (int8x8_t)(__p0);
38640  return __ret;
38641 }
38642 #endif
38643 
38644 #ifdef __LITTLE_ENDIAN__
38645 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38646  int8x8_t __ret;
38647  __ret = (int8x8_t)(__p0);
38648  return __ret;
38649 }
38650 #else
38651 __ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38652  int8x8_t __ret;
38653  __ret = (int8x8_t)(__p0);
38654  return __ret;
38655 }
38656 #endif
38657 
38658 #ifdef __LITTLE_ENDIAN__
38659 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38660  int8x8_t __ret;
38661  __ret = (int8x8_t)(__p0);
38662  return __ret;
38663 }
38664 #else
38665 __ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38666  int8x8_t __ret;
38667  __ret = (int8x8_t)(__p0);
38668  return __ret;
38669 }
38670 #endif
38671 
38672 #ifdef __LITTLE_ENDIAN__
38673 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38674  int8x8_t __ret;
38675  __ret = (int8x8_t)(__p0);
38676  return __ret;
38677 }
38678 #else
38679 __ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38680  int8x8_t __ret;
38681  __ret = (int8x8_t)(__p0);
38682  return __ret;
38683 }
38684 #endif
38685 
38686 #ifdef __LITTLE_ENDIAN__
38687 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38688  int8x8_t __ret;
38689  __ret = (int8x8_t)(__p0);
38690  return __ret;
38691 }
38692 #else
38693 __ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38694  int8x8_t __ret;
38695  __ret = (int8x8_t)(__p0);
38696  return __ret;
38697 }
38698 #endif
38699 
38700 #ifdef __LITTLE_ENDIAN__
38701 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38702  int8x8_t __ret;
38703  __ret = (int8x8_t)(__p0);
38704  return __ret;
38705 }
38706 #else
38707 __ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38708  int8x8_t __ret;
38709  __ret = (int8x8_t)(__p0);
38710  return __ret;
38711 }
38712 #endif
38713 
38714 #ifdef __LITTLE_ENDIAN__
38715 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38716  int8x8_t __ret;
38717  __ret = (int8x8_t)(__p0);
38718  return __ret;
38719 }
38720 #else
38721 __ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38722  int8x8_t __ret;
38723  __ret = (int8x8_t)(__p0);
38724  return __ret;
38725 }
38726 #endif
38727 
38728 #ifdef __LITTLE_ENDIAN__
38729 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38730  int8x8_t __ret;
38731  __ret = (int8x8_t)(__p0);
38732  return __ret;
38733 }
38734 #else
38735 __ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38736  int8x8_t __ret;
38737  __ret = (int8x8_t)(__p0);
38738  return __ret;
38739 }
38740 #endif
38741 
38742 #ifdef __LITTLE_ENDIAN__
38743 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38744  int8x8_t __ret;
38745  __ret = (int8x8_t)(__p0);
38746  return __ret;
38747 }
38748 #else
38749 __ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38750  int8x8_t __ret;
38751  __ret = (int8x8_t)(__p0);
38752  return __ret;
38753 }
38754 #endif
38755 
38756 #ifdef __LITTLE_ENDIAN__
38757 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38758  int8x8_t __ret;
38759  __ret = (int8x8_t)(__p0);
38760  return __ret;
38761 }
38762 #else
38763 __ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38764  int8x8_t __ret;
38765  __ret = (int8x8_t)(__p0);
38766  return __ret;
38767 }
38768 #endif
38769 
38770 #ifdef __LITTLE_ENDIAN__
38771 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38772  int8x8_t __ret;
38773  __ret = (int8x8_t)(__p0);
38774  return __ret;
38775 }
38776 #else
38777 __ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38778  int8x8_t __ret;
38779  __ret = (int8x8_t)(__p0);
38780  return __ret;
38781 }
38782 #endif
38783 
38784 #ifdef __LITTLE_ENDIAN__
38785 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38786  float64x1_t __ret;
38787  __ret = (float64x1_t)(__p0);
38788  return __ret;
38789 }
38790 #else
38791 __ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38792  float64x1_t __ret;
38793  __ret = (float64x1_t)(__p0);
38794  return __ret;
38795 }
38796 #endif
38797 
38798 #ifdef __LITTLE_ENDIAN__
38799 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38800  float64x1_t __ret;
38801  __ret = (float64x1_t)(__p0);
38802  return __ret;
38803 }
38804 #else
38805 __ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38806  float64x1_t __ret;
38807  __ret = (float64x1_t)(__p0);
38808  return __ret;
38809 }
38810 #endif
38811 
38812 #ifdef __LITTLE_ENDIAN__
38813 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38814  float64x1_t __ret;
38815  __ret = (float64x1_t)(__p0);
38816  return __ret;
38817 }
38818 #else
38819 __ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38820  float64x1_t __ret;
38821  __ret = (float64x1_t)(__p0);
38822  return __ret;
38823 }
38824 #endif
38825 
38826 #ifdef __LITTLE_ENDIAN__
38827 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38828  float64x1_t __ret;
38829  __ret = (float64x1_t)(__p0);
38830  return __ret;
38831 }
38832 #else
38833 __ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38834  float64x1_t __ret;
38835  __ret = (float64x1_t)(__p0);
38836  return __ret;
38837 }
38838 #endif
38839 
38840 #ifdef __LITTLE_ENDIAN__
38841 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38842  float64x1_t __ret;
38843  __ret = (float64x1_t)(__p0);
38844  return __ret;
38845 }
38846 #else
38847 __ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38848  float64x1_t __ret;
38849  __ret = (float64x1_t)(__p0);
38850  return __ret;
38851 }
38852 #endif
38853 
38854 #ifdef __LITTLE_ENDIAN__
38855 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38856  float64x1_t __ret;
38857  __ret = (float64x1_t)(__p0);
38858  return __ret;
38859 }
38860 #else
38861 __ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38862  float64x1_t __ret;
38863  __ret = (float64x1_t)(__p0);
38864  return __ret;
38865 }
38866 #endif
38867 
38868 #ifdef __LITTLE_ENDIAN__
38869 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38870  float64x1_t __ret;
38871  __ret = (float64x1_t)(__p0);
38872  return __ret;
38873 }
38874 #else
38875 __ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38876  float64x1_t __ret;
38877  __ret = (float64x1_t)(__p0);
38878  return __ret;
38879 }
38880 #endif
38881 
38882 #ifdef __LITTLE_ENDIAN__
38883 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38884  float64x1_t __ret;
38885  __ret = (float64x1_t)(__p0);
38886  return __ret;
38887 }
38888 #else
38889 __ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38890  float64x1_t __ret;
38891  __ret = (float64x1_t)(__p0);
38892  return __ret;
38893 }
38894 #endif
38895 
38896 #ifdef __LITTLE_ENDIAN__
38897 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38898  float64x1_t __ret;
38899  __ret = (float64x1_t)(__p0);
38900  return __ret;
38901 }
38902 #else
38903 __ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38904  float64x1_t __ret;
38905  __ret = (float64x1_t)(__p0);
38906  return __ret;
38907 }
38908 #endif
38909 
38910 #ifdef __LITTLE_ENDIAN__
38911 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38912  float64x1_t __ret;
38913  __ret = (float64x1_t)(__p0);
38914  return __ret;
38915 }
38916 #else
38917 __ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38918  float64x1_t __ret;
38919  __ret = (float64x1_t)(__p0);
38920  return __ret;
38921 }
38922 #endif
38923 
38924 #ifdef __LITTLE_ENDIAN__
38925 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38926  float64x1_t __ret;
38927  __ret = (float64x1_t)(__p0);
38928  return __ret;
38929 }
38930 #else
38931 __ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38932  float64x1_t __ret;
38933  __ret = (float64x1_t)(__p0);
38934  return __ret;
38935 }
38936 #endif
38937 
38938 #ifdef __LITTLE_ENDIAN__
38939 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38940  float64x1_t __ret;
38941  __ret = (float64x1_t)(__p0);
38942  return __ret;
38943 }
38944 #else
38945 __ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38946  float64x1_t __ret;
38947  __ret = (float64x1_t)(__p0);
38948  return __ret;
38949 }
38950 #endif
38951 
38952 #ifdef __LITTLE_ENDIAN__
38953 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38954  float64x1_t __ret;
38955  __ret = (float64x1_t)(__p0);
38956  return __ret;
38957 }
38958 #else
38959 __ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38960  float64x1_t __ret;
38961  __ret = (float64x1_t)(__p0);
38962  return __ret;
38963 }
38964 #endif
38965 
38966 #ifdef __LITTLE_ENDIAN__
38967 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38968  float32x2_t __ret;
38969  __ret = (float32x2_t)(__p0);
38970  return __ret;
38971 }
38972 #else
38973 __ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38974  float32x2_t __ret;
38975  __ret = (float32x2_t)(__p0);
38976  return __ret;
38977 }
38978 #endif
38979 
38980 #ifdef __LITTLE_ENDIAN__
38981 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38982  float32x2_t __ret;
38983  __ret = (float32x2_t)(__p0);
38984  return __ret;
38985 }
38986 #else
38987 __ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38988  float32x2_t __ret;
38989  __ret = (float32x2_t)(__p0);
38990  return __ret;
38991 }
38992 #endif
38993 
38994 #ifdef __LITTLE_ENDIAN__
38995 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38996  float32x2_t __ret;
38997  __ret = (float32x2_t)(__p0);
38998  return __ret;
38999 }
39000 #else
39001 __ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
39002  float32x2_t __ret;
39003  __ret = (float32x2_t)(__p0);
39004  return __ret;
39005 }
39006 #endif
39007 
39008 #ifdef __LITTLE_ENDIAN__
39009 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
39010  float32x2_t __ret;
39011  __ret = (float32x2_t)(__p0);
39012  return __ret;
39013 }
39014 #else
39015 __ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
39016  float32x2_t __ret;
39017  __ret = (float32x2_t)(__p0);
39018  return __ret;
39019 }
39020 #endif
39021 
39022 #ifdef __LITTLE_ENDIAN__
39023 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
39024  float32x2_t __ret;
39025  __ret = (float32x2_t)(__p0);
39026  return __ret;
39027 }
39028 #else
39029 __ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
39030  float32x2_t __ret;
39031  __ret = (float32x2_t)(__p0);
39032  return __ret;
39033 }
39034 #endif
39035 
39036 #ifdef __LITTLE_ENDIAN__
39037 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
39038  float32x2_t __ret;
39039  __ret = (float32x2_t)(__p0);
39040  return __ret;
39041 }
39042 #else
39043 __ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
39044  float32x2_t __ret;
39045  __ret = (float32x2_t)(__p0);
39046  return __ret;
39047 }
39048 #endif
39049 
39050 #ifdef __LITTLE_ENDIAN__
39051 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39052  float32x2_t __ret;
39053  __ret = (float32x2_t)(__p0);
39054  return __ret;
39055 }
39056 #else
39057 __ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39058  float32x2_t __ret;
39059  __ret = (float32x2_t)(__p0);
39060  return __ret;
39061 }
39062 #endif
39063 
39064 #ifdef __LITTLE_ENDIAN__
39065 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39066  float32x2_t __ret;
39067  __ret = (float32x2_t)(__p0);
39068  return __ret;
39069 }
39070 #else
39071 __ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39072  float32x2_t __ret;
39073  __ret = (float32x2_t)(__p0);
39074  return __ret;
39075 }
39076 #endif
39077 
39078 #ifdef __LITTLE_ENDIAN__
39079 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39080  float32x2_t __ret;
39081  __ret = (float32x2_t)(__p0);
39082  return __ret;
39083 }
39084 #else
39085 __ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39086  float32x2_t __ret;
39087  __ret = (float32x2_t)(__p0);
39088  return __ret;
39089 }
39090 #endif
39091 
39092 #ifdef __LITTLE_ENDIAN__
39093 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39094  float32x2_t __ret;
39095  __ret = (float32x2_t)(__p0);
39096  return __ret;
39097 }
39098 #else
39099 __ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39100  float32x2_t __ret;
39101  __ret = (float32x2_t)(__p0);
39102  return __ret;
39103 }
39104 #endif
39105 
39106 #ifdef __LITTLE_ENDIAN__
39107 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39108  float32x2_t __ret;
39109  __ret = (float32x2_t)(__p0);
39110  return __ret;
39111 }
39112 #else
39113 __ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39114  float32x2_t __ret;
39115  __ret = (float32x2_t)(__p0);
39116  return __ret;
39117 }
39118 #endif
39119 
39120 #ifdef __LITTLE_ENDIAN__
39121 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39122  float32x2_t __ret;
39123  __ret = (float32x2_t)(__p0);
39124  return __ret;
39125 }
39126 #else
39127 __ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39128  float32x2_t __ret;
39129  __ret = (float32x2_t)(__p0);
39130  return __ret;
39131 }
39132 #endif
39133 
39134 #ifdef __LITTLE_ENDIAN__
39135 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39136  float32x2_t __ret;
39137  __ret = (float32x2_t)(__p0);
39138  return __ret;
39139 }
39140 #else
39141 __ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39142  float32x2_t __ret;
39143  __ret = (float32x2_t)(__p0);
39144  return __ret;
39145 }
39146 #endif
39147 
39148 #ifdef __LITTLE_ENDIAN__
39149 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39150  float16x4_t __ret;
39151  __ret = (float16x4_t)(__p0);
39152  return __ret;
39153 }
39154 #else
39155 __ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39156  float16x4_t __ret;
39157  __ret = (float16x4_t)(__p0);
39158  return __ret;
39159 }
39160 #endif
39161 
39162 #ifdef __LITTLE_ENDIAN__
39163 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39164  float16x4_t __ret;
39165  __ret = (float16x4_t)(__p0);
39166  return __ret;
39167 }
39168 #else
39169 __ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39170  float16x4_t __ret;
39171  __ret = (float16x4_t)(__p0);
39172  return __ret;
39173 }
39174 #endif
39175 
39176 #ifdef __LITTLE_ENDIAN__
39177 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39178  float16x4_t __ret;
39179  __ret = (float16x4_t)(__p0);
39180  return __ret;
39181 }
39182 #else
39183 __ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39184  float16x4_t __ret;
39185  __ret = (float16x4_t)(__p0);
39186  return __ret;
39187 }
39188 #endif
39189 
39190 #ifdef __LITTLE_ENDIAN__
39191 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39192  float16x4_t __ret;
39193  __ret = (float16x4_t)(__p0);
39194  return __ret;
39195 }
39196 #else
39197 __ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39198  float16x4_t __ret;
39199  __ret = (float16x4_t)(__p0);
39200  return __ret;
39201 }
39202 #endif
39203 
39204 #ifdef __LITTLE_ENDIAN__
39205 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39206  float16x4_t __ret;
39207  __ret = (float16x4_t)(__p0);
39208  return __ret;
39209 }
39210 #else
39211 __ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39212  float16x4_t __ret;
39213  __ret = (float16x4_t)(__p0);
39214  return __ret;
39215 }
39216 #endif
39217 
39218 #ifdef __LITTLE_ENDIAN__
39219 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39220  float16x4_t __ret;
39221  __ret = (float16x4_t)(__p0);
39222  return __ret;
39223 }
39224 #else
39225 __ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39226  float16x4_t __ret;
39227  __ret = (float16x4_t)(__p0);
39228  return __ret;
39229 }
39230 #endif
39231 
39232 #ifdef __LITTLE_ENDIAN__
39233 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39234  float16x4_t __ret;
39235  __ret = (float16x4_t)(__p0);
39236  return __ret;
39237 }
39238 #else
39239 __ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39240  float16x4_t __ret;
39241  __ret = (float16x4_t)(__p0);
39242  return __ret;
39243 }
39244 #endif
39245 
39246 #ifdef __LITTLE_ENDIAN__
39247 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39248  float16x4_t __ret;
39249  __ret = (float16x4_t)(__p0);
39250  return __ret;
39251 }
39252 #else
39253 __ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39254  float16x4_t __ret;
39255  __ret = (float16x4_t)(__p0);
39256  return __ret;
39257 }
39258 #endif
39259 
39260 #ifdef __LITTLE_ENDIAN__
39261 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39262  float16x4_t __ret;
39263  __ret = (float16x4_t)(__p0);
39264  return __ret;
39265 }
39266 #else
39267 __ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39268  float16x4_t __ret;
39269  __ret = (float16x4_t)(__p0);
39270  return __ret;
39271 }
39272 #endif
39273 
39274 #ifdef __LITTLE_ENDIAN__
39275 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39276  float16x4_t __ret;
39277  __ret = (float16x4_t)(__p0);
39278  return __ret;
39279 }
39280 #else
39281 __ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39282  float16x4_t __ret;
39283  __ret = (float16x4_t)(__p0);
39284  return __ret;
39285 }
39286 #endif
39287 
39288 #ifdef __LITTLE_ENDIAN__
39289 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39290  float16x4_t __ret;
39291  __ret = (float16x4_t)(__p0);
39292  return __ret;
39293 }
39294 #else
39295 __ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39296  float16x4_t __ret;
39297  __ret = (float16x4_t)(__p0);
39298  return __ret;
39299 }
39300 #endif
39301 
39302 #ifdef __LITTLE_ENDIAN__
39303 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39304  float16x4_t __ret;
39305  __ret = (float16x4_t)(__p0);
39306  return __ret;
39307 }
39308 #else
39309 __ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39310  float16x4_t __ret;
39311  __ret = (float16x4_t)(__p0);
39312  return __ret;
39313 }
39314 #endif
39315 
39316 #ifdef __LITTLE_ENDIAN__
39317 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39318  float16x4_t __ret;
39319  __ret = (float16x4_t)(__p0);
39320  return __ret;
39321 }
39322 #else
39323 __ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39324  float16x4_t __ret;
39325  __ret = (float16x4_t)(__p0);
39326  return __ret;
39327 }
39328 #endif
39329 
39330 #ifdef __LITTLE_ENDIAN__
39331 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39332  int32x2_t __ret;
39333  __ret = (int32x2_t)(__p0);
39334  return __ret;
39335 }
39336 #else
39337 __ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39338  int32x2_t __ret;
39339  __ret = (int32x2_t)(__p0);
39340  return __ret;
39341 }
39342 #endif
39343 
39344 #ifdef __LITTLE_ENDIAN__
39345 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39346  int32x2_t __ret;
39347  __ret = (int32x2_t)(__p0);
39348  return __ret;
39349 }
39350 #else
39351 __ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39352  int32x2_t __ret;
39353  __ret = (int32x2_t)(__p0);
39354  return __ret;
39355 }
39356 #endif
39357 
39358 #ifdef __LITTLE_ENDIAN__
39359 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39360  int32x2_t __ret;
39361  __ret = (int32x2_t)(__p0);
39362  return __ret;
39363 }
39364 #else
39365 __ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39366  int32x2_t __ret;
39367  __ret = (int32x2_t)(__p0);
39368  return __ret;
39369 }
39370 #endif
39371 
39372 #ifdef __LITTLE_ENDIAN__
39373 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39374  int32x2_t __ret;
39375  __ret = (int32x2_t)(__p0);
39376  return __ret;
39377 }
39378 #else
39379 __ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39380  int32x2_t __ret;
39381  __ret = (int32x2_t)(__p0);
39382  return __ret;
39383 }
39384 #endif
39385 
39386 #ifdef __LITTLE_ENDIAN__
39387 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39388  int32x2_t __ret;
39389  __ret = (int32x2_t)(__p0);
39390  return __ret;
39391 }
39392 #else
39393 __ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39394  int32x2_t __ret;
39395  __ret = (int32x2_t)(__p0);
39396  return __ret;
39397 }
39398 #endif
39399 
39400 #ifdef __LITTLE_ENDIAN__
39401 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39402  int32x2_t __ret;
39403  __ret = (int32x2_t)(__p0);
39404  return __ret;
39405 }
39406 #else
39407 __ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39408  int32x2_t __ret;
39409  __ret = (int32x2_t)(__p0);
39410  return __ret;
39411 }
39412 #endif
39413 
39414 #ifdef __LITTLE_ENDIAN__
39415 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39416  int32x2_t __ret;
39417  __ret = (int32x2_t)(__p0);
39418  return __ret;
39419 }
39420 #else
39421 __ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39422  int32x2_t __ret;
39423  __ret = (int32x2_t)(__p0);
39424  return __ret;
39425 }
39426 #endif
39427 
39428 #ifdef __LITTLE_ENDIAN__
39429 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39430  int32x2_t __ret;
39431  __ret = (int32x2_t)(__p0);
39432  return __ret;
39433 }
39434 #else
39435 __ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39436  int32x2_t __ret;
39437  __ret = (int32x2_t)(__p0);
39438  return __ret;
39439 }
39440 #endif
39441 
39442 #ifdef __LITTLE_ENDIAN__
39443 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39444  int32x2_t __ret;
39445  __ret = (int32x2_t)(__p0);
39446  return __ret;
39447 }
39448 #else
39449 __ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39450  int32x2_t __ret;
39451  __ret = (int32x2_t)(__p0);
39452  return __ret;
39453 }
39454 #endif
39455 
39456 #ifdef __LITTLE_ENDIAN__
39457 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39458  int32x2_t __ret;
39459  __ret = (int32x2_t)(__p0);
39460  return __ret;
39461 }
39462 #else
39463 __ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39464  int32x2_t __ret;
39465  __ret = (int32x2_t)(__p0);
39466  return __ret;
39467 }
39468 #endif
39469 
39470 #ifdef __LITTLE_ENDIAN__
39471 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39472  int32x2_t __ret;
39473  __ret = (int32x2_t)(__p0);
39474  return __ret;
39475 }
39476 #else
39477 __ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39478  int32x2_t __ret;
39479  __ret = (int32x2_t)(__p0);
39480  return __ret;
39481 }
39482 #endif
39483 
39484 #ifdef __LITTLE_ENDIAN__
39485 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39486  int32x2_t __ret;
39487  __ret = (int32x2_t)(__p0);
39488  return __ret;
39489 }
39490 #else
39491 __ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39492  int32x2_t __ret;
39493  __ret = (int32x2_t)(__p0);
39494  return __ret;
39495 }
39496 #endif
39497 
39498 #ifdef __LITTLE_ENDIAN__
39499 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39500  int32x2_t __ret;
39501  __ret = (int32x2_t)(__p0);
39502  return __ret;
39503 }
39504 #else
39505 __ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39506  int32x2_t __ret;
39507  __ret = (int32x2_t)(__p0);
39508  return __ret;
39509 }
39510 #endif
39511 
39512 #ifdef __LITTLE_ENDIAN__
39513 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39514  int64x1_t __ret;
39515  __ret = (int64x1_t)(__p0);
39516  return __ret;
39517 }
39518 #else
39519 __ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39520  int64x1_t __ret;
39521  __ret = (int64x1_t)(__p0);
39522  return __ret;
39523 }
39524 #endif
39525 
39526 #ifdef __LITTLE_ENDIAN__
39527 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39528  int64x1_t __ret;
39529  __ret = (int64x1_t)(__p0);
39530  return __ret;
39531 }
39532 #else
39533 __ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39534  int64x1_t __ret;
39535  __ret = (int64x1_t)(__p0);
39536  return __ret;
39537 }
39538 #endif
39539 
39540 #ifdef __LITTLE_ENDIAN__
39541 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39542  int64x1_t __ret;
39543  __ret = (int64x1_t)(__p0);
39544  return __ret;
39545 }
39546 #else
39547 __ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39548  int64x1_t __ret;
39549  __ret = (int64x1_t)(__p0);
39550  return __ret;
39551 }
39552 #endif
39553 
39554 #ifdef __LITTLE_ENDIAN__
39555 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39556  int64x1_t __ret;
39557  __ret = (int64x1_t)(__p0);
39558  return __ret;
39559 }
39560 #else
39561 __ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39562  int64x1_t __ret;
39563  __ret = (int64x1_t)(__p0);
39564  return __ret;
39565 }
39566 #endif
39567 
39568 #ifdef __LITTLE_ENDIAN__
39569 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39570  int64x1_t __ret;
39571  __ret = (int64x1_t)(__p0);
39572  return __ret;
39573 }
39574 #else
39575 __ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39576  int64x1_t __ret;
39577  __ret = (int64x1_t)(__p0);
39578  return __ret;
39579 }
39580 #endif
39581 
39582 #ifdef __LITTLE_ENDIAN__
39583 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39584  int64x1_t __ret;
39585  __ret = (int64x1_t)(__p0);
39586  return __ret;
39587 }
39588 #else
39589 __ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39590  int64x1_t __ret;
39591  __ret = (int64x1_t)(__p0);
39592  return __ret;
39593 }
39594 #endif
39595 
39596 #ifdef __LITTLE_ENDIAN__
39597 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39598  int64x1_t __ret;
39599  __ret = (int64x1_t)(__p0);
39600  return __ret;
39601 }
39602 #else
39603 __ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39604  int64x1_t __ret;
39605  __ret = (int64x1_t)(__p0);
39606  return __ret;
39607 }
39608 #endif
39609 
39610 #ifdef __LITTLE_ENDIAN__
39611 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39612  int64x1_t __ret;
39613  __ret = (int64x1_t)(__p0);
39614  return __ret;
39615 }
39616 #else
39617 __ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39618  int64x1_t __ret;
39619  __ret = (int64x1_t)(__p0);
39620  return __ret;
39621 }
39622 #endif
39623 
39624 #ifdef __LITTLE_ENDIAN__
39625 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39626  int64x1_t __ret;
39627  __ret = (int64x1_t)(__p0);
39628  return __ret;
39629 }
39630 #else
39631 __ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39632  int64x1_t __ret;
39633  __ret = (int64x1_t)(__p0);
39634  return __ret;
39635 }
39636 #endif
39637 
39638 #ifdef __LITTLE_ENDIAN__
39639 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39640  int64x1_t __ret;
39641  __ret = (int64x1_t)(__p0);
39642  return __ret;
39643 }
39644 #else
39645 __ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39646  int64x1_t __ret;
39647  __ret = (int64x1_t)(__p0);
39648  return __ret;
39649 }
39650 #endif
39651 
39652 #ifdef __LITTLE_ENDIAN__
39653 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39654  int64x1_t __ret;
39655  __ret = (int64x1_t)(__p0);
39656  return __ret;
39657 }
39658 #else
39659 __ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39660  int64x1_t __ret;
39661  __ret = (int64x1_t)(__p0);
39662  return __ret;
39663 }
39664 #endif
39665 
39666 #ifdef __LITTLE_ENDIAN__
39667 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39668  int64x1_t __ret;
39669  __ret = (int64x1_t)(__p0);
39670  return __ret;
39671 }
39672 #else
39673 __ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39674  int64x1_t __ret;
39675  __ret = (int64x1_t)(__p0);
39676  return __ret;
39677 }
39678 #endif
39679 
39680 #ifdef __LITTLE_ENDIAN__
39681 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39682  int64x1_t __ret;
39683  __ret = (int64x1_t)(__p0);
39684  return __ret;
39685 }
39686 #else
39687 __ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39688  int64x1_t __ret;
39689  __ret = (int64x1_t)(__p0);
39690  return __ret;
39691 }
39692 #endif
39693 
39694 #ifdef __LITTLE_ENDIAN__
39695 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39696  int16x4_t __ret;
39697  __ret = (int16x4_t)(__p0);
39698  return __ret;
39699 }
39700 #else
39701 __ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39702  int16x4_t __ret;
39703  __ret = (int16x4_t)(__p0);
39704  return __ret;
39705 }
39706 #endif
39707 
39708 #ifdef __LITTLE_ENDIAN__
39709 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39710  int16x4_t __ret;
39711  __ret = (int16x4_t)(__p0);
39712  return __ret;
39713 }
39714 #else
39715 __ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39716  int16x4_t __ret;
39717  __ret = (int16x4_t)(__p0);
39718  return __ret;
39719 }
39720 #endif
39721 
39722 #ifdef __LITTLE_ENDIAN__
39723 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39724  int16x4_t __ret;
39725  __ret = (int16x4_t)(__p0);
39726  return __ret;
39727 }
39728 #else
39729 __ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39730  int16x4_t __ret;
39731  __ret = (int16x4_t)(__p0);
39732  return __ret;
39733 }
39734 #endif
39735 
39736 #ifdef __LITTLE_ENDIAN__
39737 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39738  int16x4_t __ret;
39739  __ret = (int16x4_t)(__p0);
39740  return __ret;
39741 }
39742 #else
39743 __ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39744  int16x4_t __ret;
39745  __ret = (int16x4_t)(__p0);
39746  return __ret;
39747 }
39748 #endif
39749 
39750 #ifdef __LITTLE_ENDIAN__
39751 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39752  int16x4_t __ret;
39753  __ret = (int16x4_t)(__p0);
39754  return __ret;
39755 }
39756 #else
39757 __ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39758  int16x4_t __ret;
39759  __ret = (int16x4_t)(__p0);
39760  return __ret;
39761 }
39762 #endif
39763 
39764 #ifdef __LITTLE_ENDIAN__
39765 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39766  int16x4_t __ret;
39767  __ret = (int16x4_t)(__p0);
39768  return __ret;
39769 }
39770 #else
39771 __ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39772  int16x4_t __ret;
39773  __ret = (int16x4_t)(__p0);
39774  return __ret;
39775 }
39776 #endif
39777 
39778 #ifdef __LITTLE_ENDIAN__
39779 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39780  int16x4_t __ret;
39781  __ret = (int16x4_t)(__p0);
39782  return __ret;
39783 }
39784 #else
39785 __ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39786  int16x4_t __ret;
39787  __ret = (int16x4_t)(__p0);
39788  return __ret;
39789 }
39790 #endif
39791 
39792 #ifdef __LITTLE_ENDIAN__
39793 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39794  int16x4_t __ret;
39795  __ret = (int16x4_t)(__p0);
39796  return __ret;
39797 }
39798 #else
39799 __ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39800  int16x4_t __ret;
39801  __ret = (int16x4_t)(__p0);
39802  return __ret;
39803 }
39804 #endif
39805 
39806 #ifdef __LITTLE_ENDIAN__
39807 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39808  int16x4_t __ret;
39809  __ret = (int16x4_t)(__p0);
39810  return __ret;
39811 }
39812 #else
39813 __ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39814  int16x4_t __ret;
39815  __ret = (int16x4_t)(__p0);
39816  return __ret;
39817 }
39818 #endif
39819 
39820 #ifdef __LITTLE_ENDIAN__
39821 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39822  int16x4_t __ret;
39823  __ret = (int16x4_t)(__p0);
39824  return __ret;
39825 }
39826 #else
39827 __ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39828  int16x4_t __ret;
39829  __ret = (int16x4_t)(__p0);
39830  return __ret;
39831 }
39832 #endif
39833 
39834 #ifdef __LITTLE_ENDIAN__
39835 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39836  int16x4_t __ret;
39837  __ret = (int16x4_t)(__p0);
39838  return __ret;
39839 }
39840 #else
39841 __ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39842  int16x4_t __ret;
39843  __ret = (int16x4_t)(__p0);
39844  return __ret;
39845 }
39846 #endif
39847 
39848 #ifdef __LITTLE_ENDIAN__
39849 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39850  int16x4_t __ret;
39851  __ret = (int16x4_t)(__p0);
39852  return __ret;
39853 }
39854 #else
39855 __ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39856  int16x4_t __ret;
39857  __ret = (int16x4_t)(__p0);
39858  return __ret;
39859 }
39860 #endif
39861 
39862 #ifdef __LITTLE_ENDIAN__
39863 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39864  int16x4_t __ret;
39865  __ret = (int16x4_t)(__p0);
39866  return __ret;
39867 }
39868 #else
39869 __ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39870  int16x4_t __ret;
39871  __ret = (int16x4_t)(__p0);
39872  return __ret;
39873 }
39874 #endif
39875 
39876 #endif
39877 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
39878 #ifdef __LITTLE_ENDIAN__
39879 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
39880  float64x2_t __ret;
39881  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
39882  return __ret;
39883 }
39884 #else
39885 __ai float64x2_t vrndq_f64(float64x2_t __p0) {
39886  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39887  float64x2_t __ret;
39888  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
39889  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39890  return __ret;
39891 }
39892 #endif
39893 
39894 #ifdef __LITTLE_ENDIAN__
39895 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
39896  float64x1_t __ret;
39897  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39898  return __ret;
39899 }
39900 #else
39901 __ai float64x1_t vrnd_f64(float64x1_t __p0) {
39902  float64x1_t __ret;
39903  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39904  return __ret;
39905 }
39906 #endif
39907 
39908 #ifdef __LITTLE_ENDIAN__
39909 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39910  float64x2_t __ret;
39911  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
39912  return __ret;
39913 }
39914 #else
39915 __ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39916  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39917  float64x2_t __ret;
39918  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
39919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39920  return __ret;
39921 }
39922 #endif
39923 
39924 #ifdef __LITTLE_ENDIAN__
39925 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
39926  float64x1_t __ret;
39927  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39928  return __ret;
39929 }
39930 #else
39931 __ai float64x1_t vrnda_f64(float64x1_t __p0) {
39932  float64x1_t __ret;
39933  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39934  return __ret;
39935 }
39936 #endif
39937 
39938 #ifdef __LITTLE_ENDIAN__
39939 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39940  float64x2_t __ret;
39941  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
39942  return __ret;
39943 }
39944 #else
39945 __ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39946  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39947  float64x2_t __ret;
39948  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
39949  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39950  return __ret;
39951 }
39952 #endif
39953 
39954 #ifdef __LITTLE_ENDIAN__
39955 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39956  float32x4_t __ret;
39957  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
39958  return __ret;
39959 }
39960 #else
39961 __ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39962  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39963  float32x4_t __ret;
39964  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
39965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39966  return __ret;
39967 }
39968 #endif
39969 
39970 #ifdef __LITTLE_ENDIAN__
39971 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
39972  float64x1_t __ret;
39973  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39974  return __ret;
39975 }
39976 #else
39977 __ai float64x1_t vrndi_f64(float64x1_t __p0) {
39978  float64x1_t __ret;
39979  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39980  return __ret;
39981 }
39982 #endif
39983 
39984 #ifdef __LITTLE_ENDIAN__
39985 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
39986  float32x2_t __ret;
39987  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
39988  return __ret;
39989 }
39990 #else
39991 __ai float32x2_t vrndi_f32(float32x2_t __p0) {
39992  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39993  float32x2_t __ret;
39994  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
39995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39996  return __ret;
39997 }
39998 #endif
39999 
40000 #ifdef __LITTLE_ENDIAN__
40001 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
40002  float64x2_t __ret;
40003  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
40004  return __ret;
40005 }
40006 #else
40007 __ai float64x2_t vrndmq_f64(float64x2_t __p0) {
40008  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40009  float64x2_t __ret;
40010  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
40011  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40012  return __ret;
40013 }
40014 #endif
40015 
40016 #ifdef __LITTLE_ENDIAN__
40017 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
40018  float64x1_t __ret;
40019  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
40020  return __ret;
40021 }
40022 #else
40023 __ai float64x1_t vrndm_f64(float64x1_t __p0) {
40024  float64x1_t __ret;
40025  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
40026  return __ret;
40027 }
40028 #endif
40029 
40030 #ifdef __LITTLE_ENDIAN__
40031 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
40032  float64x2_t __ret;
40033  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
40034  return __ret;
40035 }
40036 #else
40037 __ai float64x2_t vrndnq_f64(float64x2_t __p0) {
40038  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40039  float64x2_t __ret;
40040  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
40041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40042  return __ret;
40043 }
40044 #endif
40045 
40046 #ifdef __LITTLE_ENDIAN__
40047 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
40048  float64x1_t __ret;
40049  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40050  return __ret;
40051 }
40052 #else
40053 __ai float64x1_t vrndn_f64(float64x1_t __p0) {
40054  float64x1_t __ret;
40055  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40056  return __ret;
40057 }
40058 #endif
40059 
40060 #ifdef __LITTLE_ENDIAN__
40061 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40062  float64x2_t __ret;
40063  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
40064  return __ret;
40065 }
40066 #else
40067 __ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40068  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40069  float64x2_t __ret;
40070  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
40071  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40072  return __ret;
40073 }
40074 #endif
40075 
40076 #ifdef __LITTLE_ENDIAN__
40077 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
40078  float64x1_t __ret;
40079  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40080  return __ret;
40081 }
40082 #else
40083 __ai float64x1_t vrndp_f64(float64x1_t __p0) {
40084  float64x1_t __ret;
40085  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40086  return __ret;
40087 }
40088 #endif
40089 
40090 #ifdef __LITTLE_ENDIAN__
40091 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40092  float64x2_t __ret;
40093  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
40094  return __ret;
40095 }
40096 #else
40097 __ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40098  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40099  float64x2_t __ret;
40100  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
40101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40102  return __ret;
40103 }
40104 #endif
40105 
40106 #ifdef __LITTLE_ENDIAN__
40107 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
40108  float64x1_t __ret;
40109  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40110  return __ret;
40111 }
40112 #else
40113 __ai float64x1_t vrndx_f64(float64x1_t __p0) {
40114  float64x1_t __ret;
40115  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40116  return __ret;
40117 }
40118 #endif
40119 
40120 #endif
40121 #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
40122 #ifdef __LITTLE_ENDIAN__
40123 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40124  float64x2_t __ret;
40125  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40126  return __ret;
40127 }
40128 #else
40129 __ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40130  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40131  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40132  float64x2_t __ret;
40133  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40134  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40135  return __ret;
40136 }
40137 #endif
40138 
40139 #ifdef __LITTLE_ENDIAN__
40140 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40141  float64x1_t __ret;
40142  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40143  return __ret;
40144 }
40145 #else
40146 __ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40147  float64x1_t __ret;
40148  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40149  return __ret;
40150 }
40151 #endif
40152 
40153 #ifdef __LITTLE_ENDIAN__
40154 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40155  float64x2_t __ret;
40156  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40157  return __ret;
40158 }
40159 #else
40160 __ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40161  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40162  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40163  float64x2_t __ret;
40164  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40166  return __ret;
40167 }
40168 #endif
40169 
40170 #ifdef __LITTLE_ENDIAN__
40171 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40172  float64x1_t __ret;
40173  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40174  return __ret;
40175 }
40176 #else
40177 __ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40178  float64x1_t __ret;
40179  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40180  return __ret;
40181 }
40182 #endif
40183 
40184 #endif
40185 #if __ARM_FEATURE_CRYPTO
40186 #ifdef __LITTLE_ENDIAN__
40187 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40188  uint8x16_t __ret;
40189  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40190  return __ret;
40191 }
40192 #else
40193 __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40194  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40195  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40196  uint8x16_t __ret;
40197  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40199  return __ret;
40200 }
40201 #endif
40202 
40203 #ifdef __LITTLE_ENDIAN__
40204 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40205  uint8x16_t __ret;
40206  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40207  return __ret;
40208 }
40209 #else
40210 __ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40211  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40212  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40213  uint8x16_t __ret;
40214  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40216  return __ret;
40217 }
40218 #endif
40219 
40220 #ifdef __LITTLE_ENDIAN__
40221 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40222  uint8x16_t __ret;
40223  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
40224  return __ret;
40225 }
40226 #else
40227 __ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40228  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40229  uint8x16_t __ret;
40230  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
40231  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40232  return __ret;
40233 }
40234 #endif
40235 
40236 #ifdef __LITTLE_ENDIAN__
40237 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40238  uint8x16_t __ret;
40239  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
40240  return __ret;
40241 }
40242 #else
40243 __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40244  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40245  uint8x16_t __ret;
40246  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
40247  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40248  return __ret;
40249 }
40250 #endif
40251 
40252 #ifdef __LITTLE_ENDIAN__
40253 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40254  uint32x4_t __ret;
40255  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40256  return __ret;
40257 }
40258 #else
40259 __ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40260  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40261  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40262  uint32x4_t __ret;
40263  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40264  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40265  return __ret;
40266 }
40267 #endif
40268 
40269 #ifdef __LITTLE_ENDIAN__
40270 __ai uint32_t vsha1h_u32(uint32_t __p0) {
40271  uint32_t __ret;
40272  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40273  return __ret;
40274 }
40275 #else
40276 __ai uint32_t vsha1h_u32(uint32_t __p0) {
40277  uint32_t __ret;
40278  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40279  return __ret;
40280 }
40281 #endif
40282 
40283 #ifdef __LITTLE_ENDIAN__
40284 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40285  uint32x4_t __ret;
40286  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40287  return __ret;
40288 }
40289 #else
40290 __ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40291  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40292  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40293  uint32x4_t __ret;
40294  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40295  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40296  return __ret;
40297 }
40298 #endif
40299 
40300 #ifdef __LITTLE_ENDIAN__
40301 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40302  uint32x4_t __ret;
40303  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40304  return __ret;
40305 }
40306 #else
40307 __ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40308  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40309  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40310  uint32x4_t __ret;
40311  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40312  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40313  return __ret;
40314 }
40315 #endif
40316 
40317 #ifdef __LITTLE_ENDIAN__
40318 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40319  uint32x4_t __ret;
40320  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40321  return __ret;
40322 }
40323 #else
40324 __ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40325  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40326  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40327  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40328  uint32x4_t __ret;
40329  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40330  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40331  return __ret;
40332 }
40333 #endif
40334 
40335 #ifdef __LITTLE_ENDIAN__
40336 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40337  uint32x4_t __ret;
40338  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40339  return __ret;
40340 }
40341 #else
40342 __ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40343  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40344  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40345  uint32x4_t __ret;
40346  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40347  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40348  return __ret;
40349 }
40350 #endif
40351 
40352 #ifdef __LITTLE_ENDIAN__
40353 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40354  uint32x4_t __ret;
40355  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40356  return __ret;
40357 }
40358 #else
40359 __ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40360  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40361  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40362  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40363  uint32x4_t __ret;
40364  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40366  return __ret;
40367 }
40368 #endif
40369 
40370 #ifdef __LITTLE_ENDIAN__
40371 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40372  uint32x4_t __ret;
40373  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40374  return __ret;
40375 }
40376 #else
40377 __ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40378  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40379  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40380  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40381  uint32x4_t __ret;
40382  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40383  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40384  return __ret;
40385 }
40386 #endif
40387 
40388 #ifdef __LITTLE_ENDIAN__
40389 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40390  uint32x4_t __ret;
40391  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40392  return __ret;
40393 }
40394 #else
40395 __ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40396  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40397  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40398  uint32x4_t __ret;
40399  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40400  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40401  return __ret;
40402 }
40403 #endif
40404 
40405 #ifdef __LITTLE_ENDIAN__
40406 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40407  uint32x4_t __ret;
40408  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40409  return __ret;
40410 }
40411 #else
40412 __ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40413  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40414  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40415  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40416  uint32x4_t __ret;
40417  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40419  return __ret;
40420 }
40421 #endif
40422 
40423 #endif
40424 #if defined(__ARM_FEATURE_QRDMX)
40425 #ifdef __LITTLE_ENDIAN__
40426 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40427  int32x4_t __ret;
40428  __ret = vqaddq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40429  return __ret;
40430 }
40431 #else
40432 __ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40433  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40434  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40435  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40436  int32x4_t __ret;
40437  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40438  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40439  return __ret;
40440 }
40441 #endif
40442 
40443 #ifdef __LITTLE_ENDIAN__
40444 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40445  int16x8_t __ret;
40446  __ret = vqaddq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40447  return __ret;
40448 }
40449 #else
40450 __ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40451  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40452  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40453  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40454  int16x8_t __ret;
40455  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40456  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40457  return __ret;
40458 }
40459 #endif
40460 
40461 #ifdef __LITTLE_ENDIAN__
40462 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40463  int32x2_t __ret;
40464  __ret = vqadd_s32(__p0, vqrdmulh_s32(__p1, __p2));
40465  return __ret;
40466 }
40467 #else
40468 __ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40469  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40470  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40471  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40472  int32x2_t __ret;
40473  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40475  return __ret;
40476 }
40477 #endif
40478 
40479 #ifdef __LITTLE_ENDIAN__
40480 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40481  int16x4_t __ret;
40482  __ret = vqadd_s16(__p0, vqrdmulh_s16(__p1, __p2));
40483  return __ret;
40484 }
40485 #else
40486 __ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40487  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40488  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40489  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40490  int16x4_t __ret;
40491  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40492  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40493  return __ret;
40494 }
40495 #endif
40496 
40497 #ifdef __LITTLE_ENDIAN__
40498 #define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40499  int32x4_t __s0 = __p0; \
40500  int32x4_t __s1 = __p1; \
40501  int32x2_t __s2 = __p2; \
40502  int32x4_t __ret; \
40503  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40504  __ret; \
40505 })
40506 #else
40507 #define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40508  int32x4_t __s0 = __p0; \
40509  int32x4_t __s1 = __p1; \
40510  int32x2_t __s2 = __p2; \
40511  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40512  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40513  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40514  int32x4_t __ret; \
40515  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40517  __ret; \
40518 })
40519 #endif
40520 
40521 #ifdef __LITTLE_ENDIAN__
40522 #define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40523  int16x8_t __s0 = __p0; \
40524  int16x8_t __s1 = __p1; \
40525  int16x4_t __s2 = __p2; \
40526  int16x8_t __ret; \
40527  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40528  __ret; \
40529 })
40530 #else
40531 #define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40532  int16x8_t __s0 = __p0; \
40533  int16x8_t __s1 = __p1; \
40534  int16x4_t __s2 = __p2; \
40535  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40536  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40537  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40538  int16x8_t __ret; \
40539  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40540  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40541  __ret; \
40542 })
40543 #endif
40544 
40545 #ifdef __LITTLE_ENDIAN__
40546 #define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40547  int32x2_t __s0 = __p0; \
40548  int32x2_t __s1 = __p1; \
40549  int32x2_t __s2 = __p2; \
40550  int32x2_t __ret; \
40551  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40552  __ret; \
40553 })
40554 #else
40555 #define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40556  int32x2_t __s0 = __p0; \
40557  int32x2_t __s1 = __p1; \
40558  int32x2_t __s2 = __p2; \
40559  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40560  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40561  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40562  int32x2_t __ret; \
40563  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40565  __ret; \
40566 })
40567 #endif
40568 
40569 #ifdef __LITTLE_ENDIAN__
40570 #define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40571  int16x4_t __s0 = __p0; \
40572  int16x4_t __s1 = __p1; \
40573  int16x4_t __s2 = __p2; \
40574  int16x4_t __ret; \
40575  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40576  __ret; \
40577 })
40578 #else
40579 #define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40580  int16x4_t __s0 = __p0; \
40581  int16x4_t __s1 = __p1; \
40582  int16x4_t __s2 = __p2; \
40583  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40584  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40585  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40586  int16x4_t __ret; \
40587  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40588  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40589  __ret; \
40590 })
40591 #endif
40592 
40593 #ifdef __LITTLE_ENDIAN__
40594 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40595  int32x4_t __ret;
40596  __ret = vqsubq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40597  return __ret;
40598 }
40599 #else
40600 __ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40601  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40602  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40603  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40604  int32x4_t __ret;
40605  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40606  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40607  return __ret;
40608 }
40609 #endif
40610 
40611 #ifdef __LITTLE_ENDIAN__
40612 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40613  int16x8_t __ret;
40614  __ret = vqsubq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40615  return __ret;
40616 }
40617 #else
40618 __ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40619  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40620  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40621  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40622  int16x8_t __ret;
40623  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40625  return __ret;
40626 }
40627 #endif
40628 
40629 #ifdef __LITTLE_ENDIAN__
40630 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40631  int32x2_t __ret;
40632  __ret = vqsub_s32(__p0, vqrdmulh_s32(__p1, __p2));
40633  return __ret;
40634 }
40635 #else
40636 __ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40637  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40638  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40639  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40640  int32x2_t __ret;
40641  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40642  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40643  return __ret;
40644 }
40645 #endif
40646 
40647 #ifdef __LITTLE_ENDIAN__
40648 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40649  int16x4_t __ret;
40650  __ret = vqsub_s16(__p0, vqrdmulh_s16(__p1, __p2));
40651  return __ret;
40652 }
40653 #else
40654 __ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40655  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40656  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40657  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40658  int16x4_t __ret;
40659  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40660  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40661  return __ret;
40662 }
40663 #endif
40664 
40665 #ifdef __LITTLE_ENDIAN__
40666 #define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40667  int32x4_t __s0 = __p0; \
40668  int32x4_t __s1 = __p1; \
40669  int32x2_t __s2 = __p2; \
40670  int32x4_t __ret; \
40671  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40672  __ret; \
40673 })
40674 #else
40675 #define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40676  int32x4_t __s0 = __p0; \
40677  int32x4_t __s1 = __p1; \
40678  int32x2_t __s2 = __p2; \
40679  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40680  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40681  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40682  int32x4_t __ret; \
40683  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40684  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40685  __ret; \
40686 })
40687 #endif
40688 
40689 #ifdef __LITTLE_ENDIAN__
40690 #define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40691  int16x8_t __s0 = __p0; \
40692  int16x8_t __s1 = __p1; \
40693  int16x4_t __s2 = __p2; \
40694  int16x8_t __ret; \
40695  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40696  __ret; \
40697 })
40698 #else
40699 #define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40700  int16x8_t __s0 = __p0; \
40701  int16x8_t __s1 = __p1; \
40702  int16x4_t __s2 = __p2; \
40703  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40704  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40705  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40706  int16x8_t __ret; \
40707  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40708  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40709  __ret; \
40710 })
40711 #endif
40712 
40713 #ifdef __LITTLE_ENDIAN__
40714 #define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40715  int32x2_t __s0 = __p0; \
40716  int32x2_t __s1 = __p1; \
40717  int32x2_t __s2 = __p2; \
40718  int32x2_t __ret; \
40719  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40720  __ret; \
40721 })
40722 #else
40723 #define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40724  int32x2_t __s0 = __p0; \
40725  int32x2_t __s1 = __p1; \
40726  int32x2_t __s2 = __p2; \
40727  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40728  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40729  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40730  int32x2_t __ret; \
40731  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40733  __ret; \
40734 })
40735 #endif
40736 
40737 #ifdef __LITTLE_ENDIAN__
40738 #define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40739  int16x4_t __s0 = __p0; \
40740  int16x4_t __s1 = __p1; \
40741  int16x4_t __s2 = __p2; \
40742  int16x4_t __ret; \
40743  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40744  __ret; \
40745 })
40746 #else
40747 #define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40748  int16x4_t __s0 = __p0; \
40749  int16x4_t __s1 = __p1; \
40750  int16x4_t __s2 = __p2; \
40751  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40752  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40753  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40754  int16x4_t __ret; \
40755  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40756  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40757  __ret; \
40758 })
40759 #endif
40760 
40761 #endif
40762 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
40763 #ifdef __LITTLE_ENDIAN__
40764 #define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40765  int32x4_t __s0 = __p0; \
40766  int32x4_t __s1 = __p1; \
40767  int32x4_t __s2 = __p2; \
40768  int32x4_t __ret; \
40769  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40770  __ret; \
40771 })
40772 #else
40773 #define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40774  int32x4_t __s0 = __p0; \
40775  int32x4_t __s1 = __p1; \
40776  int32x4_t __s2 = __p2; \
40777  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40778  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40779  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40780  int32x4_t __ret; \
40781  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40783  __ret; \
40784 })
40785 #endif
40786 
40787 #ifdef __LITTLE_ENDIAN__
40788 #define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40789  int16x8_t __s0 = __p0; \
40790  int16x8_t __s1 = __p1; \
40791  int16x8_t __s2 = __p2; \
40792  int16x8_t __ret; \
40793  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40794  __ret; \
40795 })
40796 #else
40797 #define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40798  int16x8_t __s0 = __p0; \
40799  int16x8_t __s1 = __p1; \
40800  int16x8_t __s2 = __p2; \
40801  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40802  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40803  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40804  int16x8_t __ret; \
40805  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40806  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40807  __ret; \
40808 })
40809 #endif
40810 
40811 #ifdef __LITTLE_ENDIAN__
40812 #define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40813  int32x2_t __s0 = __p0; \
40814  int32x2_t __s1 = __p1; \
40815  int32x4_t __s2 = __p2; \
40816  int32x2_t __ret; \
40817  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40818  __ret; \
40819 })
40820 #else
40821 #define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40822  int32x2_t __s0 = __p0; \
40823  int32x2_t __s1 = __p1; \
40824  int32x4_t __s2 = __p2; \
40825  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40826  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40827  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40828  int32x2_t __ret; \
40829  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40830  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40831  __ret; \
40832 })
40833 #endif
40834 
40835 #ifdef __LITTLE_ENDIAN__
40836 #define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40837  int16x4_t __s0 = __p0; \
40838  int16x4_t __s1 = __p1; \
40839  int16x8_t __s2 = __p2; \
40840  int16x4_t __ret; \
40841  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40842  __ret; \
40843 })
40844 #else
40845 #define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40846  int16x4_t __s0 = __p0; \
40847  int16x4_t __s1 = __p1; \
40848  int16x8_t __s2 = __p2; \
40849  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40850  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40851  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40852  int16x4_t __ret; \
40853  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40855  __ret; \
40856 })
40857 #endif
40858 
40859 #ifdef __LITTLE_ENDIAN__
40860 #define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40861  int32x4_t __s0 = __p0; \
40862  int32x4_t __s1 = __p1; \
40863  int32x4_t __s2 = __p2; \
40864  int32x4_t __ret; \
40865  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40866  __ret; \
40867 })
40868 #else
40869 #define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40870  int32x4_t __s0 = __p0; \
40871  int32x4_t __s1 = __p1; \
40872  int32x4_t __s2 = __p2; \
40873  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40874  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40875  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40876  int32x4_t __ret; \
40877  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40878  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40879  __ret; \
40880 })
40881 #endif
40882 
40883 #ifdef __LITTLE_ENDIAN__
40884 #define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40885  int16x8_t __s0 = __p0; \
40886  int16x8_t __s1 = __p1; \
40887  int16x8_t __s2 = __p2; \
40888  int16x8_t __ret; \
40889  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40890  __ret; \
40891 })
40892 #else
40893 #define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40894  int16x8_t __s0 = __p0; \
40895  int16x8_t __s1 = __p1; \
40896  int16x8_t __s2 = __p2; \
40897  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40898  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40899  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40900  int16x8_t __ret; \
40901  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40903  __ret; \
40904 })
40905 #endif
40906 
40907 #ifdef __LITTLE_ENDIAN__
40908 #define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40909  int32x2_t __s0 = __p0; \
40910  int32x2_t __s1 = __p1; \
40911  int32x4_t __s2 = __p2; \
40912  int32x2_t __ret; \
40913  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40914  __ret; \
40915 })
40916 #else
40917 #define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40918  int32x2_t __s0 = __p0; \
40919  int32x2_t __s1 = __p1; \
40920  int32x4_t __s2 = __p2; \
40921  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40922  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40923  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40924  int32x2_t __ret; \
40925  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40927  __ret; \
40928 })
40929 #endif
40930 
40931 #ifdef __LITTLE_ENDIAN__
40932 #define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40933  int16x4_t __s0 = __p0; \
40934  int16x4_t __s1 = __p1; \
40935  int16x8_t __s2 = __p2; \
40936  int16x4_t __ret; \
40937  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40938  __ret; \
40939 })
40940 #else
40941 #define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40942  int16x4_t __s0 = __p0; \
40943  int16x4_t __s1 = __p1; \
40944  int16x8_t __s2 = __p2; \
40945  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40946  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40947  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40948  int16x4_t __ret; \
40949  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40950  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40951  __ret; \
40952 })
40953 #endif
40954 
40955 #endif
40956 #if defined(__aarch64__)
40957 #ifdef __LITTLE_ENDIAN__
40958 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
40959  float64x2_t __ret;
40960  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40961  return __ret;
40962 }
40963 #else
40964 __ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
40965  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40966  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40967  float64x2_t __ret;
40968  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40970  return __ret;
40971 }
40972 #endif
40973 
40974 #ifdef __LITTLE_ENDIAN__
40975 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
40976  float64x1_t __ret;
40977  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40978  return __ret;
40979 }
40980 #else
40981 __ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
40982  float64x1_t __ret;
40983  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40984  return __ret;
40985 }
40986 #endif
40987 
40988 #ifdef __LITTLE_ENDIAN__
40989 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
40990  float64_t __ret;
40991  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
40992  return __ret;
40993 }
40994 #else
40995 __ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
40996  float64_t __ret;
40997  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
40998  return __ret;
40999 }
41000 #endif
41001 
41002 #ifdef __LITTLE_ENDIAN__
41003 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41004  float32_t __ret;
41005  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41006  return __ret;
41007 }
41008 #else
41009 __ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41010  float32_t __ret;
41011  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41012  return __ret;
41013 }
41014 #endif
41015 
41016 #ifdef __LITTLE_ENDIAN__
41017 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
41018  float64x2_t __ret;
41019  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
41020  return __ret;
41021 }
41022 #else
41023 __ai float64x2_t vabsq_f64(float64x2_t __p0) {
41024  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41025  float64x2_t __ret;
41026  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
41027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41028  return __ret;
41029 }
41030 #endif
41031 
41032 #ifdef __LITTLE_ENDIAN__
41033 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
41034  int64x2_t __ret;
41035  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
41036  return __ret;
41037 }
41038 #else
41039 __ai int64x2_t vabsq_s64(int64x2_t __p0) {
41040  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41041  int64x2_t __ret;
41042  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
41043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41044  return __ret;
41045 }
41046 #endif
41047 
41048 #ifdef __LITTLE_ENDIAN__
41049 __ai float64x1_t vabs_f64(float64x1_t __p0) {
41050  float64x1_t __ret;
41051  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41052  return __ret;
41053 }
41054 #else
41055 __ai float64x1_t vabs_f64(float64x1_t __p0) {
41056  float64x1_t __ret;
41057  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41058  return __ret;
41059 }
41060 #endif
41061 
41062 #ifdef __LITTLE_ENDIAN__
41063 __ai int64x1_t vabs_s64(int64x1_t __p0) {
41064  int64x1_t __ret;
41065  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41066  return __ret;
41067 }
41068 #else
41069 __ai int64x1_t vabs_s64(int64x1_t __p0) {
41070  int64x1_t __ret;
41071  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41072  return __ret;
41073 }
41074 #endif
41075 
41076 #ifdef __LITTLE_ENDIAN__
41077 __ai int64_t vabsd_s64(int64_t __p0) {
41078  int64_t __ret;
41079  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41080  return __ret;
41081 }
41082 #else
41083 __ai int64_t vabsd_s64(int64_t __p0) {
41084  int64_t __ret;
41085  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41086  return __ret;
41087 }
41088 #endif
41089 
41090 #ifdef __LITTLE_ENDIAN__
41091 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41092  float64x2_t __ret;
41093  __ret = __p0 + __p1;
41094  return __ret;
41095 }
41096 #else
41097 __ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41098  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41099  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41100  float64x2_t __ret;
41101  __ret = __rev0 + __rev1;
41102  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41103  return __ret;
41104 }
41105 #endif
41106 
41107 #ifdef __LITTLE_ENDIAN__
41108 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41109  float64x1_t __ret;
41110  __ret = __p0 + __p1;
41111  return __ret;
41112 }
41113 #else
41114 __ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41115  float64x1_t __ret;
41116  __ret = __p0 + __p1;
41117  return __ret;
41118 }
41119 #endif
41120 
41121 #ifdef __LITTLE_ENDIAN__
41122 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41123  uint64_t __ret;
41124  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41125  return __ret;
41126 }
41127 #else
41128 __ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41129  uint64_t __ret;
41130  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41131  return __ret;
41132 }
41133 #endif
41134 
41135 #ifdef __LITTLE_ENDIAN__
41136 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41137  int64_t __ret;
41138  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41139  return __ret;
41140 }
41141 #else
41142 __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41143  int64_t __ret;
41144  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41145  return __ret;
41146 }
41147 #endif
41148 
41149 #ifdef __LITTLE_ENDIAN__
41150 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41151  uint16x8_t __ret;
41152  __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
41153  return __ret;
41154 }
41155 #else
41156 __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41157  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41158  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41159  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41160  uint16x8_t __ret;
41161  __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
41162  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41163  return __ret;
41164 }
41165 #endif
41166 
41167 #ifdef __LITTLE_ENDIAN__
41168 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41169  uint32x4_t __ret;
41170  __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
41171  return __ret;
41172 }
41173 #else
41174 __ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41175  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41176  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41177  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41178  uint32x4_t __ret;
41179  __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
41180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41181  return __ret;
41182 }
41183 #endif
41184 
41185 #ifdef __LITTLE_ENDIAN__
41186 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41187  uint8x16_t __ret;
41188  __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
41189  return __ret;
41190 }
41191 #else
41192 __ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41193  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41194  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41195  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41196  uint8x16_t __ret;
41197  __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
41198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41199  return __ret;
41200 }
41201 #endif
41202 
41203 #ifdef __LITTLE_ENDIAN__
41204 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41205  int16x8_t __ret;
41206  __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
41207  return __ret;
41208 }
41209 #else
41210 __ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41211  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41212  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41213  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41214  int16x8_t __ret;
41215  __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
41216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41217  return __ret;
41218 }
41219 #endif
41220 
41221 #ifdef __LITTLE_ENDIAN__
41222 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41223  int32x4_t __ret;
41224  __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
41225  return __ret;
41226 }
41227 #else
41228 __ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41229  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41230  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41231  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41232  int32x4_t __ret;
41233  __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
41234  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41235  return __ret;
41236 }
41237 #endif
41238 
41239 #ifdef __LITTLE_ENDIAN__
41240 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41241  int8x16_t __ret;
41242  __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
41243  return __ret;
41244 }
41245 #else
41246 __ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41247  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41248  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41249  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41250  int8x16_t __ret;
41251  __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
41252  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41253  return __ret;
41254 }
41255 #endif
41256 
41257 #ifdef __LITTLE_ENDIAN__
41258 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41259  uint16_t __ret;
41260  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__p0);
41261  return __ret;
41262 }
41263 #else
41264 __ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41265  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41266  uint16_t __ret;
41267  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__rev0);
41268  return __ret;
41269 }
41270 #endif
41271 
41272 #ifdef __LITTLE_ENDIAN__
41273 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41274  uint64_t __ret;
41275  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__p0);
41276  return __ret;
41277 }
41278 #else
41279 __ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41280  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41281  uint64_t __ret;
41282  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__rev0);
41283  return __ret;
41284 }
41285 #endif
41286 
41287 #ifdef __LITTLE_ENDIAN__
41288 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41289  uint32_t __ret;
41290  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__p0);
41291  return __ret;
41292 }
41293 #else
41294 __ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41295  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41296  uint32_t __ret;
41297  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__rev0);
41298  return __ret;
41299 }
41300 #endif
41301 
41302 #ifdef __LITTLE_ENDIAN__
41303 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
41304  int16_t __ret;
41305  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__p0);
41306  return __ret;
41307 }
41308 #else
41309 __ai int16_t vaddlvq_s8(int8x16_t __p0) {
41310  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41311  int16_t __ret;
41312  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__rev0);
41313  return __ret;
41314 }
41315 #endif
41316 
41317 #ifdef __LITTLE_ENDIAN__
41318 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
41319  int64_t __ret;
41320  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__p0);
41321  return __ret;
41322 }
41323 #else
41324 __ai int64_t vaddlvq_s32(int32x4_t __p0) {
41325  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41326  int64_t __ret;
41327  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__rev0);
41328  return __ret;
41329 }
41330 #endif
41331 
41332 #ifdef __LITTLE_ENDIAN__
41333 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
41334  int32_t __ret;
41335  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__p0);
41336  return __ret;
41337 }
41338 #else
41339 __ai int32_t vaddlvq_s16(int16x8_t __p0) {
41340  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41341  int32_t __ret;
41342  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__rev0);
41343  return __ret;
41344 }
41345 #endif
41346 
41347 #ifdef __LITTLE_ENDIAN__
41348 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41349  uint16_t __ret;
41350  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__p0);
41351  return __ret;
41352 }
41353 #else
41354 __ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41355  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41356  uint16_t __ret;
41357  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__rev0);
41358  return __ret;
41359 }
41360 #endif
41361 
41362 #ifdef __LITTLE_ENDIAN__
41363 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41364  uint64_t __ret;
41365  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__p0);
41366  return __ret;
41367 }
41368 #else
41369 __ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41370  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41371  uint64_t __ret;
41372  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__rev0);
41373  return __ret;
41374 }
41375 #endif
41376 
41377 #ifdef __LITTLE_ENDIAN__
41378 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41379  uint32_t __ret;
41380  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__p0);
41381  return __ret;
41382 }
41383 #else
41384 __ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41385  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41386  uint32_t __ret;
41387  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__rev0);
41388  return __ret;
41389 }
41390 #endif
41391 
41392 #ifdef __LITTLE_ENDIAN__
41393 __ai int16_t vaddlv_s8(int8x8_t __p0) {
41394  int16_t __ret;
41395  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__p0);
41396  return __ret;
41397 }
41398 #else
41399 __ai int16_t vaddlv_s8(int8x8_t __p0) {
41400  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41401  int16_t __ret;
41402  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__rev0);
41403  return __ret;
41404 }
41405 #endif
41406 
41407 #ifdef __LITTLE_ENDIAN__
41408 __ai int64_t vaddlv_s32(int32x2_t __p0) {
41409  int64_t __ret;
41410  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__p0);
41411  return __ret;
41412 }
41413 #else
41414 __ai int64_t vaddlv_s32(int32x2_t __p0) {
41415  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41416  int64_t __ret;
41417  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__rev0);
41418  return __ret;
41419 }
41420 #endif
41421 
41422 #ifdef __LITTLE_ENDIAN__
41423 __ai int32_t vaddlv_s16(int16x4_t __p0) {
41424  int32_t __ret;
41425  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__p0);
41426  return __ret;
41427 }
41428 #else
41429 __ai int32_t vaddlv_s16(int16x4_t __p0) {
41430  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41431  int32_t __ret;
41432  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__rev0);
41433  return __ret;
41434 }
41435 #endif
41436 
41437 #ifdef __LITTLE_ENDIAN__
41438 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41439  uint8_t __ret;
41440  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__p0);
41441  return __ret;
41442 }
41443 #else
41444 __ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41445  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41446  uint8_t __ret;
41447  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__rev0);
41448  return __ret;
41449 }
41450 #endif
41451 
41452 #ifdef __LITTLE_ENDIAN__
41453 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41454  uint32_t __ret;
41455  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__p0);
41456  return __ret;
41457 }
41458 #else
41459 __ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41460  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41461  uint32_t __ret;
41462  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__rev0);
41463  return __ret;
41464 }
41465 #endif
41466 
41467 #ifdef __LITTLE_ENDIAN__
41468 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41469  uint64_t __ret;
41470  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__p0);
41471  return __ret;
41472 }
41473 #else
41474 __ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41475  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41476  uint64_t __ret;
41477  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__rev0);
41478  return __ret;
41479 }
41480 #endif
41481 
41482 #ifdef __LITTLE_ENDIAN__
41483 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41484  uint16_t __ret;
41485  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__p0);
41486  return __ret;
41487 }
41488 #else
41489 __ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41490  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41491  uint16_t __ret;
41492  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__rev0);
41493  return __ret;
41494 }
41495 #endif
41496 
41497 #ifdef __LITTLE_ENDIAN__
41498 __ai int8_t vaddvq_s8(int8x16_t __p0) {
41499  int8_t __ret;
41500  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__p0);
41501  return __ret;
41502 }
41503 #else
41504 __ai int8_t vaddvq_s8(int8x16_t __p0) {
41505  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41506  int8_t __ret;
41507  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__rev0);
41508  return __ret;
41509 }
41510 #endif
41511 
41512 #ifdef __LITTLE_ENDIAN__
41513 __ai float64_t vaddvq_f64(float64x2_t __p0) {
41514  float64_t __ret;
41515  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__p0);
41516  return __ret;
41517 }
41518 #else
41519 __ai float64_t vaddvq_f64(float64x2_t __p0) {
41520  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41521  float64_t __ret;
41522  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__rev0);
41523  return __ret;
41524 }
41525 #endif
41526 
41527 #ifdef __LITTLE_ENDIAN__
41528 __ai float32_t vaddvq_f32(float32x4_t __p0) {
41529  float32_t __ret;
41530  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__p0);
41531  return __ret;
41532 }
41533 #else
41534 __ai float32_t vaddvq_f32(float32x4_t __p0) {
41535  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41536  float32_t __ret;
41537  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__rev0);
41538  return __ret;
41539 }
41540 #endif
41541 
41542 #ifdef __LITTLE_ENDIAN__
41543 __ai int32_t vaddvq_s32(int32x4_t __p0) {
41544  int32_t __ret;
41545  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__p0);
41546  return __ret;
41547 }
41548 #else
41549 __ai int32_t vaddvq_s32(int32x4_t __p0) {
41550  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41551  int32_t __ret;
41552  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__rev0);
41553  return __ret;
41554 }
41555 #endif
41556 
41557 #ifdef __LITTLE_ENDIAN__
41558 __ai int64_t vaddvq_s64(int64x2_t __p0) {
41559  int64_t __ret;
41560  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__p0);
41561  return __ret;
41562 }
41563 #else
41564 __ai int64_t vaddvq_s64(int64x2_t __p0) {
41565  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41566  int64_t __ret;
41567  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__rev0);
41568  return __ret;
41569 }
41570 #endif
41571 
41572 #ifdef __LITTLE_ENDIAN__
41573 __ai int16_t vaddvq_s16(int16x8_t __p0) {
41574  int16_t __ret;
41575  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__p0);
41576  return __ret;
41577 }
41578 #else
41579 __ai int16_t vaddvq_s16(int16x8_t __p0) {
41580  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41581  int16_t __ret;
41582  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__rev0);
41583  return __ret;
41584 }
41585 #endif
41586 
41587 #ifdef __LITTLE_ENDIAN__
41588 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
41589  uint8_t __ret;
41590  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__p0);
41591  return __ret;
41592 }
41593 #else
41594 __ai uint8_t vaddv_u8(uint8x8_t __p0) {
41595  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41596  uint8_t __ret;
41597  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__rev0);
41598  return __ret;
41599 }
41600 #endif
41601 
41602 #ifdef __LITTLE_ENDIAN__
41603 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
41604  uint32_t __ret;
41605  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__p0);
41606  return __ret;
41607 }
41608 #else
41609 __ai uint32_t vaddv_u32(uint32x2_t __p0) {
41610  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41611  uint32_t __ret;
41612  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__rev0);
41613  return __ret;
41614 }
41615 #endif
41616 
41617 #ifdef __LITTLE_ENDIAN__
41618 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
41619  uint16_t __ret;
41620  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__p0);
41621  return __ret;
41622 }
41623 #else
41624 __ai uint16_t vaddv_u16(uint16x4_t __p0) {
41625  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41626  uint16_t __ret;
41627  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__rev0);
41628  return __ret;
41629 }
41630 #endif
41631 
41632 #ifdef __LITTLE_ENDIAN__
41633 __ai int8_t vaddv_s8(int8x8_t __p0) {
41634  int8_t __ret;
41635  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__p0);
41636  return __ret;
41637 }
41638 #else
41639 __ai int8_t vaddv_s8(int8x8_t __p0) {
41640  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41641  int8_t __ret;
41642  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__rev0);
41643  return __ret;
41644 }
41645 #endif
41646 
41647 #ifdef __LITTLE_ENDIAN__
41648 __ai float32_t vaddv_f32(float32x2_t __p0) {
41649  float32_t __ret;
41650  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__p0);
41651  return __ret;
41652 }
41653 #else
41654 __ai float32_t vaddv_f32(float32x2_t __p0) {
41655  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41656  float32_t __ret;
41657  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__rev0);
41658  return __ret;
41659 }
41660 #endif
41661 
41662 #ifdef __LITTLE_ENDIAN__
41663 __ai int32_t vaddv_s32(int32x2_t __p0) {
41664  int32_t __ret;
41665  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__p0);
41666  return __ret;
41667 }
41668 #else
41669 __ai int32_t vaddv_s32(int32x2_t __p0) {
41670  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41671  int32_t __ret;
41672  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__rev0);
41673  return __ret;
41674 }
41675 #endif
41676 
41677 #ifdef __LITTLE_ENDIAN__
41678 __ai int16_t vaddv_s16(int16x4_t __p0) {
41679  int16_t __ret;
41680  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__p0);
41681  return __ret;
41682 }
41683 #else
41684 __ai int16_t vaddv_s16(int16x4_t __p0) {
41685  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41686  int16_t __ret;
41687  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__rev0);
41688  return __ret;
41689 }
41690 #endif
41691 
41692 #ifdef __LITTLE_ENDIAN__
41693 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41694  poly64x1_t __ret;
41695  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41696  return __ret;
41697 }
41698 #else
41699 __ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41700  poly64x1_t __ret;
41701  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41702  return __ret;
41703 }
41704 #endif
41705 
41706 #ifdef __LITTLE_ENDIAN__
41707 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41708  poly64x2_t __ret;
41709  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
41710  return __ret;
41711 }
41712 #else
41713 __ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41714  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41715  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41716  poly64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41717  poly64x2_t __ret;
41718  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
41719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41720  return __ret;
41721 }
41722 #endif
41723 
41724 #ifdef __LITTLE_ENDIAN__
41725 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41726  float64x2_t __ret;
41727  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
41728  return __ret;
41729 }
41730 #else
41731 __ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41732  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41733  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41734  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41735  float64x2_t __ret;
41736  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
41737  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41738  return __ret;
41739 }
41740 #endif
41741 
41742 #ifdef __LITTLE_ENDIAN__
41743 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41744  float64x1_t __ret;
41745  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41746  return __ret;
41747 }
41748 #else
41749 __ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41750  float64x1_t __ret;
41751  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41752  return __ret;
41753 }
41754 #endif
41755 
41756 #ifdef __LITTLE_ENDIAN__
41757 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41758  uint64x2_t __ret;
41759  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41760  return __ret;
41761 }
41762 #else
41763 __ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41764  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41765  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41766  uint64x2_t __ret;
41767  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41768  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41769  return __ret;
41770 }
41771 #endif
41772 
41773 #ifdef __LITTLE_ENDIAN__
41774 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41775  uint64x1_t __ret;
41776  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41777  return __ret;
41778 }
41779 #else
41780 __ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41781  uint64x1_t __ret;
41782  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41783  return __ret;
41784 }
41785 #endif
41786 
41787 #ifdef __LITTLE_ENDIAN__
41788 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41789  uint64_t __ret;
41790  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41791  return __ret;
41792 }
41793 #else
41794 __ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41795  uint64_t __ret;
41796  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41797  return __ret;
41798 }
41799 #endif
41800 
41801 #ifdef __LITTLE_ENDIAN__
41802 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41803  uint32_t __ret;
41804  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41805  return __ret;
41806 }
41807 #else
41808 __ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41809  uint32_t __ret;
41810  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41811  return __ret;
41812 }
41813 #endif
41814 
41815 #ifdef __LITTLE_ENDIAN__
41816 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41817  uint64x2_t __ret;
41818  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41819  return __ret;
41820 }
41821 #else
41822 __ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41823  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41824  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41825  uint64x2_t __ret;
41826  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41828  return __ret;
41829 }
41830 #endif
41831 
41832 #ifdef __LITTLE_ENDIAN__
41833 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41834  uint64x1_t __ret;
41835  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41836  return __ret;
41837 }
41838 #else
41839 __ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41840  uint64x1_t __ret;
41841  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41842  return __ret;
41843 }
41844 #endif
41845 
41846 #ifdef __LITTLE_ENDIAN__
41847 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41848  uint64_t __ret;
41849  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41850  return __ret;
41851 }
41852 #else
41853 __ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41854  uint64_t __ret;
41855  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41856  return __ret;
41857 }
41858 #endif
41859 
41860 #ifdef __LITTLE_ENDIAN__
41861 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41862  uint32_t __ret;
41863  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41864  return __ret;
41865 }
41866 #else
41867 __ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41868  uint32_t __ret;
41869  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41870  return __ret;
41871 }
41872 #endif
41873 
41874 #ifdef __LITTLE_ENDIAN__
41875 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41876  uint64x2_t __ret;
41877  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41878  return __ret;
41879 }
41880 #else
41881 __ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41882  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41883  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41884  uint64x2_t __ret;
41885  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41886  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41887  return __ret;
41888 }
41889 #endif
41890 
41891 #ifdef __LITTLE_ENDIAN__
41892 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41893  uint64x1_t __ret;
41894  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41895  return __ret;
41896 }
41897 #else
41898 __ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41899  uint64x1_t __ret;
41900  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41901  return __ret;
41902 }
41903 #endif
41904 
41905 #ifdef __LITTLE_ENDIAN__
41906 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41907  uint64_t __ret;
41908  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41909  return __ret;
41910 }
41911 #else
41912 __ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41913  uint64_t __ret;
41914  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41915  return __ret;
41916 }
41917 #endif
41918 
41919 #ifdef __LITTLE_ENDIAN__
41920 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41921  uint32_t __ret;
41922  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41923  return __ret;
41924 }
41925 #else
41926 __ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41927  uint32_t __ret;
41928  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41929  return __ret;
41930 }
41931 #endif
41932 
41933 #ifdef __LITTLE_ENDIAN__
41934 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41935  uint64x2_t __ret;
41936  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41937  return __ret;
41938 }
41939 #else
41940 __ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41941  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41942  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41943  uint64x2_t __ret;
41944  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41945  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41946  return __ret;
41947 }
41948 #endif
41949 
41950 #ifdef __LITTLE_ENDIAN__
41951 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41952  uint64x1_t __ret;
41953  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41954  return __ret;
41955 }
41956 #else
41957 __ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41958  uint64x1_t __ret;
41959  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41960  return __ret;
41961 }
41962 #endif
41963 
41964 #ifdef __LITTLE_ENDIAN__
41965 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
41966  uint64_t __ret;
41967  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
41968  return __ret;
41969 }
41970 #else
41971 __ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
41972  uint64_t __ret;
41973  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
41974  return __ret;
41975 }
41976 #endif
41977 
41978 #ifdef __LITTLE_ENDIAN__
41979 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
41980  uint32_t __ret;
41981  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
41982  return __ret;
41983 }
41984 #else
41985 __ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
41986  uint32_t __ret;
41987  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
41988  return __ret;
41989 }
41990 #endif
41991 
41992 #ifdef __LITTLE_ENDIAN__
41993 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
41994  uint64x1_t __ret;
41995  __ret = (uint64x1_t)(__p0 == __p1);
41996  return __ret;
41997 }
41998 #else
41999 __ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
42000  uint64x1_t __ret;
42001  __ret = (uint64x1_t)(__p0 == __p1);
42002  return __ret;
42003 }
42004 #endif
42005 
42006 #ifdef __LITTLE_ENDIAN__
42007 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42008  uint64x2_t __ret;
42009  __ret = (uint64x2_t)(__p0 == __p1);
42010  return __ret;
42011 }
42012 #else
42013 __ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42014  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42015  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42016  uint64x2_t __ret;
42017  __ret = (uint64x2_t)(__rev0 == __rev1);
42018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42019  return __ret;
42020 }
42021 #endif
42022 
42023 #ifdef __LITTLE_ENDIAN__
42024 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42025  uint64x2_t __ret;
42026  __ret = (uint64x2_t)(__p0 == __p1);
42027  return __ret;
42028 }
42029 #else
42030 __ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42031  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42032  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42033  uint64x2_t __ret;
42034  __ret = (uint64x2_t)(__rev0 == __rev1);
42035  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42036  return __ret;
42037 }
42038 #endif
42039 
42040 #ifdef __LITTLE_ENDIAN__
42041 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42042  uint64x2_t __ret;
42043  __ret = (uint64x2_t)(__p0 == __p1);
42044  return __ret;
42045 }
42046 #else
42047 __ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42048  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42049  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42050  uint64x2_t __ret;
42051  __ret = (uint64x2_t)(__rev0 == __rev1);
42052  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42053  return __ret;
42054 }
42055 #endif
42056 
42057 #ifdef __LITTLE_ENDIAN__
42058 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42059  uint64x2_t __ret;
42060  __ret = (uint64x2_t)(__p0 == __p1);
42061  return __ret;
42062 }
42063 #else
42064 __ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42065  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42066  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42067  uint64x2_t __ret;
42068  __ret = (uint64x2_t)(__rev0 == __rev1);
42069  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42070  return __ret;
42071 }
42072 #endif
42073 
42074 #ifdef __LITTLE_ENDIAN__
42075 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42076  uint64x1_t __ret;
42077  __ret = (uint64x1_t)(__p0 == __p1);
42078  return __ret;
42079 }
42080 #else
42081 __ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42082  uint64x1_t __ret;
42083  __ret = (uint64x1_t)(__p0 == __p1);
42084  return __ret;
42085 }
42086 #endif
42087 
42088 #ifdef __LITTLE_ENDIAN__
42089 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42090  uint64x1_t __ret;
42091  __ret = (uint64x1_t)(__p0 == __p1);
42092  return __ret;
42093 }
42094 #else
42095 __ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42096  uint64x1_t __ret;
42097  __ret = (uint64x1_t)(__p0 == __p1);
42098  return __ret;
42099 }
42100 #endif
42101 
42102 #ifdef __LITTLE_ENDIAN__
42103 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42104  uint64x1_t __ret;
42105  __ret = (uint64x1_t)(__p0 == __p1);
42106  return __ret;
42107 }
42108 #else
42109 __ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42110  uint64x1_t __ret;
42111  __ret = (uint64x1_t)(__p0 == __p1);
42112  return __ret;
42113 }
42114 #endif
42115 
42116 #ifdef __LITTLE_ENDIAN__
42117 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42118  uint64_t __ret;
42119  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42120  return __ret;
42121 }
42122 #else
42123 __ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42124  uint64_t __ret;
42125  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42126  return __ret;
42127 }
42128 #endif
42129 
42130 #ifdef __LITTLE_ENDIAN__
42131 __ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42132  int64_t __ret;
42133  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42134  return __ret;
42135 }
42136 #else
42137 __ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42138  int64_t __ret;
42139  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42140  return __ret;
42141 }
42142 #endif
42143 
42144 #ifdef __LITTLE_ENDIAN__
42145 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42146  uint64_t __ret;
42147  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42148  return __ret;
42149 }
42150 #else
42151 __ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42152  uint64_t __ret;
42153  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42154  return __ret;
42155 }
42156 #endif
42157 
42158 #ifdef __LITTLE_ENDIAN__
42159 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42160  uint32_t __ret;
42161  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42162  return __ret;
42163 }
42164 #else
42165 __ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42166  uint32_t __ret;
42167  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42168  return __ret;
42169 }
42170 #endif
42171 
42172 #ifdef __LITTLE_ENDIAN__
42173 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42174  uint8x8_t __ret;
42175  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42176  return __ret;
42177 }
42178 #else
42179 __ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42180  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42181  uint8x8_t __ret;
42182  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42183  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42184  return __ret;
42185 }
42186 #endif
42187 
42188 #ifdef __LITTLE_ENDIAN__
42189 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42190  uint64x1_t __ret;
42191  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42192  return __ret;
42193 }
42194 #else
42195 __ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42196  uint64x1_t __ret;
42197  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42198  return __ret;
42199 }
42200 #endif
42201 
42202 #ifdef __LITTLE_ENDIAN__
42203 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42204  uint16x4_t __ret;
42205  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42206  return __ret;
42207 }
42208 #else
42209 __ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42210  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42211  uint16x4_t __ret;
42212  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42213  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42214  return __ret;
42215 }
42216 #endif
42217 
42218 #ifdef __LITTLE_ENDIAN__
42219 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42220  uint8x16_t __ret;
42221  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42222  return __ret;
42223 }
42224 #else
42225 __ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42226  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42227  uint8x16_t __ret;
42228  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42229  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42230  return __ret;
42231 }
42232 #endif
42233 
42234 #ifdef __LITTLE_ENDIAN__
42235 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42236  uint64x2_t __ret;
42237  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42238  return __ret;
42239 }
42240 #else
42241 __ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42242  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42243  uint64x2_t __ret;
42244  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42245  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42246  return __ret;
42247 }
42248 #endif
42249 
42250 #ifdef __LITTLE_ENDIAN__
42251 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42252  uint16x8_t __ret;
42253  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42254  return __ret;
42255 }
42256 #else
42257 __ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42258  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42259  uint16x8_t __ret;
42260  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42261  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42262  return __ret;
42263 }
42264 #endif
42265 
42266 #ifdef __LITTLE_ENDIAN__
42267 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42268  uint8x16_t __ret;
42269  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42270  return __ret;
42271 }
42272 #else
42273 __ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42274  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42275  uint8x16_t __ret;
42276  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42277  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42278  return __ret;
42279 }
42280 #endif
42281 
42282 #ifdef __LITTLE_ENDIAN__
42283 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42284  uint32x4_t __ret;
42285  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42286  return __ret;
42287 }
42288 #else
42289 __ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42290  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42291  uint32x4_t __ret;
42292  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42293  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42294  return __ret;
42295 }
42296 #endif
42297 
42298 #ifdef __LITTLE_ENDIAN__
42299 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42300  uint64x2_t __ret;
42301  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42302  return __ret;
42303 }
42304 #else
42305 __ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42306  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42307  uint64x2_t __ret;
42308  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42310  return __ret;
42311 }
42312 #endif
42313 
42314 #ifdef __LITTLE_ENDIAN__
42315 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42316  uint16x8_t __ret;
42317  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42318  return __ret;
42319 }
42320 #else
42321 __ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42322  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42323  uint16x8_t __ret;
42324  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42325  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42326  return __ret;
42327 }
42328 #endif
42329 
42330 #ifdef __LITTLE_ENDIAN__
42331 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42332  uint8x16_t __ret;
42333  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42334  return __ret;
42335 }
42336 #else
42337 __ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42338  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42339  uint8x16_t __ret;
42340  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42341  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42342  return __ret;
42343 }
42344 #endif
42345 
42346 #ifdef __LITTLE_ENDIAN__
42347 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42348  uint64x2_t __ret;
42349  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42350  return __ret;
42351 }
42352 #else
42353 __ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42354  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42355  uint64x2_t __ret;
42356  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42357  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42358  return __ret;
42359 }
42360 #endif
42361 
42362 #ifdef __LITTLE_ENDIAN__
42363 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42364  uint32x4_t __ret;
42365  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42366  return __ret;
42367 }
42368 #else
42369 __ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42370  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42371  uint32x4_t __ret;
42372  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42374  return __ret;
42375 }
42376 #endif
42377 
42378 #ifdef __LITTLE_ENDIAN__
42379 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42380  uint32x4_t __ret;
42381  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42382  return __ret;
42383 }
42384 #else
42385 __ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42386  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42387  uint32x4_t __ret;
42388  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42389  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42390  return __ret;
42391 }
42392 #endif
42393 
42394 #ifdef __LITTLE_ENDIAN__
42395 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42396  uint64x2_t __ret;
42397  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42398  return __ret;
42399 }
42400 #else
42401 __ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42402  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42403  uint64x2_t __ret;
42404  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42406  return __ret;
42407 }
42408 #endif
42409 
42410 #ifdef __LITTLE_ENDIAN__
42411 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42412  uint16x8_t __ret;
42413  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42414  return __ret;
42415 }
42416 #else
42417 __ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42418  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42419  uint16x8_t __ret;
42420  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42422  return __ret;
42423 }
42424 #endif
42425 
42426 #ifdef __LITTLE_ENDIAN__
42427 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42428  uint8x8_t __ret;
42429  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42430  return __ret;
42431 }
42432 #else
42433 __ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42434  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42435  uint8x8_t __ret;
42436  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42437  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42438  return __ret;
42439 }
42440 #endif
42441 
42442 #ifdef __LITTLE_ENDIAN__
42443 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42444  uint32x2_t __ret;
42445  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42446  return __ret;
42447 }
42448 #else
42449 __ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42450  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42451  uint32x2_t __ret;
42452  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42454  return __ret;
42455 }
42456 #endif
42457 
42458 #ifdef __LITTLE_ENDIAN__
42459 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42460  uint64x1_t __ret;
42461  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42462  return __ret;
42463 }
42464 #else
42465 __ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42466  uint64x1_t __ret;
42467  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42468  return __ret;
42469 }
42470 #endif
42471 
42472 #ifdef __LITTLE_ENDIAN__
42473 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42474  uint16x4_t __ret;
42475  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42476  return __ret;
42477 }
42478 #else
42479 __ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42480  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42481  uint16x4_t __ret;
42482  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42483  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42484  return __ret;
42485 }
42486 #endif
42487 
42488 #ifdef __LITTLE_ENDIAN__
42489 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42490  uint8x8_t __ret;
42491  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42492  return __ret;
42493 }
42494 #else
42495 __ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42496  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42497  uint8x8_t __ret;
42498  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42500  return __ret;
42501 }
42502 #endif
42503 
42504 #ifdef __LITTLE_ENDIAN__
42505 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42506  uint64x1_t __ret;
42507  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42508  return __ret;
42509 }
42510 #else
42511 __ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42512  uint64x1_t __ret;
42513  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42514  return __ret;
42515 }
42516 #endif
42517 
42518 #ifdef __LITTLE_ENDIAN__
42519 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42520  uint32x2_t __ret;
42521  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42522  return __ret;
42523 }
42524 #else
42525 __ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42526  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42527  uint32x2_t __ret;
42528  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42529  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42530  return __ret;
42531 }
42532 #endif
42533 
42534 #ifdef __LITTLE_ENDIAN__
42535 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42536  uint32x2_t __ret;
42537  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42538  return __ret;
42539 }
42540 #else
42541 __ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42542  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42543  uint32x2_t __ret;
42544  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42546  return __ret;
42547 }
42548 #endif
42549 
42550 #ifdef __LITTLE_ENDIAN__
42551 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42552  uint64x1_t __ret;
42553  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42554  return __ret;
42555 }
42556 #else
42557 __ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42558  uint64x1_t __ret;
42559  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42560  return __ret;
42561 }
42562 #endif
42563 
42564 #ifdef __LITTLE_ENDIAN__
42565 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42566  uint16x4_t __ret;
42567  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42568  return __ret;
42569 }
42570 #else
42571 __ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42572  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42573  uint16x4_t __ret;
42574  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42575  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42576  return __ret;
42577 }
42578 #endif
42579 
42580 #ifdef __LITTLE_ENDIAN__
42581 __ai uint64_t vceqzd_u64(uint64_t __p0) {
42582  uint64_t __ret;
42583  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42584  return __ret;
42585 }
42586 #else
42587 __ai uint64_t vceqzd_u64(uint64_t __p0) {
42588  uint64_t __ret;
42589  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42590  return __ret;
42591 }
42592 #endif
42593 
42594 #ifdef __LITTLE_ENDIAN__
42595 __ai int64_t vceqzd_s64(int64_t __p0) {
42596  int64_t __ret;
42597  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42598  return __ret;
42599 }
42600 #else
42601 __ai int64_t vceqzd_s64(int64_t __p0) {
42602  int64_t __ret;
42603  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42604  return __ret;
42605 }
42606 #endif
42607 
42608 #ifdef __LITTLE_ENDIAN__
42609 __ai uint64_t vceqzd_f64(float64_t __p0) {
42610  uint64_t __ret;
42611  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42612  return __ret;
42613 }
42614 #else
42615 __ai uint64_t vceqzd_f64(float64_t __p0) {
42616  uint64_t __ret;
42617  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42618  return __ret;
42619 }
42620 #endif
42621 
42622 #ifdef __LITTLE_ENDIAN__
42623 __ai uint32_t vceqzs_f32(float32_t __p0) {
42624  uint32_t __ret;
42625  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42626  return __ret;
42627 }
42628 #else
42629 __ai uint32_t vceqzs_f32(float32_t __p0) {
42630  uint32_t __ret;
42631  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42632  return __ret;
42633 }
42634 #endif
42635 
42636 #ifdef __LITTLE_ENDIAN__
42637 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42638  uint64x2_t __ret;
42639  __ret = (uint64x2_t)(__p0 >= __p1);
42640  return __ret;
42641 }
42642 #else
42643 __ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42644  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42645  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42646  uint64x2_t __ret;
42647  __ret = (uint64x2_t)(__rev0 >= __rev1);
42648  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42649  return __ret;
42650 }
42651 #endif
42652 
42653 #ifdef __LITTLE_ENDIAN__
42654 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42655  uint64x2_t __ret;
42656  __ret = (uint64x2_t)(__p0 >= __p1);
42657  return __ret;
42658 }
42659 #else
42660 __ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42661  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42662  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42663  uint64x2_t __ret;
42664  __ret = (uint64x2_t)(__rev0 >= __rev1);
42665  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42666  return __ret;
42667 }
42668 #endif
42669 
42670 #ifdef __LITTLE_ENDIAN__
42671 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42672  uint64x2_t __ret;
42673  __ret = (uint64x2_t)(__p0 >= __p1);
42674  return __ret;
42675 }
42676 #else
42677 __ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42678  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42679  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42680  uint64x2_t __ret;
42681  __ret = (uint64x2_t)(__rev0 >= __rev1);
42682  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42683  return __ret;
42684 }
42685 #endif
42686 
42687 #ifdef __LITTLE_ENDIAN__
42688 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42689  uint64x1_t __ret;
42690  __ret = (uint64x1_t)(__p0 >= __p1);
42691  return __ret;
42692 }
42693 #else
42694 __ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42695  uint64x1_t __ret;
42696  __ret = (uint64x1_t)(__p0 >= __p1);
42697  return __ret;
42698 }
42699 #endif
42700 
42701 #ifdef __LITTLE_ENDIAN__
42702 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42703  uint64x1_t __ret;
42704  __ret = (uint64x1_t)(__p0 >= __p1);
42705  return __ret;
42706 }
42707 #else
42708 __ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42709  uint64x1_t __ret;
42710  __ret = (uint64x1_t)(__p0 >= __p1);
42711  return __ret;
42712 }
42713 #endif
42714 
42715 #ifdef __LITTLE_ENDIAN__
42716 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42717  uint64x1_t __ret;
42718  __ret = (uint64x1_t)(__p0 >= __p1);
42719  return __ret;
42720 }
42721 #else
42722 __ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42723  uint64x1_t __ret;
42724  __ret = (uint64x1_t)(__p0 >= __p1);
42725  return __ret;
42726 }
42727 #endif
42728 
42729 #ifdef __LITTLE_ENDIAN__
42730 __ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42731  int64_t __ret;
42732  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42733  return __ret;
42734 }
42735 #else
42736 __ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42737  int64_t __ret;
42738  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42739  return __ret;
42740 }
42741 #endif
42742 
42743 #ifdef __LITTLE_ENDIAN__
42744 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42745  uint64_t __ret;
42746  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42747  return __ret;
42748 }
42749 #else
42750 __ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42751  uint64_t __ret;
42752  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42753  return __ret;
42754 }
42755 #endif
42756 
42757 #ifdef __LITTLE_ENDIAN__
42758 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42759  uint64_t __ret;
42760  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42761  return __ret;
42762 }
42763 #else
42764 __ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42765  uint64_t __ret;
42766  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42767  return __ret;
42768 }
42769 #endif
42770 
42771 #ifdef __LITTLE_ENDIAN__
42772 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42773  uint32_t __ret;
42774  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42775  return __ret;
42776 }
42777 #else
42778 __ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42779  uint32_t __ret;
42780  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42781  return __ret;
42782 }
42783 #endif
42784 
42785 #ifdef __LITTLE_ENDIAN__
42786 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42787  uint8x16_t __ret;
42788  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
42789  return __ret;
42790 }
42791 #else
42792 __ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42793  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42794  uint8x16_t __ret;
42795  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
42796  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42797  return __ret;
42798 }
42799 #endif
42800 
42801 #ifdef __LITTLE_ENDIAN__
42802 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42803  uint64x2_t __ret;
42804  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42805  return __ret;
42806 }
42807 #else
42808 __ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42809  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42810  uint64x2_t __ret;
42811  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42812  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42813  return __ret;
42814 }
42815 #endif
42816 
42817 #ifdef __LITTLE_ENDIAN__
42818 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42819  uint32x4_t __ret;
42820  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42821  return __ret;
42822 }
42823 #else
42824 __ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42825  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42826  uint32x4_t __ret;
42827  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42828  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42829  return __ret;
42830 }
42831 #endif
42832 
42833 #ifdef __LITTLE_ENDIAN__
42834 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42835  uint32x4_t __ret;
42836  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42837  return __ret;
42838 }
42839 #else
42840 __ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42841  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42842  uint32x4_t __ret;
42843  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42845  return __ret;
42846 }
42847 #endif
42848 
42849 #ifdef __LITTLE_ENDIAN__
42850 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42851  uint64x2_t __ret;
42852  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42853  return __ret;
42854 }
42855 #else
42856 __ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42857  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42858  uint64x2_t __ret;
42859  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42860  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42861  return __ret;
42862 }
42863 #endif
42864 
42865 #ifdef __LITTLE_ENDIAN__
42866 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42867  uint16x8_t __ret;
42868  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
42869  return __ret;
42870 }
42871 #else
42872 __ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42873  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42874  uint16x8_t __ret;
42875  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
42876  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42877  return __ret;
42878 }
42879 #endif
42880 
42881 #ifdef __LITTLE_ENDIAN__
42882 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42883  uint8x8_t __ret;
42884  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
42885  return __ret;
42886 }
42887 #else
42888 __ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42889  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42890  uint8x8_t __ret;
42891  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
42892  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42893  return __ret;
42894 }
42895 #endif
42896 
42897 #ifdef __LITTLE_ENDIAN__
42898 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42899  uint64x1_t __ret;
42900  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42901  return __ret;
42902 }
42903 #else
42904 __ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42905  uint64x1_t __ret;
42906  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42907  return __ret;
42908 }
42909 #endif
42910 
42911 #ifdef __LITTLE_ENDIAN__
42912 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42913  uint32x2_t __ret;
42914  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42915  return __ret;
42916 }
42917 #else
42918 __ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42919  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42920  uint32x2_t __ret;
42921  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42922  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42923  return __ret;
42924 }
42925 #endif
42926 
42927 #ifdef __LITTLE_ENDIAN__
42928 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42929  uint32x2_t __ret;
42930  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42931  return __ret;
42932 }
42933 #else
42934 __ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42935  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42936  uint32x2_t __ret;
42937  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42938  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42939  return __ret;
42940 }
42941 #endif
42942 
42943 #ifdef __LITTLE_ENDIAN__
42944 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42945  uint64x1_t __ret;
42946  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42947  return __ret;
42948 }
42949 #else
42950 __ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42951  uint64x1_t __ret;
42952  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42953  return __ret;
42954 }
42955 #endif
42956 
42957 #ifdef __LITTLE_ENDIAN__
42958 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42959  uint16x4_t __ret;
42960  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
42961  return __ret;
42962 }
42963 #else
42964 __ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42965  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42966  uint16x4_t __ret;
42967  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
42968  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42969  return __ret;
42970 }
42971 #endif
42972 
42973 #ifdef __LITTLE_ENDIAN__
42974 __ai int64_t vcgezd_s64(int64_t __p0) {
42975  int64_t __ret;
42976  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
42977  return __ret;
42978 }
42979 #else
42980 __ai int64_t vcgezd_s64(int64_t __p0) {
42981  int64_t __ret;
42982  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
42983  return __ret;
42984 }
42985 #endif
42986 
42987 #ifdef __LITTLE_ENDIAN__
42988 __ai uint64_t vcgezd_f64(float64_t __p0) {
42989  uint64_t __ret;
42990  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
42991  return __ret;
42992 }
42993 #else
42994 __ai uint64_t vcgezd_f64(float64_t __p0) {
42995  uint64_t __ret;
42996  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
42997  return __ret;
42998 }
42999 #endif
43000 
43001 #ifdef __LITTLE_ENDIAN__
43002 __ai uint32_t vcgezs_f32(float32_t __p0) {
43003  uint32_t __ret;
43004  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43005  return __ret;
43006 }
43007 #else
43008 __ai uint32_t vcgezs_f32(float32_t __p0) {
43009  uint32_t __ret;
43010  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43011  return __ret;
43012 }
43013 #endif
43014 
43015 #ifdef __LITTLE_ENDIAN__
43016 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43017  uint64x2_t __ret;
43018  __ret = (uint64x2_t)(__p0 > __p1);
43019  return __ret;
43020 }
43021 #else
43022 __ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43023  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43024  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43025  uint64x2_t __ret;
43026  __ret = (uint64x2_t)(__rev0 > __rev1);
43027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43028  return __ret;
43029 }
43030 #endif
43031 
43032 #ifdef __LITTLE_ENDIAN__
43033 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43034  uint64x2_t __ret;
43035  __ret = (uint64x2_t)(__p0 > __p1);
43036  return __ret;
43037 }
43038 #else
43039 __ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43040  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43041  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43042  uint64x2_t __ret;
43043  __ret = (uint64x2_t)(__rev0 > __rev1);
43044  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43045  return __ret;
43046 }
43047 #endif
43048 
43049 #ifdef __LITTLE_ENDIAN__
43050 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43051  uint64x2_t __ret;
43052  __ret = (uint64x2_t)(__p0 > __p1);
43053  return __ret;
43054 }
43055 #else
43056 __ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43057  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43058  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43059  uint64x2_t __ret;
43060  __ret = (uint64x2_t)(__rev0 > __rev1);
43061  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43062  return __ret;
43063 }
43064 #endif
43065 
43066 #ifdef __LITTLE_ENDIAN__
43067 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43068  uint64x1_t __ret;
43069  __ret = (uint64x1_t)(__p0 > __p1);
43070  return __ret;
43071 }
43072 #else
43073 __ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43074  uint64x1_t __ret;
43075  __ret = (uint64x1_t)(__p0 > __p1);
43076  return __ret;
43077 }
43078 #endif
43079 
43080 #ifdef __LITTLE_ENDIAN__
43081 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43082  uint64x1_t __ret;
43083  __ret = (uint64x1_t)(__p0 > __p1);
43084  return __ret;
43085 }
43086 #else
43087 __ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43088  uint64x1_t __ret;
43089  __ret = (uint64x1_t)(__p0 > __p1);
43090  return __ret;
43091 }
43092 #endif
43093 
43094 #ifdef __LITTLE_ENDIAN__
43095 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43096  uint64x1_t __ret;
43097  __ret = (uint64x1_t)(__p0 > __p1);
43098  return __ret;
43099 }
43100 #else
43101 __ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43102  uint64x1_t __ret;
43103  __ret = (uint64x1_t)(__p0 > __p1);
43104  return __ret;
43105 }
43106 #endif
43107 
43108 #ifdef __LITTLE_ENDIAN__
43109 __ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43110  int64_t __ret;
43111  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43112  return __ret;
43113 }
43114 #else
43115 __ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43116  int64_t __ret;
43117  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43118  return __ret;
43119 }
43120 #endif
43121 
43122 #ifdef __LITTLE_ENDIAN__
43123 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43124  uint64_t __ret;
43125  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43126  return __ret;
43127 }
43128 #else
43129 __ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43130  uint64_t __ret;
43131  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43132  return __ret;
43133 }
43134 #endif
43135 
43136 #ifdef __LITTLE_ENDIAN__
43137 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43138  uint64_t __ret;
43139  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43140  return __ret;
43141 }
43142 #else
43143 __ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43144  uint64_t __ret;
43145  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43146  return __ret;
43147 }
43148 #endif
43149 
43150 #ifdef __LITTLE_ENDIAN__
43151 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43152  uint32_t __ret;
43153  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43154  return __ret;
43155 }
43156 #else
43157 __ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43158  uint32_t __ret;
43159  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43160  return __ret;
43161 }
43162 #endif
43163 
43164 #ifdef __LITTLE_ENDIAN__
43165 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43166  uint8x16_t __ret;
43167  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
43168  return __ret;
43169 }
43170 #else
43171 __ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43172  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43173  uint8x16_t __ret;
43174  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
43175  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43176  return __ret;
43177 }
43178 #endif
43179 
43180 #ifdef __LITTLE_ENDIAN__
43181 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43182  uint64x2_t __ret;
43183  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43184  return __ret;
43185 }
43186 #else
43187 __ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43188  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43189  uint64x2_t __ret;
43190  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43191  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43192  return __ret;
43193 }
43194 #endif
43195 
43196 #ifdef __LITTLE_ENDIAN__
43197 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43198  uint32x4_t __ret;
43199  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43200  return __ret;
43201 }
43202 #else
43203 __ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43204  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43205  uint32x4_t __ret;
43206  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43207  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43208  return __ret;
43209 }
43210 #endif
43211 
43212 #ifdef __LITTLE_ENDIAN__
43213 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43214  uint32x4_t __ret;
43215  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43216  return __ret;
43217 }
43218 #else
43219 __ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43220  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43221  uint32x4_t __ret;
43222  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43223  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43224  return __ret;
43225 }
43226 #endif
43227 
43228 #ifdef __LITTLE_ENDIAN__
43229 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43230  uint64x2_t __ret;
43231  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43232  return __ret;
43233 }
43234 #else
43235 __ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43236  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43237  uint64x2_t __ret;
43238  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43240  return __ret;
43241 }
43242 #endif
43243 
43244 #ifdef __LITTLE_ENDIAN__
43245 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43246  uint16x8_t __ret;
43247  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
43248  return __ret;
43249 }
43250 #else
43251 __ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43252  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43253  uint16x8_t __ret;
43254  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
43255  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43256  return __ret;
43257 }
43258 #endif
43259 
43260 #ifdef __LITTLE_ENDIAN__
43261 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43262  uint8x8_t __ret;
43263  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
43264  return __ret;
43265 }
43266 #else
43267 __ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43268  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43269  uint8x8_t __ret;
43270  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
43271  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43272  return __ret;
43273 }
43274 #endif
43275 
43276 #ifdef __LITTLE_ENDIAN__
43277 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43278  uint64x1_t __ret;
43279  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43280  return __ret;
43281 }
43282 #else
43283 __ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43284  uint64x1_t __ret;
43285  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43286  return __ret;
43287 }
43288 #endif
43289 
43290 #ifdef __LITTLE_ENDIAN__
43291 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43292  uint32x2_t __ret;
43293  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43294  return __ret;
43295 }
43296 #else
43297 __ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43298  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43299  uint32x2_t __ret;
43300  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43301  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43302  return __ret;
43303 }
43304 #endif
43305 
43306 #ifdef __LITTLE_ENDIAN__
43307 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43308  uint32x2_t __ret;
43309  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43310  return __ret;
43311 }
43312 #else
43313 __ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43314  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43315  uint32x2_t __ret;
43316  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43317  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43318  return __ret;
43319 }
43320 #endif
43321 
43322 #ifdef __LITTLE_ENDIAN__
43323 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43324  uint64x1_t __ret;
43325  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43326  return __ret;
43327 }
43328 #else
43329 __ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43330  uint64x1_t __ret;
43331  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43332  return __ret;
43333 }
43334 #endif
43335 
43336 #ifdef __LITTLE_ENDIAN__
43337 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43338  uint16x4_t __ret;
43339  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
43340  return __ret;
43341 }
43342 #else
43343 __ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43344  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43345  uint16x4_t __ret;
43346  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
43347  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43348  return __ret;
43349 }
43350 #endif
43351 
43352 #ifdef __LITTLE_ENDIAN__
43353 __ai int64_t vcgtzd_s64(int64_t __p0) {
43354  int64_t __ret;
43355  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43356  return __ret;
43357 }
43358 #else
43359 __ai int64_t vcgtzd_s64(int64_t __p0) {
43360  int64_t __ret;
43361  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43362  return __ret;
43363 }
43364 #endif
43365 
43366 #ifdef __LITTLE_ENDIAN__
43367 __ai uint64_t vcgtzd_f64(float64_t __p0) {
43368  uint64_t __ret;
43369  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43370  return __ret;
43371 }
43372 #else
43373 __ai uint64_t vcgtzd_f64(float64_t __p0) {
43374  uint64_t __ret;
43375  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43376  return __ret;
43377 }
43378 #endif
43379 
43380 #ifdef __LITTLE_ENDIAN__
43381 __ai uint32_t vcgtzs_f32(float32_t __p0) {
43382  uint32_t __ret;
43383  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43384  return __ret;
43385 }
43386 #else
43387 __ai uint32_t vcgtzs_f32(float32_t __p0) {
43388  uint32_t __ret;
43389  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43390  return __ret;
43391 }
43392 #endif
43393 
43394 #ifdef __LITTLE_ENDIAN__
43395 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43396  uint64x2_t __ret;
43397  __ret = (uint64x2_t)(__p0 <= __p1);
43398  return __ret;
43399 }
43400 #else
43401 __ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43402  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43403  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43404  uint64x2_t __ret;
43405  __ret = (uint64x2_t)(__rev0 <= __rev1);
43406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43407  return __ret;
43408 }
43409 #endif
43410 
43411 #ifdef __LITTLE_ENDIAN__
43412 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43413  uint64x2_t __ret;
43414  __ret = (uint64x2_t)(__p0 <= __p1);
43415  return __ret;
43416 }
43417 #else
43418 __ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43419  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43420  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43421  uint64x2_t __ret;
43422  __ret = (uint64x2_t)(__rev0 <= __rev1);
43423  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43424  return __ret;
43425 }
43426 #endif
43427 
43428 #ifdef __LITTLE_ENDIAN__
43429 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43430  uint64x2_t __ret;
43431  __ret = (uint64x2_t)(__p0 <= __p1);
43432  return __ret;
43433 }
43434 #else
43435 __ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43436  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43437  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43438  uint64x2_t __ret;
43439  __ret = (uint64x2_t)(__rev0 <= __rev1);
43440  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43441  return __ret;
43442 }
43443 #endif
43444 
43445 #ifdef __LITTLE_ENDIAN__
43446 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43447  uint64x1_t __ret;
43448  __ret = (uint64x1_t)(__p0 <= __p1);
43449  return __ret;
43450 }
43451 #else
43452 __ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43453  uint64x1_t __ret;
43454  __ret = (uint64x1_t)(__p0 <= __p1);
43455  return __ret;
43456 }
43457 #endif
43458 
43459 #ifdef __LITTLE_ENDIAN__
43460 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43461  uint64x1_t __ret;
43462  __ret = (uint64x1_t)(__p0 <= __p1);
43463  return __ret;
43464 }
43465 #else
43466 __ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43467  uint64x1_t __ret;
43468  __ret = (uint64x1_t)(__p0 <= __p1);
43469  return __ret;
43470 }
43471 #endif
43472 
43473 #ifdef __LITTLE_ENDIAN__
43474 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43475  uint64x1_t __ret;
43476  __ret = (uint64x1_t)(__p0 <= __p1);
43477  return __ret;
43478 }
43479 #else
43480 __ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43481  uint64x1_t __ret;
43482  __ret = (uint64x1_t)(__p0 <= __p1);
43483  return __ret;
43484 }
43485 #endif
43486 
43487 #ifdef __LITTLE_ENDIAN__
43488 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43489  uint64_t __ret;
43490  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43491  return __ret;
43492 }
43493 #else
43494 __ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43495  uint64_t __ret;
43496  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43497  return __ret;
43498 }
43499 #endif
43500 
43501 #ifdef __LITTLE_ENDIAN__
43502 __ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43503  int64_t __ret;
43504  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43505  return __ret;
43506 }
43507 #else
43508 __ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43509  int64_t __ret;
43510  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43511  return __ret;
43512 }
43513 #endif
43514 
43515 #ifdef __LITTLE_ENDIAN__
43516 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43517  uint64_t __ret;
43518  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43519  return __ret;
43520 }
43521 #else
43522 __ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43523  uint64_t __ret;
43524  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43525  return __ret;
43526 }
43527 #endif
43528 
43529 #ifdef __LITTLE_ENDIAN__
43530 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43531  uint32_t __ret;
43532  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43533  return __ret;
43534 }
43535 #else
43536 __ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43537  uint32_t __ret;
43538  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43539  return __ret;
43540 }
43541 #endif
43542 
43543 #ifdef __LITTLE_ENDIAN__
43544 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43545  uint8x16_t __ret;
43546  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
43547  return __ret;
43548 }
43549 #else
43550 __ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43551  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43552  uint8x16_t __ret;
43553  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
43554  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43555  return __ret;
43556 }
43557 #endif
43558 
43559 #ifdef __LITTLE_ENDIAN__
43560 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43561  uint64x2_t __ret;
43562  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43563  return __ret;
43564 }
43565 #else
43566 __ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43567  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43568  uint64x2_t __ret;
43569  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43570  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43571  return __ret;
43572 }
43573 #endif
43574 
43575 #ifdef __LITTLE_ENDIAN__
43576 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43577  uint32x4_t __ret;
43578  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43579  return __ret;
43580 }
43581 #else
43582 __ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43583  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43584  uint32x4_t __ret;
43585  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43586  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43587  return __ret;
43588 }
43589 #endif
43590 
43591 #ifdef __LITTLE_ENDIAN__
43592 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43593  uint32x4_t __ret;
43594  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43595  return __ret;
43596 }
43597 #else
43598 __ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43599  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43600  uint32x4_t __ret;
43601  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43602  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43603  return __ret;
43604 }
43605 #endif
43606 
43607 #ifdef __LITTLE_ENDIAN__
43608 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43609  uint64x2_t __ret;
43610  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43611  return __ret;
43612 }
43613 #else
43614 __ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43615  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43616  uint64x2_t __ret;
43617  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43618  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43619  return __ret;
43620 }
43621 #endif
43622 
43623 #ifdef __LITTLE_ENDIAN__
43624 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43625  uint16x8_t __ret;
43626  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
43627  return __ret;
43628 }
43629 #else
43630 __ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43631  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43632  uint16x8_t __ret;
43633  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
43634  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43635  return __ret;
43636 }
43637 #endif
43638 
43639 #ifdef __LITTLE_ENDIAN__
43640 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
43641  uint8x8_t __ret;
43642  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
43643  return __ret;
43644 }
43645 #else
43646 __ai uint8x8_t vclez_s8(int8x8_t __p0) {
43647  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43648  uint8x8_t __ret;
43649  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
43650  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43651  return __ret;
43652 }
43653 #endif
43654 
43655 #ifdef __LITTLE_ENDIAN__
43656 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
43657  uint64x1_t __ret;
43658  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43659  return __ret;
43660 }
43661 #else
43662 __ai uint64x1_t vclez_f64(float64x1_t __p0) {
43663  uint64x1_t __ret;
43664  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43665  return __ret;
43666 }
43667 #endif
43668 
43669 #ifdef __LITTLE_ENDIAN__
43670 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
43671  uint32x2_t __ret;
43672  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43673  return __ret;
43674 }
43675 #else
43676 __ai uint32x2_t vclez_f32(float32x2_t __p0) {
43677  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43678  uint32x2_t __ret;
43679  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43681  return __ret;
43682 }
43683 #endif
43684 
43685 #ifdef __LITTLE_ENDIAN__
43686 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
43687  uint32x2_t __ret;
43688  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43689  return __ret;
43690 }
43691 #else
43692 __ai uint32x2_t vclez_s32(int32x2_t __p0) {
43693  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43694  uint32x2_t __ret;
43695  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43696  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43697  return __ret;
43698 }
43699 #endif
43700 
43701 #ifdef __LITTLE_ENDIAN__
43702 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
43703  uint64x1_t __ret;
43704  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43705  return __ret;
43706 }
43707 #else
43708 __ai uint64x1_t vclez_s64(int64x1_t __p0) {
43709  uint64x1_t __ret;
43710  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43711  return __ret;
43712 }
43713 #endif
43714 
43715 #ifdef __LITTLE_ENDIAN__
43716 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
43717  uint16x4_t __ret;
43718  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
43719  return __ret;
43720 }
43721 #else
43722 __ai uint16x4_t vclez_s16(int16x4_t __p0) {
43723  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43724  uint16x4_t __ret;
43725  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
43726  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43727  return __ret;
43728 }
43729 #endif
43730 
43731 #ifdef __LITTLE_ENDIAN__
43732 __ai int64_t vclezd_s64(int64_t __p0) {
43733  int64_t __ret;
43734  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43735  return __ret;
43736 }
43737 #else
43738 __ai int64_t vclezd_s64(int64_t __p0) {
43739  int64_t __ret;
43740  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43741  return __ret;
43742 }
43743 #endif
43744 
43745 #ifdef __LITTLE_ENDIAN__
43746 __ai uint64_t vclezd_f64(float64_t __p0) {
43747  uint64_t __ret;
43748  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43749  return __ret;
43750 }
43751 #else
43752 __ai uint64_t vclezd_f64(float64_t __p0) {
43753  uint64_t __ret;
43754  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43755  return __ret;
43756 }
43757 #endif
43758 
43759 #ifdef __LITTLE_ENDIAN__
43760 __ai uint32_t vclezs_f32(float32_t __p0) {
43761  uint32_t __ret;
43762  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43763  return __ret;
43764 }
43765 #else
43766 __ai uint32_t vclezs_f32(float32_t __p0) {
43767  uint32_t __ret;
43768  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43769  return __ret;
43770 }
43771 #endif
43772 
43773 #ifdef __LITTLE_ENDIAN__
43774 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43775  uint64x2_t __ret;
43776  __ret = (uint64x2_t)(__p0 < __p1);
43777  return __ret;
43778 }
43779 #else
43780 __ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43781  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43782  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43783  uint64x2_t __ret;
43784  __ret = (uint64x2_t)(__rev0 < __rev1);
43785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43786  return __ret;
43787 }
43788 #endif
43789 
43790 #ifdef __LITTLE_ENDIAN__
43791 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43792  uint64x2_t __ret;
43793  __ret = (uint64x2_t)(__p0 < __p1);
43794  return __ret;
43795 }
43796 #else
43797 __ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43798  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43799  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43800  uint64x2_t __ret;
43801  __ret = (uint64x2_t)(__rev0 < __rev1);
43802  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43803  return __ret;
43804 }
43805 #endif
43806 
43807 #ifdef __LITTLE_ENDIAN__
43808 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43809  uint64x2_t __ret;
43810  __ret = (uint64x2_t)(__p0 < __p1);
43811  return __ret;
43812 }
43813 #else
43814 __ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43815  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43816  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43817  uint64x2_t __ret;
43818  __ret = (uint64x2_t)(__rev0 < __rev1);
43819  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43820  return __ret;
43821 }
43822 #endif
43823 
43824 #ifdef __LITTLE_ENDIAN__
43825 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43826  uint64x1_t __ret;
43827  __ret = (uint64x1_t)(__p0 < __p1);
43828  return __ret;
43829 }
43830 #else
43831 __ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43832  uint64x1_t __ret;
43833  __ret = (uint64x1_t)(__p0 < __p1);
43834  return __ret;
43835 }
43836 #endif
43837 
43838 #ifdef __LITTLE_ENDIAN__
43839 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43840  uint64x1_t __ret;
43841  __ret = (uint64x1_t)(__p0 < __p1);
43842  return __ret;
43843 }
43844 #else
43845 __ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43846  uint64x1_t __ret;
43847  __ret = (uint64x1_t)(__p0 < __p1);
43848  return __ret;
43849 }
43850 #endif
43851 
43852 #ifdef __LITTLE_ENDIAN__
43853 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43854  uint64x1_t __ret;
43855  __ret = (uint64x1_t)(__p0 < __p1);
43856  return __ret;
43857 }
43858 #else
43859 __ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43860  uint64x1_t __ret;
43861  __ret = (uint64x1_t)(__p0 < __p1);
43862  return __ret;
43863 }
43864 #endif
43865 
43866 #ifdef __LITTLE_ENDIAN__
43867 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43868  uint64_t __ret;
43869  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43870  return __ret;
43871 }
43872 #else
43873 __ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43874  uint64_t __ret;
43875  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43876  return __ret;
43877 }
43878 #endif
43879 
43880 #ifdef __LITTLE_ENDIAN__
43881 __ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43882  int64_t __ret;
43883  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43884  return __ret;
43885 }
43886 #else
43887 __ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43888  int64_t __ret;
43889  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43890  return __ret;
43891 }
43892 #endif
43893 
43894 #ifdef __LITTLE_ENDIAN__
43895 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43896  uint64_t __ret;
43897  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43898  return __ret;
43899 }
43900 #else
43901 __ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43902  uint64_t __ret;
43903  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43904  return __ret;
43905 }
43906 #endif
43907 
43908 #ifdef __LITTLE_ENDIAN__
43909 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43910  uint32_t __ret;
43911  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43912  return __ret;
43913 }
43914 #else
43915 __ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43916  uint32_t __ret;
43917  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43918  return __ret;
43919 }
43920 #endif
43921 
43922 #ifdef __LITTLE_ENDIAN__
43923 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43924  uint8x16_t __ret;
43925  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
43926  return __ret;
43927 }
43928 #else
43929 __ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43930  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43931  uint8x16_t __ret;
43932  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
43933  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43934  return __ret;
43935 }
43936 #endif
43937 
43938 #ifdef __LITTLE_ENDIAN__
43939 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43940  uint64x2_t __ret;
43941  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43942  return __ret;
43943 }
43944 #else
43945 __ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43946  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43947  uint64x2_t __ret;
43948  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43949  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43950  return __ret;
43951 }
43952 #endif
43953 
43954 #ifdef __LITTLE_ENDIAN__
43955 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43956  uint32x4_t __ret;
43957  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43958  return __ret;
43959 }
43960 #else
43961 __ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43962  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43963  uint32x4_t __ret;
43964  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43966  return __ret;
43967 }
43968 #endif
43969 
43970 #ifdef __LITTLE_ENDIAN__
43971 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43972  uint32x4_t __ret;
43973  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43974  return __ret;
43975 }
43976 #else
43977 __ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43978  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43979  uint32x4_t __ret;
43980  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43981  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43982  return __ret;
43983 }
43984 #endif
43985 
43986 #ifdef __LITTLE_ENDIAN__
43987 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43988  uint64x2_t __ret;
43989  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43990  return __ret;
43991 }
43992 #else
43993 __ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43994  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43995  uint64x2_t __ret;
43996  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43997  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43998  return __ret;
43999 }
44000 #endif
44001 
44002 #ifdef __LITTLE_ENDIAN__
44003 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44004  uint16x8_t __ret;
44005  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
44006  return __ret;
44007 }
44008 #else
44009 __ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44010  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44011  uint16x8_t __ret;
44012  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
44013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44014  return __ret;
44015 }
44016 #endif
44017 
44018 #ifdef __LITTLE_ENDIAN__
44019 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44020  uint8x8_t __ret;
44021  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
44022  return __ret;
44023 }
44024 #else
44025 __ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44026  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44027  uint8x8_t __ret;
44028  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
44029  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44030  return __ret;
44031 }
44032 #endif
44033 
44034 #ifdef __LITTLE_ENDIAN__
44035 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44036  uint64x1_t __ret;
44037  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44038  return __ret;
44039 }
44040 #else
44041 __ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44042  uint64x1_t __ret;
44043  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44044  return __ret;
44045 }
44046 #endif
44047 
44048 #ifdef __LITTLE_ENDIAN__
44049 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44050  uint32x2_t __ret;
44051  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44052  return __ret;
44053 }
44054 #else
44055 __ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44056  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44057  uint32x2_t __ret;
44058  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44059  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44060  return __ret;
44061 }
44062 #endif
44063 
44064 #ifdef __LITTLE_ENDIAN__
44065 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44066  uint32x2_t __ret;
44067  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44068  return __ret;
44069 }
44070 #else
44071 __ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44072  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44073  uint32x2_t __ret;
44074  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44076  return __ret;
44077 }
44078 #endif
44079 
44080 #ifdef __LITTLE_ENDIAN__
44081 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44082  uint64x1_t __ret;
44083  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44084  return __ret;
44085 }
44086 #else
44087 __ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44088  uint64x1_t __ret;
44089  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44090  return __ret;
44091 }
44092 #endif
44093 
44094 #ifdef __LITTLE_ENDIAN__
44095 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44096  uint16x4_t __ret;
44097  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
44098  return __ret;
44099 }
44100 #else
44101 __ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44102  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44103  uint16x4_t __ret;
44104  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
44105  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44106  return __ret;
44107 }
44108 #endif
44109 
44110 #ifdef __LITTLE_ENDIAN__
44111 __ai int64_t vcltzd_s64(int64_t __p0) {
44112  int64_t __ret;
44113  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44114  return __ret;
44115 }
44116 #else
44117 __ai int64_t vcltzd_s64(int64_t __p0) {
44118  int64_t __ret;
44119  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44120  return __ret;
44121 }
44122 #endif
44123 
44124 #ifdef __LITTLE_ENDIAN__
44125 __ai uint64_t vcltzd_f64(float64_t __p0) {
44126  uint64_t __ret;
44127  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44128  return __ret;
44129 }
44130 #else
44131 __ai uint64_t vcltzd_f64(float64_t __p0) {
44132  uint64_t __ret;
44133  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44134  return __ret;
44135 }
44136 #endif
44137 
44138 #ifdef __LITTLE_ENDIAN__
44139 __ai uint32_t vcltzs_f32(float32_t __p0) {
44140  uint32_t __ret;
44141  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44142  return __ret;
44143 }
44144 #else
44145 __ai uint32_t vcltzs_f32(float32_t __p0) {
44146  uint32_t __ret;
44147  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44148  return __ret;
44149 }
44150 #endif
44151 
44152 #ifdef __LITTLE_ENDIAN__
44153 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44154  poly64x2_t __ret;
44155  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44156  return __ret;
44157 }
44158 #else
44159 __ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44160  poly64x2_t __ret;
44161  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44162  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44163  return __ret;
44164 }
44165 #endif
44166 
44167 #ifdef __LITTLE_ENDIAN__
44168 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44169  float64x2_t __ret;
44170  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44171  return __ret;
44172 }
44173 #else
44174 __ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44175  float64x2_t __ret;
44176  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44177  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44178  return __ret;
44179 }
44180 #endif
44181 
44182 #ifdef __LITTLE_ENDIAN__
44183 #define vcopyq_lane_p8(__p0_0, __p1_0, __p2_0, __p3_0) __extension__ ({ \
44184  poly8x16_t __s0_0 = __p0_0; \
44185  poly8x8_t __s2_0 = __p2_0; \
44186  poly8x16_t __ret_0; \
44187  __ret_0 = vsetq_lane_p8(vget_lane_p8(__s2_0, __p3_0), __s0_0, __p1_0); \
44188  __ret_0; \
44189 })
44190 #else
44191 #define vcopyq_lane_p8(__p0_1, __p1_1, __p2_1, __p3_1) __extension__ ({ \
44192  poly8x16_t __s0_1 = __p0_1; \
44193  poly8x8_t __s2_1 = __p2_1; \
44194  poly8x16_t __rev0_1; __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44195  poly8x8_t __rev2_1; __rev2_1 = __builtin_shufflevector(__s2_1, __s2_1, 7, 6, 5, 4, 3, 2, 1, 0); \
44196  poly8x16_t __ret_1; \
44197  __ret_1 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_1, __p3_1), __rev0_1, __p1_1); \
44198  __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44199  __ret_1; \
44200 })
44201 #endif
44202 
44203 #ifdef __LITTLE_ENDIAN__
44204 #define vcopyq_lane_p16(__p0_2, __p1_2, __p2_2, __p3_2) __extension__ ({ \
44205  poly16x8_t __s0_2 = __p0_2; \
44206  poly16x4_t __s2_2 = __p2_2; \
44207  poly16x8_t __ret_2; \
44208  __ret_2 = vsetq_lane_p16(vget_lane_p16(__s2_2, __p3_2), __s0_2, __p1_2); \
44209  __ret_2; \
44210 })
44211 #else
44212 #define vcopyq_lane_p16(__p0_3, __p1_3, __p2_3, __p3_3) __extension__ ({ \
44213  poly16x8_t __s0_3 = __p0_3; \
44214  poly16x4_t __s2_3 = __p2_3; \
44215  poly16x8_t __rev0_3; __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44216  poly16x4_t __rev2_3; __rev2_3 = __builtin_shufflevector(__s2_3, __s2_3, 3, 2, 1, 0); \
44217  poly16x8_t __ret_3; \
44218  __ret_3 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_3, __p3_3), __rev0_3, __p1_3); \
44219  __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44220  __ret_3; \
44221 })
44222 #endif
44223 
44224 #ifdef __LITTLE_ENDIAN__
44225 #define vcopyq_lane_u8(__p0_4, __p1_4, __p2_4, __p3_4) __extension__ ({ \
44226  uint8x16_t __s0_4 = __p0_4; \
44227  uint8x8_t __s2_4 = __p2_4; \
44228  uint8x16_t __ret_4; \
44229  __ret_4 = vsetq_lane_u8(vget_lane_u8(__s2_4, __p3_4), __s0_4, __p1_4); \
44230  __ret_4; \
44231 })
44232 #else
44233 #define vcopyq_lane_u8(__p0_5, __p1_5, __p2_5, __p3_5) __extension__ ({ \
44234  uint8x16_t __s0_5 = __p0_5; \
44235  uint8x8_t __s2_5 = __p2_5; \
44236  uint8x16_t __rev0_5; __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44237  uint8x8_t __rev2_5; __rev2_5 = __builtin_shufflevector(__s2_5, __s2_5, 7, 6, 5, 4, 3, 2, 1, 0); \
44238  uint8x16_t __ret_5; \
44239  __ret_5 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_5, __p3_5), __rev0_5, __p1_5); \
44240  __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44241  __ret_5; \
44242 })
44243 #endif
44244 
44245 #ifdef __LITTLE_ENDIAN__
44246 #define vcopyq_lane_u32(__p0_6, __p1_6, __p2_6, __p3_6) __extension__ ({ \
44247  uint32x4_t __s0_6 = __p0_6; \
44248  uint32x2_t __s2_6 = __p2_6; \
44249  uint32x4_t __ret_6; \
44250  __ret_6 = vsetq_lane_u32(vget_lane_u32(__s2_6, __p3_6), __s0_6, __p1_6); \
44251  __ret_6; \
44252 })
44253 #else
44254 #define vcopyq_lane_u32(__p0_7, __p1_7, __p2_7, __p3_7) __extension__ ({ \
44255  uint32x4_t __s0_7 = __p0_7; \
44256  uint32x2_t __s2_7 = __p2_7; \
44257  uint32x4_t __rev0_7; __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
44258  uint32x2_t __rev2_7; __rev2_7 = __builtin_shufflevector(__s2_7, __s2_7, 1, 0); \
44259  uint32x4_t __ret_7; \
44260  __ret_7 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_7, __p3_7), __rev0_7, __p1_7); \
44261  __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 3, 2, 1, 0); \
44262  __ret_7; \
44263 })
44264 #endif
44265 
44266 #ifdef __LITTLE_ENDIAN__
44267 #define vcopyq_lane_u64(__p0_8, __p1_8, __p2_8, __p3_8) __extension__ ({ \
44268  uint64x2_t __s0_8 = __p0_8; \
44269  uint64x1_t __s2_8 = __p2_8; \
44270  uint64x2_t __ret_8; \
44271  __ret_8 = vsetq_lane_u64(vget_lane_u64(__s2_8, __p3_8), __s0_8, __p1_8); \
44272  __ret_8; \
44273 })
44274 #else
44275 #define vcopyq_lane_u64(__p0_9, __p1_9, __p2_9, __p3_9) __extension__ ({ \
44276  uint64x2_t __s0_9 = __p0_9; \
44277  uint64x1_t __s2_9 = __p2_9; \
44278  uint64x2_t __rev0_9; __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 1, 0); \
44279  uint64x2_t __ret_9; \
44280  __ret_9 = __noswap_vsetq_lane_u64(__noswap_vget_lane_u64(__s2_9, __p3_9), __rev0_9, __p1_9); \
44281  __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 1, 0); \
44282  __ret_9; \
44283 })
44284 #endif
44285 
44286 #ifdef __LITTLE_ENDIAN__
44287 #define vcopyq_lane_u16(__p0_10, __p1_10, __p2_10, __p3_10) __extension__ ({ \
44288  uint16x8_t __s0_10 = __p0_10; \
44289  uint16x4_t __s2_10 = __p2_10; \
44290  uint16x8_t __ret_10; \
44291  __ret_10 = vsetq_lane_u16(vget_lane_u16(__s2_10, __p3_10), __s0_10, __p1_10); \
44292  __ret_10; \
44293 })
44294 #else
44295 #define vcopyq_lane_u16(__p0_11, __p1_11, __p2_11, __p3_11) __extension__ ({ \
44296  uint16x8_t __s0_11 = __p0_11; \
44297  uint16x4_t __s2_11 = __p2_11; \
44298  uint16x8_t __rev0_11; __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44299  uint16x4_t __rev2_11; __rev2_11 = __builtin_shufflevector(__s2_11, __s2_11, 3, 2, 1, 0); \
44300  uint16x8_t __ret_11; \
44301  __ret_11 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_11, __p3_11), __rev0_11, __p1_11); \
44302  __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44303  __ret_11; \
44304 })
44305 #endif
44306 
44307 #ifdef __LITTLE_ENDIAN__
44308 #define vcopyq_lane_s8(__p0_12, __p1_12, __p2_12, __p3_12) __extension__ ({ \
44309  int8x16_t __s0_12 = __p0_12; \
44310  int8x8_t __s2_12 = __p2_12; \
44311  int8x16_t __ret_12; \
44312  __ret_12 = vsetq_lane_s8(vget_lane_s8(__s2_12, __p3_12), __s0_12, __p1_12); \
44313  __ret_12; \
44314 })
44315 #else
44316 #define vcopyq_lane_s8(__p0_13, __p1_13, __p2_13, __p3_13) __extension__ ({ \
44317  int8x16_t __s0_13 = __p0_13; \
44318  int8x8_t __s2_13 = __p2_13; \
44319  int8x16_t __rev0_13; __rev0_13 = __builtin_shufflevector(__s0_13, __s0_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44320  int8x8_t __rev2_13; __rev2_13 = __builtin_shufflevector(__s2_13, __s2_13, 7, 6, 5, 4, 3, 2, 1, 0); \
44321  int8x16_t __ret_13; \
44322  __ret_13 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_13, __p3_13), __rev0_13, __p1_13); \
44323  __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44324  __ret_13; \
44325 })
44326 #endif
44327 
44328 #ifdef __LITTLE_ENDIAN__
44329 #define vcopyq_lane_f32(__p0_14, __p1_14, __p2_14, __p3_14) __extension__ ({ \
44330  float32x4_t __s0_14 = __p0_14; \
44331  float32x2_t __s2_14 = __p2_14; \
44332  float32x4_t __ret_14; \
44333  __ret_14 = vsetq_lane_f32(vget_lane_f32(__s2_14, __p3_14), __s0_14, __p1_14); \
44334  __ret_14; \
44335 })
44336 #else
44337 #define vcopyq_lane_f32(__p0_15, __p1_15, __p2_15, __p3_15) __extension__ ({ \
44338  float32x4_t __s0_15 = __p0_15; \
44339  float32x2_t __s2_15 = __p2_15; \
44340  float32x4_t __rev0_15; __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
44341  float32x2_t __rev2_15; __rev2_15 = __builtin_shufflevector(__s2_15, __s2_15, 1, 0); \
44342  float32x4_t __ret_15; \
44343  __ret_15 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_15, __p3_15), __rev0_15, __p1_15); \
44344  __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 3, 2, 1, 0); \
44345  __ret_15; \
44346 })
44347 #endif
44348 
44349 #ifdef __LITTLE_ENDIAN__
44350 #define vcopyq_lane_s32(__p0_16, __p1_16, __p2_16, __p3_16) __extension__ ({ \
44351  int32x4_t __s0_16 = __p0_16; \
44352  int32x2_t __s2_16 = __p2_16; \
44353  int32x4_t __ret_16; \
44354  __ret_16 = vsetq_lane_s32(vget_lane_s32(__s2_16, __p3_16), __s0_16, __p1_16); \
44355  __ret_16; \
44356 })
44357 #else
44358 #define vcopyq_lane_s32(__p0_17, __p1_17, __p2_17, __p3_17) __extension__ ({ \
44359  int32x4_t __s0_17 = __p0_17; \
44360  int32x2_t __s2_17 = __p2_17; \
44361  int32x4_t __rev0_17; __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 3, 2, 1, 0); \
44362  int32x2_t __rev2_17; __rev2_17 = __builtin_shufflevector(__s2_17, __s2_17, 1, 0); \
44363  int32x4_t __ret_17; \
44364  __ret_17 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_17, __p3_17), __rev0_17, __p1_17); \
44365  __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 3, 2, 1, 0); \
44366  __ret_17; \
44367 })
44368 #endif
44369 
44370 #ifdef __LITTLE_ENDIAN__
44371 #define vcopyq_lane_s64(__p0_18, __p1_18, __p2_18, __p3_18) __extension__ ({ \
44372  int64x2_t __s0_18 = __p0_18; \
44373  int64x1_t __s2_18 = __p2_18; \
44374  int64x2_t __ret_18; \
44375  __ret_18 = vsetq_lane_s64(vget_lane_s64(__s2_18, __p3_18), __s0_18, __p1_18); \
44376  __ret_18; \
44377 })
44378 #else
44379 #define vcopyq_lane_s64(__p0_19, __p1_19, __p2_19, __p3_19) __extension__ ({ \
44380  int64x2_t __s0_19 = __p0_19; \
44381  int64x1_t __s2_19 = __p2_19; \
44382  int64x2_t __rev0_19; __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
44383  int64x2_t __ret_19; \
44384  __ret_19 = __noswap_vsetq_lane_s64(__noswap_vget_lane_s64(__s2_19, __p3_19), __rev0_19, __p1_19); \
44385  __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 1, 0); \
44386  __ret_19; \
44387 })
44388 #endif
44389 
44390 #ifdef __LITTLE_ENDIAN__
44391 #define vcopyq_lane_s16(__p0_20, __p1_20, __p2_20, __p3_20) __extension__ ({ \
44392  int16x8_t __s0_20 = __p0_20; \
44393  int16x4_t __s2_20 = __p2_20; \
44394  int16x8_t __ret_20; \
44395  __ret_20 = vsetq_lane_s16(vget_lane_s16(__s2_20, __p3_20), __s0_20, __p1_20); \
44396  __ret_20; \
44397 })
44398 #else
44399 #define vcopyq_lane_s16(__p0_21, __p1_21, __p2_21, __p3_21) __extension__ ({ \
44400  int16x8_t __s0_21 = __p0_21; \
44401  int16x4_t __s2_21 = __p2_21; \
44402  int16x8_t __rev0_21; __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44403  int16x4_t __rev2_21; __rev2_21 = __builtin_shufflevector(__s2_21, __s2_21, 3, 2, 1, 0); \
44404  int16x8_t __ret_21; \
44405  __ret_21 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_21, __p3_21), __rev0_21, __p1_21); \
44406  __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44407  __ret_21; \
44408 })
44409 #endif
44410 
44411 #ifdef __LITTLE_ENDIAN__
44412 #define vcopy_lane_p8(__p0_22, __p1_22, __p2_22, __p3_22) __extension__ ({ \
44413  poly8x8_t __s0_22 = __p0_22; \
44414  poly8x8_t __s2_22 = __p2_22; \
44415  poly8x8_t __ret_22; \
44416  __ret_22 = vset_lane_p8(vget_lane_p8(__s2_22, __p3_22), __s0_22, __p1_22); \
44417  __ret_22; \
44418 })
44419 #else
44420 #define vcopy_lane_p8(__p0_23, __p1_23, __p2_23, __p3_23) __extension__ ({ \
44421  poly8x8_t __s0_23 = __p0_23; \
44422  poly8x8_t __s2_23 = __p2_23; \
44423  poly8x8_t __rev0_23; __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44424  poly8x8_t __rev2_23; __rev2_23 = __builtin_shufflevector(__s2_23, __s2_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44425  poly8x8_t __ret_23; \
44426  __ret_23 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_23, __p3_23), __rev0_23, __p1_23); \
44427  __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44428  __ret_23; \
44429 })
44430 #endif
44431 
44432 #ifdef __LITTLE_ENDIAN__
44433 #define vcopy_lane_p16(__p0_24, __p1_24, __p2_24, __p3_24) __extension__ ({ \
44434  poly16x4_t __s0_24 = __p0_24; \
44435  poly16x4_t __s2_24 = __p2_24; \
44436  poly16x4_t __ret_24; \
44437  __ret_24 = vset_lane_p16(vget_lane_p16(__s2_24, __p3_24), __s0_24, __p1_24); \
44438  __ret_24; \
44439 })
44440 #else
44441 #define vcopy_lane_p16(__p0_25, __p1_25, __p2_25, __p3_25) __extension__ ({ \
44442  poly16x4_t __s0_25 = __p0_25; \
44443  poly16x4_t __s2_25 = __p2_25; \
44444  poly16x4_t __rev0_25; __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
44445  poly16x4_t __rev2_25; __rev2_25 = __builtin_shufflevector(__s2_25, __s2_25, 3, 2, 1, 0); \
44446  poly16x4_t __ret_25; \
44447  __ret_25 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_25, __p3_25), __rev0_25, __p1_25); \
44448  __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 3, 2, 1, 0); \
44449  __ret_25; \
44450 })
44451 #endif
44452 
44453 #ifdef __LITTLE_ENDIAN__
44454 #define vcopy_lane_u8(__p0_26, __p1_26, __p2_26, __p3_26) __extension__ ({ \
44455  uint8x8_t __s0_26 = __p0_26; \
44456  uint8x8_t __s2_26 = __p2_26; \
44457  uint8x8_t __ret_26; \
44458  __ret_26 = vset_lane_u8(vget_lane_u8(__s2_26, __p3_26), __s0_26, __p1_26); \
44459  __ret_26; \
44460 })
44461 #else
44462 #define vcopy_lane_u8(__p0_27, __p1_27, __p2_27, __p3_27) __extension__ ({ \
44463  uint8x8_t __s0_27 = __p0_27; \
44464  uint8x8_t __s2_27 = __p2_27; \
44465  uint8x8_t __rev0_27; __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44466  uint8x8_t __rev2_27; __rev2_27 = __builtin_shufflevector(__s2_27, __s2_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44467  uint8x8_t __ret_27; \
44468  __ret_27 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_27, __p3_27), __rev0_27, __p1_27); \
44469  __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44470  __ret_27; \
44471 })
44472 #endif
44473 
44474 #ifdef __LITTLE_ENDIAN__
44475 #define vcopy_lane_u32(__p0_28, __p1_28, __p2_28, __p3_28) __extension__ ({ \
44476  uint32x2_t __s0_28 = __p0_28; \
44477  uint32x2_t __s2_28 = __p2_28; \
44478  uint32x2_t __ret_28; \
44479  __ret_28 = vset_lane_u32(vget_lane_u32(__s2_28, __p3_28), __s0_28, __p1_28); \
44480  __ret_28; \
44481 })
44482 #else
44483 #define vcopy_lane_u32(__p0_29, __p1_29, __p2_29, __p3_29) __extension__ ({ \
44484  uint32x2_t __s0_29 = __p0_29; \
44485  uint32x2_t __s2_29 = __p2_29; \
44486  uint32x2_t __rev0_29; __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
44487  uint32x2_t __rev2_29; __rev2_29 = __builtin_shufflevector(__s2_29, __s2_29, 1, 0); \
44488  uint32x2_t __ret_29; \
44489  __ret_29 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_29, __p3_29), __rev0_29, __p1_29); \
44490  __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
44491  __ret_29; \
44492 })
44493 #endif
44494 
44495 #ifdef __LITTLE_ENDIAN__
44496 #define vcopy_lane_u64(__p0_30, __p1_30, __p2_30, __p3_30) __extension__ ({ \
44497  uint64x1_t __s0_30 = __p0_30; \
44498  uint64x1_t __s2_30 = __p2_30; \
44499  uint64x1_t __ret_30; \
44500  __ret_30 = vset_lane_u64(vget_lane_u64(__s2_30, __p3_30), __s0_30, __p1_30); \
44501  __ret_30; \
44502 })
44503 #else
44504 #define vcopy_lane_u64(__p0_31, __p1_31, __p2_31, __p3_31) __extension__ ({ \
44505  uint64x1_t __s0_31 = __p0_31; \
44506  uint64x1_t __s2_31 = __p2_31; \
44507  uint64x1_t __ret_31; \
44508  __ret_31 = __noswap_vset_lane_u64(__noswap_vget_lane_u64(__s2_31, __p3_31), __s0_31, __p1_31); \
44509  __ret_31; \
44510 })
44511 #endif
44512 
44513 #ifdef __LITTLE_ENDIAN__
44514 #define vcopy_lane_u16(__p0_32, __p1_32, __p2_32, __p3_32) __extension__ ({ \
44515  uint16x4_t __s0_32 = __p0_32; \
44516  uint16x4_t __s2_32 = __p2_32; \
44517  uint16x4_t __ret_32; \
44518  __ret_32 = vset_lane_u16(vget_lane_u16(__s2_32, __p3_32), __s0_32, __p1_32); \
44519  __ret_32; \
44520 })
44521 #else
44522 #define vcopy_lane_u16(__p0_33, __p1_33, __p2_33, __p3_33) __extension__ ({ \
44523  uint16x4_t __s0_33 = __p0_33; \
44524  uint16x4_t __s2_33 = __p2_33; \
44525  uint16x4_t __rev0_33; __rev0_33 = __builtin_shufflevector(__s0_33, __s0_33, 3, 2, 1, 0); \
44526  uint16x4_t __rev2_33; __rev2_33 = __builtin_shufflevector(__s2_33, __s2_33, 3, 2, 1, 0); \
44527  uint16x4_t __ret_33; \
44528  __ret_33 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_33, __p3_33), __rev0_33, __p1_33); \
44529  __ret_33 = __builtin_shufflevector(__ret_33, __ret_33, 3, 2, 1, 0); \
44530  __ret_33; \
44531 })
44532 #endif
44533 
44534 #ifdef __LITTLE_ENDIAN__
44535 #define vcopy_lane_s8(__p0_34, __p1_34, __p2_34, __p3_34) __extension__ ({ \
44536  int8x8_t __s0_34 = __p0_34; \
44537  int8x8_t __s2_34 = __p2_34; \
44538  int8x8_t __ret_34; \
44539  __ret_34 = vset_lane_s8(vget_lane_s8(__s2_34, __p3_34), __s0_34, __p1_34); \
44540  __ret_34; \
44541 })
44542 #else
44543 #define vcopy_lane_s8(__p0_35, __p1_35, __p2_35, __p3_35) __extension__ ({ \
44544  int8x8_t __s0_35 = __p0_35; \
44545  int8x8_t __s2_35 = __p2_35; \
44546  int8x8_t __rev0_35; __rev0_35 = __builtin_shufflevector(__s0_35, __s0_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44547  int8x8_t __rev2_35; __rev2_35 = __builtin_shufflevector(__s2_35, __s2_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44548  int8x8_t __ret_35; \
44549  __ret_35 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_35, __p3_35), __rev0_35, __p1_35); \
44550  __ret_35 = __builtin_shufflevector(__ret_35, __ret_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44551  __ret_35; \
44552 })
44553 #endif
44554 
44555 #ifdef __LITTLE_ENDIAN__
44556 #define vcopy_lane_f32(__p0_36, __p1_36, __p2_36, __p3_36) __extension__ ({ \
44557  float32x2_t __s0_36 = __p0_36; \
44558  float32x2_t __s2_36 = __p2_36; \
44559  float32x2_t __ret_36; \
44560  __ret_36 = vset_lane_f32(vget_lane_f32(__s2_36, __p3_36), __s0_36, __p1_36); \
44561  __ret_36; \
44562 })
44563 #else
44564 #define vcopy_lane_f32(__p0_37, __p1_37, __p2_37, __p3_37) __extension__ ({ \
44565  float32x2_t __s0_37 = __p0_37; \
44566  float32x2_t __s2_37 = __p2_37; \
44567  float32x2_t __rev0_37; __rev0_37 = __builtin_shufflevector(__s0_37, __s0_37, 1, 0); \
44568  float32x2_t __rev2_37; __rev2_37 = __builtin_shufflevector(__s2_37, __s2_37, 1, 0); \
44569  float32x2_t __ret_37; \
44570  __ret_37 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_37, __p3_37), __rev0_37, __p1_37); \
44571  __ret_37 = __builtin_shufflevector(__ret_37, __ret_37, 1, 0); \
44572  __ret_37; \
44573 })
44574 #endif
44575 
44576 #ifdef __LITTLE_ENDIAN__
44577 #define vcopy_lane_s32(__p0_38, __p1_38, __p2_38, __p3_38) __extension__ ({ \
44578  int32x2_t __s0_38 = __p0_38; \
44579  int32x2_t __s2_38 = __p2_38; \
44580  int32x2_t __ret_38; \
44581  __ret_38 = vset_lane_s32(vget_lane_s32(__s2_38, __p3_38), __s0_38, __p1_38); \
44582  __ret_38; \
44583 })
44584 #else
44585 #define vcopy_lane_s32(__p0_39, __p1_39, __p2_39, __p3_39) __extension__ ({ \
44586  int32x2_t __s0_39 = __p0_39; \
44587  int32x2_t __s2_39 = __p2_39; \
44588  int32x2_t __rev0_39; __rev0_39 = __builtin_shufflevector(__s0_39, __s0_39, 1, 0); \
44589  int32x2_t __rev2_39; __rev2_39 = __builtin_shufflevector(__s2_39, __s2_39, 1, 0); \
44590  int32x2_t __ret_39; \
44591  __ret_39 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_39, __p3_39), __rev0_39, __p1_39); \
44592  __ret_39 = __builtin_shufflevector(__ret_39, __ret_39, 1, 0); \
44593  __ret_39; \
44594 })
44595 #endif
44596 
44597 #ifdef __LITTLE_ENDIAN__
44598 #define vcopy_lane_s64(__p0_40, __p1_40, __p2_40, __p3_40) __extension__ ({ \
44599  int64x1_t __s0_40 = __p0_40; \
44600  int64x1_t __s2_40 = __p2_40; \
44601  int64x1_t __ret_40; \
44602  __ret_40 = vset_lane_s64(vget_lane_s64(__s2_40, __p3_40), __s0_40, __p1_40); \
44603  __ret_40; \
44604 })
44605 #else
44606 #define vcopy_lane_s64(__p0_41, __p1_41, __p2_41, __p3_41) __extension__ ({ \
44607  int64x1_t __s0_41 = __p0_41; \
44608  int64x1_t __s2_41 = __p2_41; \
44609  int64x1_t __ret_41; \
44610  __ret_41 = __noswap_vset_lane_s64(__noswap_vget_lane_s64(__s2_41, __p3_41), __s0_41, __p1_41); \
44611  __ret_41; \
44612 })
44613 #endif
44614 
44615 #ifdef __LITTLE_ENDIAN__
44616 #define vcopy_lane_s16(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
44617  int16x4_t __s0_42 = __p0_42; \
44618  int16x4_t __s2_42 = __p2_42; \
44619  int16x4_t __ret_42; \
44620  __ret_42 = vset_lane_s16(vget_lane_s16(__s2_42, __p3_42), __s0_42, __p1_42); \
44621  __ret_42; \
44622 })
44623 #else
44624 #define vcopy_lane_s16(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
44625  int16x4_t __s0_43 = __p0_43; \
44626  int16x4_t __s2_43 = __p2_43; \
44627  int16x4_t __rev0_43; __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
44628  int16x4_t __rev2_43; __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 3, 2, 1, 0); \
44629  int16x4_t __ret_43; \
44630  __ret_43 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_43, __p3_43), __rev0_43, __p1_43); \
44631  __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
44632  __ret_43; \
44633 })
44634 #endif
44635 
44636 #ifdef __LITTLE_ENDIAN__
44637 #define vcopyq_laneq_p8(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
44638  poly8x16_t __s0_44 = __p0_44; \
44639  poly8x16_t __s2_44 = __p2_44; \
44640  poly8x16_t __ret_44; \
44641  __ret_44 = vsetq_lane_p8(vgetq_lane_p8(__s2_44, __p3_44), __s0_44, __p1_44); \
44642  __ret_44; \
44643 })
44644 #else
44645 #define vcopyq_laneq_p8(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
44646  poly8x16_t __s0_45 = __p0_45; \
44647  poly8x16_t __s2_45 = __p2_45; \
44648  poly8x16_t __rev0_45; __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44649  poly8x16_t __rev2_45; __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44650  poly8x16_t __ret_45; \
44651  __ret_45 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_45, __p3_45), __rev0_45, __p1_45); \
44652  __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44653  __ret_45; \
44654 })
44655 #endif
44656 
44657 #ifdef __LITTLE_ENDIAN__
44658 #define vcopyq_laneq_p16(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
44659  poly16x8_t __s0_46 = __p0_46; \
44660  poly16x8_t __s2_46 = __p2_46; \
44661  poly16x8_t __ret_46; \
44662  __ret_46 = vsetq_lane_p16(vgetq_lane_p16(__s2_46, __p3_46), __s0_46, __p1_46); \
44663  __ret_46; \
44664 })
44665 #else
44666 #define vcopyq_laneq_p16(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
44667  poly16x8_t __s0_47 = __p0_47; \
44668  poly16x8_t __s2_47 = __p2_47; \
44669  poly16x8_t __rev0_47; __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44670  poly16x8_t __rev2_47; __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44671  poly16x8_t __ret_47; \
44672  __ret_47 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_47, __p3_47), __rev0_47, __p1_47); \
44673  __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44674  __ret_47; \
44675 })
44676 #endif
44677 
44678 #ifdef __LITTLE_ENDIAN__
44679 #define vcopyq_laneq_u8(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
44680  uint8x16_t __s0_48 = __p0_48; \
44681  uint8x16_t __s2_48 = __p2_48; \
44682  uint8x16_t __ret_48; \
44683  __ret_48 = vsetq_lane_u8(vgetq_lane_u8(__s2_48, __p3_48), __s0_48, __p1_48); \
44684  __ret_48; \
44685 })
44686 #else
44687 #define vcopyq_laneq_u8(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
44688  uint8x16_t __s0_49 = __p0_49; \
44689  uint8x16_t __s2_49 = __p2_49; \
44690  uint8x16_t __rev0_49; __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44691  uint8x16_t __rev2_49; __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44692  uint8x16_t __ret_49; \
44693  __ret_49 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_49, __p3_49), __rev0_49, __p1_49); \
44694  __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44695  __ret_49; \
44696 })
44697 #endif
44698 
44699 #ifdef __LITTLE_ENDIAN__
44700 #define vcopyq_laneq_u32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
44701  uint32x4_t __s0_50 = __p0_50; \
44702  uint32x4_t __s2_50 = __p2_50; \
44703  uint32x4_t __ret_50; \
44704  __ret_50 = vsetq_lane_u32(vgetq_lane_u32(__s2_50, __p3_50), __s0_50, __p1_50); \
44705  __ret_50; \
44706 })
44707 #else
44708 #define vcopyq_laneq_u32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
44709  uint32x4_t __s0_51 = __p0_51; \
44710  uint32x4_t __s2_51 = __p2_51; \
44711  uint32x4_t __rev0_51; __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
44712  uint32x4_t __rev2_51; __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
44713  uint32x4_t __ret_51; \
44714  __ret_51 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_51, __p3_51), __rev0_51, __p1_51); \
44715  __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
44716  __ret_51; \
44717 })
44718 #endif
44719 
44720 #ifdef __LITTLE_ENDIAN__
44721 #define vcopyq_laneq_u64(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
44722  uint64x2_t __s0_52 = __p0_52; \
44723  uint64x2_t __s2_52 = __p2_52; \
44724  uint64x2_t __ret_52; \
44725  __ret_52 = vsetq_lane_u64(vgetq_lane_u64(__s2_52, __p3_52), __s0_52, __p1_52); \
44726  __ret_52; \
44727 })
44728 #else
44729 #define vcopyq_laneq_u64(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
44730  uint64x2_t __s0_53 = __p0_53; \
44731  uint64x2_t __s2_53 = __p2_53; \
44732  uint64x2_t __rev0_53; __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
44733  uint64x2_t __rev2_53; __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
44734  uint64x2_t __ret_53; \
44735  __ret_53 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_53, __p3_53), __rev0_53, __p1_53); \
44736  __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
44737  __ret_53; \
44738 })
44739 #endif
44740 
44741 #ifdef __LITTLE_ENDIAN__
44742 #define vcopyq_laneq_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
44743  uint16x8_t __s0_54 = __p0_54; \
44744  uint16x8_t __s2_54 = __p2_54; \
44745  uint16x8_t __ret_54; \
44746  __ret_54 = vsetq_lane_u16(vgetq_lane_u16(__s2_54, __p3_54), __s0_54, __p1_54); \
44747  __ret_54; \
44748 })
44749 #else
44750 #define vcopyq_laneq_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
44751  uint16x8_t __s0_55 = __p0_55; \
44752  uint16x8_t __s2_55 = __p2_55; \
44753  uint16x8_t __rev0_55; __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44754  uint16x8_t __rev2_55; __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44755  uint16x8_t __ret_55; \
44756  __ret_55 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_55, __p3_55), __rev0_55, __p1_55); \
44757  __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44758  __ret_55; \
44759 })
44760 #endif
44761 
44762 #ifdef __LITTLE_ENDIAN__
44763 #define vcopyq_laneq_s8(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
44764  int8x16_t __s0_56 = __p0_56; \
44765  int8x16_t __s2_56 = __p2_56; \
44766  int8x16_t __ret_56; \
44767  __ret_56 = vsetq_lane_s8(vgetq_lane_s8(__s2_56, __p3_56), __s0_56, __p1_56); \
44768  __ret_56; \
44769 })
44770 #else
44771 #define vcopyq_laneq_s8(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
44772  int8x16_t __s0_57 = __p0_57; \
44773  int8x16_t __s2_57 = __p2_57; \
44774  int8x16_t __rev0_57; __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44775  int8x16_t __rev2_57; __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44776  int8x16_t __ret_57; \
44777  __ret_57 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_57, __p3_57), __rev0_57, __p1_57); \
44778  __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44779  __ret_57; \
44780 })
44781 #endif
44782 
44783 #ifdef __LITTLE_ENDIAN__
44784 #define vcopyq_laneq_f32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
44785  float32x4_t __s0_58 = __p0_58; \
44786  float32x4_t __s2_58 = __p2_58; \
44787  float32x4_t __ret_58; \
44788  __ret_58 = vsetq_lane_f32(vgetq_lane_f32(__s2_58, __p3_58), __s0_58, __p1_58); \
44789  __ret_58; \
44790 })
44791 #else
44792 #define vcopyq_laneq_f32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
44793  float32x4_t __s0_59 = __p0_59; \
44794  float32x4_t __s2_59 = __p2_59; \
44795  float32x4_t __rev0_59; __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
44796  float32x4_t __rev2_59; __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
44797  float32x4_t __ret_59; \
44798  __ret_59 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_59, __p3_59), __rev0_59, __p1_59); \
44799  __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
44800  __ret_59; \
44801 })
44802 #endif
44803 
44804 #ifdef __LITTLE_ENDIAN__
44805 #define vcopyq_laneq_s32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
44806  int32x4_t __s0_60 = __p0_60; \
44807  int32x4_t __s2_60 = __p2_60; \
44808  int32x4_t __ret_60; \
44809  __ret_60 = vsetq_lane_s32(vgetq_lane_s32(__s2_60, __p3_60), __s0_60, __p1_60); \
44810  __ret_60; \
44811 })
44812 #else
44813 #define vcopyq_laneq_s32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
44814  int32x4_t __s0_61 = __p0_61; \
44815  int32x4_t __s2_61 = __p2_61; \
44816  int32x4_t __rev0_61; __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
44817  int32x4_t __rev2_61; __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
44818  int32x4_t __ret_61; \
44819  __ret_61 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_61, __p3_61), __rev0_61, __p1_61); \
44820  __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
44821  __ret_61; \
44822 })
44823 #endif
44824 
44825 #ifdef __LITTLE_ENDIAN__
44826 #define vcopyq_laneq_s64(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
44827  int64x2_t __s0_62 = __p0_62; \
44828  int64x2_t __s2_62 = __p2_62; \
44829  int64x2_t __ret_62; \
44830  __ret_62 = vsetq_lane_s64(vgetq_lane_s64(__s2_62, __p3_62), __s0_62, __p1_62); \
44831  __ret_62; \
44832 })
44833 #else
44834 #define vcopyq_laneq_s64(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
44835  int64x2_t __s0_63 = __p0_63; \
44836  int64x2_t __s2_63 = __p2_63; \
44837  int64x2_t __rev0_63; __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
44838  int64x2_t __rev2_63; __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
44839  int64x2_t __ret_63; \
44840  __ret_63 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_63, __p3_63), __rev0_63, __p1_63); \
44841  __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
44842  __ret_63; \
44843 })
44844 #endif
44845 
44846 #ifdef __LITTLE_ENDIAN__
44847 #define vcopyq_laneq_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
44848  int16x8_t __s0_64 = __p0_64; \
44849  int16x8_t __s2_64 = __p2_64; \
44850  int16x8_t __ret_64; \
44851  __ret_64 = vsetq_lane_s16(vgetq_lane_s16(__s2_64, __p3_64), __s0_64, __p1_64); \
44852  __ret_64; \
44853 })
44854 #else
44855 #define vcopyq_laneq_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
44856  int16x8_t __s0_65 = __p0_65; \
44857  int16x8_t __s2_65 = __p2_65; \
44858  int16x8_t __rev0_65; __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44859  int16x8_t __rev2_65; __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44860  int16x8_t __ret_65; \
44861  __ret_65 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_65, __p3_65), __rev0_65, __p1_65); \
44862  __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44863  __ret_65; \
44864 })
44865 #endif
44866 
44867 #ifdef __LITTLE_ENDIAN__
44868 #define vcopy_laneq_p8(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
44869  poly8x8_t __s0_66 = __p0_66; \
44870  poly8x16_t __s2_66 = __p2_66; \
44871  poly8x8_t __ret_66; \
44872  __ret_66 = vset_lane_p8(vgetq_lane_p8(__s2_66, __p3_66), __s0_66, __p1_66); \
44873  __ret_66; \
44874 })
44875 #else
44876 #define vcopy_laneq_p8(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
44877  poly8x8_t __s0_67 = __p0_67; \
44878  poly8x16_t __s2_67 = __p2_67; \
44879  poly8x8_t __rev0_67; __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44880  poly8x16_t __rev2_67; __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44881  poly8x8_t __ret_67; \
44882  __ret_67 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_67, __p3_67), __rev0_67, __p1_67); \
44883  __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44884  __ret_67; \
44885 })
44886 #endif
44887 
44888 #ifdef __LITTLE_ENDIAN__
44889 #define vcopy_laneq_p16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
44890  poly16x4_t __s0_68 = __p0_68; \
44891  poly16x8_t __s2_68 = __p2_68; \
44892  poly16x4_t __ret_68; \
44893  __ret_68 = vset_lane_p16(vgetq_lane_p16(__s2_68, __p3_68), __s0_68, __p1_68); \
44894  __ret_68; \
44895 })
44896 #else
44897 #define vcopy_laneq_p16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
44898  poly16x4_t __s0_69 = __p0_69; \
44899  poly16x8_t __s2_69 = __p2_69; \
44900  poly16x4_t __rev0_69; __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
44901  poly16x8_t __rev2_69; __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 7, 6, 5, 4, 3, 2, 1, 0); \
44902  poly16x4_t __ret_69; \
44903  __ret_69 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_69, __p3_69), __rev0_69, __p1_69); \
44904  __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
44905  __ret_69; \
44906 })
44907 #endif
44908 
44909 #ifdef __LITTLE_ENDIAN__
44910 #define vcopy_laneq_u8(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
44911  uint8x8_t __s0_70 = __p0_70; \
44912  uint8x16_t __s2_70 = __p2_70; \
44913  uint8x8_t __ret_70; \
44914  __ret_70 = vset_lane_u8(vgetq_lane_u8(__s2_70, __p3_70), __s0_70, __p1_70); \
44915  __ret_70; \
44916 })
44917 #else
44918 #define vcopy_laneq_u8(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
44919  uint8x8_t __s0_71 = __p0_71; \
44920  uint8x16_t __s2_71 = __p2_71; \
44921  uint8x8_t __rev0_71; __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44922  uint8x16_t __rev2_71; __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44923  uint8x8_t __ret_71; \
44924  __ret_71 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_71, __p3_71), __rev0_71, __p1_71); \
44925  __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44926  __ret_71; \
44927 })
44928 #endif
44929 
44930 #ifdef __LITTLE_ENDIAN__
44931 #define vcopy_laneq_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
44932  uint32x2_t __s0_72 = __p0_72; \
44933  uint32x4_t __s2_72 = __p2_72; \
44934  uint32x2_t __ret_72; \
44935  __ret_72 = vset_lane_u32(vgetq_lane_u32(__s2_72, __p3_72), __s0_72, __p1_72); \
44936  __ret_72; \
44937 })
44938 #else
44939 #define vcopy_laneq_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
44940  uint32x2_t __s0_73 = __p0_73; \
44941  uint32x4_t __s2_73 = __p2_73; \
44942  uint32x2_t __rev0_73; __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
44943  uint32x4_t __rev2_73; __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 3, 2, 1, 0); \
44944  uint32x2_t __ret_73; \
44945  __ret_73 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_73, __p3_73), __rev0_73, __p1_73); \
44946  __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
44947  __ret_73; \
44948 })
44949 #endif
44950 
44951 #ifdef __LITTLE_ENDIAN__
44952 #define vcopy_laneq_u64(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
44953  uint64x1_t __s0_74 = __p0_74; \
44954  uint64x2_t __s2_74 = __p2_74; \
44955  uint64x1_t __ret_74; \
44956  __ret_74 = vset_lane_u64(vgetq_lane_u64(__s2_74, __p3_74), __s0_74, __p1_74); \
44957  __ret_74; \
44958 })
44959 #else
44960 #define vcopy_laneq_u64(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
44961  uint64x1_t __s0_75 = __p0_75; \
44962  uint64x2_t __s2_75 = __p2_75; \
44963  uint64x2_t __rev2_75; __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 1, 0); \
44964  uint64x1_t __ret_75; \
44965  __ret_75 = __noswap_vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_75, __p3_75), __s0_75, __p1_75); \
44966  __ret_75; \
44967 })
44968 #endif
44969 
44970 #ifdef __LITTLE_ENDIAN__
44971 #define vcopy_laneq_u16(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
44972  uint16x4_t __s0_76 = __p0_76; \
44973  uint16x8_t __s2_76 = __p2_76; \
44974  uint16x4_t __ret_76; \
44975  __ret_76 = vset_lane_u16(vgetq_lane_u16(__s2_76, __p3_76), __s0_76, __p1_76); \
44976  __ret_76; \
44977 })
44978 #else
44979 #define vcopy_laneq_u16(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
44980  uint16x4_t __s0_77 = __p0_77; \
44981  uint16x8_t __s2_77 = __p2_77; \
44982  uint16x4_t __rev0_77; __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 3, 2, 1, 0); \
44983  uint16x8_t __rev2_77; __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 7, 6, 5, 4, 3, 2, 1, 0); \
44984  uint16x4_t __ret_77; \
44985  __ret_77 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_77, __p3_77), __rev0_77, __p1_77); \
44986  __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 3, 2, 1, 0); \
44987  __ret_77; \
44988 })
44989 #endif
44990 
44991 #ifdef __LITTLE_ENDIAN__
44992 #define vcopy_laneq_s8(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
44993  int8x8_t __s0_78 = __p0_78; \
44994  int8x16_t __s2_78 = __p2_78; \
44995  int8x8_t __ret_78; \
44996  __ret_78 = vset_lane_s8(vgetq_lane_s8(__s2_78, __p3_78), __s0_78, __p1_78); \
44997  __ret_78; \
44998 })
44999 #else
45000 #define vcopy_laneq_s8(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
45001  int8x8_t __s0_79 = __p0_79; \
45002  int8x16_t __s2_79 = __p2_79; \
45003  int8x8_t __rev0_79; __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45004  int8x16_t __rev2_79; __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45005  int8x8_t __ret_79; \
45006  __ret_79 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_79, __p3_79), __rev0_79, __p1_79); \
45007  __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45008  __ret_79; \
45009 })
45010 #endif
45011 
45012 #ifdef __LITTLE_ENDIAN__
45013 #define vcopy_laneq_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
45014  float32x2_t __s0_80 = __p0_80; \
45015  float32x4_t __s2_80 = __p2_80; \
45016  float32x2_t __ret_80; \
45017  __ret_80 = vset_lane_f32(vgetq_lane_f32(__s2_80, __p3_80), __s0_80, __p1_80); \
45018  __ret_80; \
45019 })
45020 #else
45021 #define vcopy_laneq_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
45022  float32x2_t __s0_81 = __p0_81; \
45023  float32x4_t __s2_81 = __p2_81; \
45024  float32x2_t __rev0_81; __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
45025  float32x4_t __rev2_81; __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
45026  float32x2_t __ret_81; \
45027  __ret_81 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_81, __p3_81), __rev0_81, __p1_81); \
45028  __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
45029  __ret_81; \
45030 })
45031 #endif
45032 
45033 #ifdef __LITTLE_ENDIAN__
45034 #define vcopy_laneq_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
45035  int32x2_t __s0_82 = __p0_82; \
45036  int32x4_t __s2_82 = __p2_82; \
45037  int32x2_t __ret_82; \
45038  __ret_82 = vset_lane_s32(vgetq_lane_s32(__s2_82, __p3_82), __s0_82, __p1_82); \
45039  __ret_82; \
45040 })
45041 #else
45042 #define vcopy_laneq_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
45043  int32x2_t __s0_83 = __p0_83; \
45044  int32x4_t __s2_83 = __p2_83; \
45045  int32x2_t __rev0_83; __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
45046  int32x4_t __rev2_83; __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 3, 2, 1, 0); \
45047  int32x2_t __ret_83; \
45048  __ret_83 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_83, __p3_83), __rev0_83, __p1_83); \
45049  __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
45050  __ret_83; \
45051 })
45052 #endif
45053 
45054 #ifdef __LITTLE_ENDIAN__
45055 #define vcopy_laneq_s64(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
45056  int64x1_t __s0_84 = __p0_84; \
45057  int64x2_t __s2_84 = __p2_84; \
45058  int64x1_t __ret_84; \
45059  __ret_84 = vset_lane_s64(vgetq_lane_s64(__s2_84, __p3_84), __s0_84, __p1_84); \
45060  __ret_84; \
45061 })
45062 #else
45063 #define vcopy_laneq_s64(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
45064  int64x1_t __s0_85 = __p0_85; \
45065  int64x2_t __s2_85 = __p2_85; \
45066  int64x2_t __rev2_85; __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 1, 0); \
45067  int64x1_t __ret_85; \
45068  __ret_85 = __noswap_vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_85, __p3_85), __s0_85, __p1_85); \
45069  __ret_85; \
45070 })
45071 #endif
45072 
45073 #ifdef __LITTLE_ENDIAN__
45074 #define vcopy_laneq_s16(__p0_86, __p1_86, __p2_86, __p3_86) __extension__ ({ \
45075  int16x4_t __s0_86 = __p0_86; \
45076  int16x8_t __s2_86 = __p2_86; \
45077  int16x4_t __ret_86; \
45078  __ret_86 = vset_lane_s16(vgetq_lane_s16(__s2_86, __p3_86), __s0_86, __p1_86); \
45079  __ret_86; \
45080 })
45081 #else
45082 #define vcopy_laneq_s16(__p0_87, __p1_87, __p2_87, __p3_87) __extension__ ({ \
45083  int16x4_t __s0_87 = __p0_87; \
45084  int16x8_t __s2_87 = __p2_87; \
45085  int16x4_t __rev0_87; __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
45086  int16x8_t __rev2_87; __rev2_87 = __builtin_shufflevector(__s2_87, __s2_87, 7, 6, 5, 4, 3, 2, 1, 0); \
45087  int16x4_t __ret_87; \
45088  __ret_87 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_87, __p3_87), __rev0_87, __p1_87); \
45089  __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
45090  __ret_87; \
45091 })
45092 #endif
45093 
45094 #ifdef __LITTLE_ENDIAN__
45095 __ai poly64x1_t vcreate_p64(uint64_t __p0) {
45096  poly64x1_t __ret;
45097  __ret = (poly64x1_t)(__p0);
45098  return __ret;
45099 }
45100 #else
45101 __ai poly64x1_t vcreate_p64(uint64_t __p0) {
45102  poly64x1_t __ret;
45103  __ret = (poly64x1_t)(__p0);
45104  return __ret;
45105 }
45106 #endif
45107 
45108 #ifdef __LITTLE_ENDIAN__
45109 __ai float64x1_t vcreate_f64(uint64_t __p0) {
45110  float64x1_t __ret;
45111  __ret = (float64x1_t)(__p0);
45112  return __ret;
45113 }
45114 #else
45115 __ai float64x1_t vcreate_f64(uint64_t __p0) {
45116  float64x1_t __ret;
45117  __ret = (float64x1_t)(__p0);
45118  return __ret;
45119 }
45120 #endif
45121 
45122 #ifdef __LITTLE_ENDIAN__
45123 __ai float32_t vcvts_f32_s32(int32_t __p0) {
45124  float32_t __ret;
45125  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45126  return __ret;
45127 }
45128 #else
45129 __ai float32_t vcvts_f32_s32(int32_t __p0) {
45130  float32_t __ret;
45131  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45132  return __ret;
45133 }
45134 #endif
45135 
45136 #ifdef __LITTLE_ENDIAN__
45137 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
45138  float32_t __ret;
45139  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45140  return __ret;
45141 }
45142 #else
45143 __ai float32_t vcvts_f32_u32(uint32_t __p0) {
45144  float32_t __ret;
45145  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45146  return __ret;
45147 }
45148 #endif
45149 
45150 #ifdef __LITTLE_ENDIAN__
45151 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45152  float32x2_t __ret;
45153  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45154  return __ret;
45155 }
45156 #else
45157 __ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45158  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45159  float32x2_t __ret;
45160  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
45161  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45162  return __ret;
45163 }
45164 __ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
45165  float32x2_t __ret;
45166  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45167  return __ret;
45168 }
45169 #endif
45170 
45171 #ifdef __LITTLE_ENDIAN__
45172 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
45173  float64_t __ret;
45174  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45175  return __ret;
45176 }
45177 #else
45178 __ai float64_t vcvtd_f64_s64(int64_t __p0) {
45179  float64_t __ret;
45180  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45181  return __ret;
45182 }
45183 #endif
45184 
45185 #ifdef __LITTLE_ENDIAN__
45186 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45187  float64_t __ret;
45188  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45189  return __ret;
45190 }
45191 #else
45192 __ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45193  float64_t __ret;
45194  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45195  return __ret;
45196 }
45197 #endif
45198 
45199 #ifdef __LITTLE_ENDIAN__
45200 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45201  float64x2_t __ret;
45202  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
45203  return __ret;
45204 }
45205 #else
45206 __ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45207  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45208  float64x2_t __ret;
45209  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
45210  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45211  return __ret;
45212 }
45213 #endif
45214 
45215 #ifdef __LITTLE_ENDIAN__
45216 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45217  float64x2_t __ret;
45218  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
45219  return __ret;
45220 }
45221 #else
45222 __ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45223  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45224  float64x2_t __ret;
45225  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
45226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45227  return __ret;
45228 }
45229 #endif
45230 
45231 #ifdef __LITTLE_ENDIAN__
45232 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45233  float64x1_t __ret;
45234  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45235  return __ret;
45236 }
45237 #else
45238 __ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45239  float64x1_t __ret;
45240  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45241  return __ret;
45242 }
45243 #endif
45244 
45245 #ifdef __LITTLE_ENDIAN__
45246 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45247  float64x1_t __ret;
45248  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45249  return __ret;
45250 }
45251 #else
45252 __ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45253  float64x1_t __ret;
45254  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45255  return __ret;
45256 }
45257 #endif
45258 
45259 #ifdef __LITTLE_ENDIAN__
45260 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45261  float64x2_t __ret;
45262  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45263  return __ret;
45264 }
45265 #else
45266 __ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45267  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45268  float64x2_t __ret;
45269  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
45270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45271  return __ret;
45272 }
45273 __ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
45274  float64x2_t __ret;
45275  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45276  return __ret;
45277 }
45278 #endif
45279 
45280 #ifdef __LITTLE_ENDIAN__
45281 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45282  float16x8_t __ret;
45283  __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
45284  return __ret;
45285 }
45286 #else
45287 __ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45288  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45289  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45290  float16x8_t __ret;
45291  __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
45292  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45293  return __ret;
45294 }
45295 #endif
45296 
45297 #ifdef __LITTLE_ENDIAN__
45298 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45299  float32x4_t __ret;
45300  __ret = vcvt_f32_f16(vget_high_f16(__p0));
45301  return __ret;
45302 }
45303 #else
45304 __ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45305  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45306  float32x4_t __ret;
45308  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45309  return __ret;
45310 }
45311 #endif
45312 
45313 #ifdef __LITTLE_ENDIAN__
45314 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45315  float32x4_t __ret;
45316  __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
45317  return __ret;
45318 }
45319 #else
45320 __ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45321  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45322  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45323  float32x4_t __ret;
45324  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
45325  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45326  return __ret;
45327 }
45328 #endif
45329 
45330 #ifdef __LITTLE_ENDIAN__
45331 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45332  float64x2_t __ret;
45333  __ret = vcvt_f64_f32(vget_high_f32(__p0));
45334  return __ret;
45335 }
45336 #else
45337 __ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45338  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45339  float64x2_t __ret;
45340  __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
45341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45342  return __ret;
45343 }
45344 #endif
45345 
45346 #ifdef __LITTLE_ENDIAN__
45347 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45348  uint32_t __s0 = __p0; \
45349  float32_t __ret; \
45350  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45351  __ret; \
45352 })
45353 #else
45354 #define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45355  uint32_t __s0 = __p0; \
45356  float32_t __ret; \
45357  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45358  __ret; \
45359 })
45360 #endif
45361 
45362 #ifdef __LITTLE_ENDIAN__
45363 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45364  int32_t __s0 = __p0; \
45365  float32_t __ret; \
45366  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45367  __ret; \
45368 })
45369 #else
45370 #define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45371  int32_t __s0 = __p0; \
45372  float32_t __ret; \
45373  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45374  __ret; \
45375 })
45376 #endif
45377 
45378 #ifdef __LITTLE_ENDIAN__
45379 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45380  uint64x2_t __s0 = __p0; \
45381  float64x2_t __ret; \
45382  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
45383  __ret; \
45384 })
45385 #else
45386 #define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45387  uint64x2_t __s0 = __p0; \
45388  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45389  float64x2_t __ret; \
45390  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
45391  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45392  __ret; \
45393 })
45394 #endif
45395 
45396 #ifdef __LITTLE_ENDIAN__
45397 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45398  int64x2_t __s0 = __p0; \
45399  float64x2_t __ret; \
45400  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
45401  __ret; \
45402 })
45403 #else
45404 #define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45405  int64x2_t __s0 = __p0; \
45406  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45407  float64x2_t __ret; \
45408  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
45409  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45410  __ret; \
45411 })
45412 #endif
45413 
45414 #ifdef __LITTLE_ENDIAN__
45415 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45416  uint64x1_t __s0 = __p0; \
45417  float64x1_t __ret; \
45418  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45419  __ret; \
45420 })
45421 #else
45422 #define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45423  uint64x1_t __s0 = __p0; \
45424  float64x1_t __ret; \
45425  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45426  __ret; \
45427 })
45428 #endif
45429 
45430 #ifdef __LITTLE_ENDIAN__
45431 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45432  int64x1_t __s0 = __p0; \
45433  float64x1_t __ret; \
45434  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45435  __ret; \
45436 })
45437 #else
45438 #define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45439  int64x1_t __s0 = __p0; \
45440  float64x1_t __ret; \
45441  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45442  __ret; \
45443 })
45444 #endif
45445 
45446 #ifdef __LITTLE_ENDIAN__
45447 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45448  uint64_t __s0 = __p0; \
45449  float64_t __ret; \
45450  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45451  __ret; \
45452 })
45453 #else
45454 #define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45455  uint64_t __s0 = __p0; \
45456  float64_t __ret; \
45457  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45458  __ret; \
45459 })
45460 #endif
45461 
45462 #ifdef __LITTLE_ENDIAN__
45463 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45464  int64_t __s0 = __p0; \
45465  float64_t __ret; \
45466  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45467  __ret; \
45468 })
45469 #else
45470 #define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45471  int64_t __s0 = __p0; \
45472  float64_t __ret; \
45473  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45474  __ret; \
45475 })
45476 #endif
45477 
45478 #ifdef __LITTLE_ENDIAN__
45479 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45480  float32_t __s0 = __p0; \
45481  int32_t __ret; \
45482  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45483  __ret; \
45484 })
45485 #else
45486 #define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45487  float32_t __s0 = __p0; \
45488  int32_t __ret; \
45489  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45490  __ret; \
45491 })
45492 #endif
45493 
45494 #ifdef __LITTLE_ENDIAN__
45495 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45496  float64x2_t __s0 = __p0; \
45497  int64x2_t __ret; \
45498  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
45499  __ret; \
45500 })
45501 #else
45502 #define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45503  float64x2_t __s0 = __p0; \
45504  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45505  int64x2_t __ret; \
45506  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
45507  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45508  __ret; \
45509 })
45510 #endif
45511 
45512 #ifdef __LITTLE_ENDIAN__
45513 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45514  float64x1_t __s0 = __p0; \
45515  int64x1_t __ret; \
45516  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45517  __ret; \
45518 })
45519 #else
45520 #define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45521  float64x1_t __s0 = __p0; \
45522  int64x1_t __ret; \
45523  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45524  __ret; \
45525 })
45526 #endif
45527 
45528 #ifdef __LITTLE_ENDIAN__
45529 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45530  float64_t __s0 = __p0; \
45531  int64_t __ret; \
45532  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45533  __ret; \
45534 })
45535 #else
45536 #define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45537  float64_t __s0 = __p0; \
45538  int64_t __ret; \
45539  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45540  __ret; \
45541 })
45542 #endif
45543 
45544 #ifdef __LITTLE_ENDIAN__
45545 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45546  float32_t __s0 = __p0; \
45547  uint32_t __ret; \
45548  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45549  __ret; \
45550 })
45551 #else
45552 #define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45553  float32_t __s0 = __p0; \
45554  uint32_t __ret; \
45555  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45556  __ret; \
45557 })
45558 #endif
45559 
45560 #ifdef __LITTLE_ENDIAN__
45561 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45562  float64x2_t __s0 = __p0; \
45563  uint64x2_t __ret; \
45564  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
45565  __ret; \
45566 })
45567 #else
45568 #define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45569  float64x2_t __s0 = __p0; \
45570  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45571  uint64x2_t __ret; \
45572  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
45573  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45574  __ret; \
45575 })
45576 #endif
45577 
45578 #ifdef __LITTLE_ENDIAN__
45579 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45580  float64x1_t __s0 = __p0; \
45581  uint64x1_t __ret; \
45582  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45583  __ret; \
45584 })
45585 #else
45586 #define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45587  float64x1_t __s0 = __p0; \
45588  uint64x1_t __ret; \
45589  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45590  __ret; \
45591 })
45592 #endif
45593 
45594 #ifdef __LITTLE_ENDIAN__
45595 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45596  float64_t __s0 = __p0; \
45597  uint64_t __ret; \
45598  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45599  __ret; \
45600 })
45601 #else
45602 #define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45603  float64_t __s0 = __p0; \
45604  uint64_t __ret; \
45605  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45606  __ret; \
45607 })
45608 #endif
45609 
45610 #ifdef __LITTLE_ENDIAN__
45611 __ai int32_t vcvts_s32_f32(float32_t __p0) {
45612  int32_t __ret;
45613  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45614  return __ret;
45615 }
45616 #else
45617 __ai int32_t vcvts_s32_f32(float32_t __p0) {
45618  int32_t __ret;
45619  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45620  return __ret;
45621 }
45622 #endif
45623 
45624 #ifdef __LITTLE_ENDIAN__
45625 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
45626  int64_t __ret;
45627  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45628  return __ret;
45629 }
45630 #else
45631 __ai int64_t vcvtd_s64_f64(float64_t __p0) {
45632  int64_t __ret;
45633  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45634  return __ret;
45635 }
45636 #endif
45637 
45638 #ifdef __LITTLE_ENDIAN__
45639 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45640  int64x2_t __ret;
45641  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
45642  return __ret;
45643 }
45644 #else
45645 __ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45646  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45647  int64x2_t __ret;
45648  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
45649  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45650  return __ret;
45651 }
45652 #endif
45653 
45654 #ifdef __LITTLE_ENDIAN__
45655 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45656  int64x1_t __ret;
45657  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45658  return __ret;
45659 }
45660 #else
45661 __ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45662  int64x1_t __ret;
45663  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45664  return __ret;
45665 }
45666 #endif
45667 
45668 #ifdef __LITTLE_ENDIAN__
45669 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
45670  uint32_t __ret;
45671  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45672  return __ret;
45673 }
45674 #else
45675 __ai uint32_t vcvts_u32_f32(float32_t __p0) {
45676  uint32_t __ret;
45677  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45678  return __ret;
45679 }
45680 #endif
45681 
45682 #ifdef __LITTLE_ENDIAN__
45683 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45684  uint64_t __ret;
45685  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45686  return __ret;
45687 }
45688 #else
45689 __ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45690  uint64_t __ret;
45691  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45692  return __ret;
45693 }
45694 #endif
45695 
45696 #ifdef __LITTLE_ENDIAN__
45697 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45698  uint64x2_t __ret;
45699  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
45700  return __ret;
45701 }
45702 #else
45703 __ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45704  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45705  uint64x2_t __ret;
45706  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
45707  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45708  return __ret;
45709 }
45710 #endif
45711 
45712 #ifdef __LITTLE_ENDIAN__
45713 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45714  uint64x1_t __ret;
45715  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45716  return __ret;
45717 }
45718 #else
45719 __ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45720  uint64x1_t __ret;
45721  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45722  return __ret;
45723 }
45724 #endif
45725 
45726 #ifdef __LITTLE_ENDIAN__
45727 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
45728  int32_t __ret;
45729  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45730  return __ret;
45731 }
45732 #else
45733 __ai int32_t vcvtas_s32_f32(float32_t __p0) {
45734  int32_t __ret;
45735  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45736  return __ret;
45737 }
45738 #endif
45739 
45740 #ifdef __LITTLE_ENDIAN__
45741 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
45742  int64_t __ret;
45743  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45744  return __ret;
45745 }
45746 #else
45747 __ai int64_t vcvtad_s64_f64(float64_t __p0) {
45748  int64_t __ret;
45749  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45750  return __ret;
45751 }
45752 #endif
45753 
45754 #ifdef __LITTLE_ENDIAN__
45755 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45756  uint32_t __ret;
45757  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45758  return __ret;
45759 }
45760 #else
45761 __ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45762  uint32_t __ret;
45763  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45764  return __ret;
45765 }
45766 #endif
45767 
45768 #ifdef __LITTLE_ENDIAN__
45769 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45770  uint64_t __ret;
45771  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45772  return __ret;
45773 }
45774 #else
45775 __ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45776  uint64_t __ret;
45777  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45778  return __ret;
45779 }
45780 #endif
45781 
45782 #ifdef __LITTLE_ENDIAN__
45783 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
45784  int32_t __ret;
45785  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45786  return __ret;
45787 }
45788 #else
45789 __ai int32_t vcvtms_s32_f32(float32_t __p0) {
45790  int32_t __ret;
45791  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45792  return __ret;
45793 }
45794 #endif
45795 
45796 #ifdef __LITTLE_ENDIAN__
45797 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45798  int64_t __ret;
45799  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45800  return __ret;
45801 }
45802 #else
45803 __ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45804  int64_t __ret;
45805  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45806  return __ret;
45807 }
45808 #endif
45809 
45810 #ifdef __LITTLE_ENDIAN__
45811 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45812  uint32_t __ret;
45813  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45814  return __ret;
45815 }
45816 #else
45817 __ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45818  uint32_t __ret;
45819  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45820  return __ret;
45821 }
45822 #endif
45823 
45824 #ifdef __LITTLE_ENDIAN__
45825 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45826  uint64_t __ret;
45827  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45828  return __ret;
45829 }
45830 #else
45831 __ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45832  uint64_t __ret;
45833  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45834  return __ret;
45835 }
45836 #endif
45837 
45838 #ifdef __LITTLE_ENDIAN__
45839 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
45840  int32_t __ret;
45841  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45842  return __ret;
45843 }
45844 #else
45845 __ai int32_t vcvtns_s32_f32(float32_t __p0) {
45846  int32_t __ret;
45847  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45848  return __ret;
45849 }
45850 #endif
45851 
45852 #ifdef __LITTLE_ENDIAN__
45853 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45854  int64_t __ret;
45855  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45856  return __ret;
45857 }
45858 #else
45859 __ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45860  int64_t __ret;
45861  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45862  return __ret;
45863 }
45864 #endif
45865 
45866 #ifdef __LITTLE_ENDIAN__
45867 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45868  uint32_t __ret;
45869  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45870  return __ret;
45871 }
45872 #else
45873 __ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45874  uint32_t __ret;
45875  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45876  return __ret;
45877 }
45878 #endif
45879 
45880 #ifdef __LITTLE_ENDIAN__
45881 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45882  uint64_t __ret;
45883  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45884  return __ret;
45885 }
45886 #else
45887 __ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45888  uint64_t __ret;
45889  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45890  return __ret;
45891 }
45892 #endif
45893 
45894 #ifdef __LITTLE_ENDIAN__
45895 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
45896  int32_t __ret;
45897  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45898  return __ret;
45899 }
45900 #else
45901 __ai int32_t vcvtps_s32_f32(float32_t __p0) {
45902  int32_t __ret;
45903  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45904  return __ret;
45905 }
45906 #endif
45907 
45908 #ifdef __LITTLE_ENDIAN__
45909 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45910  int64_t __ret;
45911  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45912  return __ret;
45913 }
45914 #else
45915 __ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45916  int64_t __ret;
45917  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45918  return __ret;
45919 }
45920 #endif
45921 
45922 #ifdef __LITTLE_ENDIAN__
45923 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45924  uint32_t __ret;
45925  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45926  return __ret;
45927 }
45928 #else
45929 __ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45930  uint32_t __ret;
45931  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45932  return __ret;
45933 }
45934 #endif
45935 
45936 #ifdef __LITTLE_ENDIAN__
45937 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45938  uint64_t __ret;
45939  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45940  return __ret;
45941 }
45942 #else
45943 __ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45944  uint64_t __ret;
45945  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45946  return __ret;
45947 }
45948 #endif
45949 
45950 #ifdef __LITTLE_ENDIAN__
45951 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45952  float32_t __ret;
45953  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45954  return __ret;
45955 }
45956 #else
45957 __ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45958  float32_t __ret;
45959  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45960  return __ret;
45961 }
45962 #endif
45963 
45964 #ifdef __LITTLE_ENDIAN__
45965 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45966  float32x2_t __ret;
45967  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45968  return __ret;
45969 }
45970 #else
45971 __ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45972  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45973  float32x2_t __ret;
45974  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
45975  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45976  return __ret;
45977 }
45978 __ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
45979  float32x2_t __ret;
45980  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45981  return __ret;
45982 }
45983 #endif
45984 
45985 #ifdef __LITTLE_ENDIAN__
45986 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45987  float32x4_t __ret;
45988  __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
45989  return __ret;
45990 }
45991 #else
45992 __ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45993  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45994  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45995  float32x4_t __ret;
45996  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
45997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45998  return __ret;
45999 }
46000 #endif
46001 
46002 #ifdef __LITTLE_ENDIAN__
46003 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46004  float64x2_t __ret;
46005  __ret = __p0 / __p1;
46006  return __ret;
46007 }
46008 #else
46009 __ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46010  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46011  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46012  float64x2_t __ret;
46013  __ret = __rev0 / __rev1;
46014  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46015  return __ret;
46016 }
46017 #endif
46018 
46019 #ifdef __LITTLE_ENDIAN__
46020 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46021  float32x4_t __ret;
46022  __ret = __p0 / __p1;
46023  return __ret;
46024 }
46025 #else
46026 __ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46027  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46028  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46029  float32x4_t __ret;
46030  __ret = __rev0 / __rev1;
46031  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46032  return __ret;
46033 }
46034 #endif
46035 
46036 #ifdef __LITTLE_ENDIAN__
46037 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46038  float64x1_t __ret;
46039  __ret = __p0 / __p1;
46040  return __ret;
46041 }
46042 #else
46043 __ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46044  float64x1_t __ret;
46045  __ret = __p0 / __p1;
46046  return __ret;
46047 }
46048 #endif
46049 
46050 #ifdef __LITTLE_ENDIAN__
46051 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46052  float32x2_t __ret;
46053  __ret = __p0 / __p1;
46054  return __ret;
46055 }
46056 #else
46057 __ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46058  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46059  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46060  float32x2_t __ret;
46061  __ret = __rev0 / __rev1;
46062  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46063  return __ret;
46064 }
46065 #endif
46066 
46067 #ifdef __LITTLE_ENDIAN__
46068 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46069  poly8x8_t __s0 = __p0; \
46070  poly8_t __ret; \
46071  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46072  __ret; \
46073 })
46074 #else
46075 #define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46076  poly8x8_t __s0 = __p0; \
46077  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46078  poly8_t __ret; \
46079  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46080  __ret; \
46081 })
46082 #endif
46083 
46084 #ifdef __LITTLE_ENDIAN__
46085 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46086  poly16x4_t __s0 = __p0; \
46087  poly16_t __ret; \
46088  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46089  __ret; \
46090 })
46091 #else
46092 #define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46093  poly16x4_t __s0 = __p0; \
46094  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46095  poly16_t __ret; \
46096  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46097  __ret; \
46098 })
46099 #endif
46100 
46101 #ifdef __LITTLE_ENDIAN__
46102 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46103  uint8x8_t __s0 = __p0; \
46104  uint8_t __ret; \
46105  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46106  __ret; \
46107 })
46108 #else
46109 #define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46110  uint8x8_t __s0 = __p0; \
46111  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46112  uint8_t __ret; \
46113  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46114  __ret; \
46115 })
46116 #endif
46117 
46118 #ifdef __LITTLE_ENDIAN__
46119 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46120  uint32x2_t __s0 = __p0; \
46121  uint32_t __ret; \
46122  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46123  __ret; \
46124 })
46125 #else
46126 #define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46127  uint32x2_t __s0 = __p0; \
46128  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46129  uint32_t __ret; \
46130  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46131  __ret; \
46132 })
46133 #endif
46134 
46135 #ifdef __LITTLE_ENDIAN__
46136 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46137  uint64x1_t __s0 = __p0; \
46138  uint64_t __ret; \
46139  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46140  __ret; \
46141 })
46142 #else
46143 #define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46144  uint64x1_t __s0 = __p0; \
46145  uint64_t __ret; \
46146  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46147  __ret; \
46148 })
46149 #endif
46150 
46151 #ifdef __LITTLE_ENDIAN__
46152 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46153  uint16x4_t __s0 = __p0; \
46154  uint16_t __ret; \
46155  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46156  __ret; \
46157 })
46158 #else
46159 #define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46160  uint16x4_t __s0 = __p0; \
46161  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46162  uint16_t __ret; \
46163  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46164  __ret; \
46165 })
46166 #endif
46167 
46168 #ifdef __LITTLE_ENDIAN__
46169 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46170  int8x8_t __s0 = __p0; \
46171  int8_t __ret; \
46172  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46173  __ret; \
46174 })
46175 #else
46176 #define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46177  int8x8_t __s0 = __p0; \
46178  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46179  int8_t __ret; \
46180  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46181  __ret; \
46182 })
46183 #endif
46184 
46185 #ifdef __LITTLE_ENDIAN__
46186 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46187  float64x1_t __s0 = __p0; \
46188  float64_t __ret; \
46189  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46190  __ret; \
46191 })
46192 #else
46193 #define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46194  float64x1_t __s0 = __p0; \
46195  float64_t __ret; \
46196  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46197  __ret; \
46198 })
46199 #endif
46200 
46201 #ifdef __LITTLE_ENDIAN__
46202 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46203  float32x2_t __s0 = __p0; \
46204  float32_t __ret; \
46205  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__s0, __p1); \
46206  __ret; \
46207 })
46208 #else
46209 #define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46210  float32x2_t __s0 = __p0; \
46211  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46212  float32_t __ret; \
46213  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__rev0, __p1); \
46214  __ret; \
46215 })
46216 #endif
46217 
46218 #ifdef __LITTLE_ENDIAN__
46219 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46220  int32x2_t __s0 = __p0; \
46221  int32_t __ret; \
46222  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46223  __ret; \
46224 })
46225 #else
46226 #define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46227  int32x2_t __s0 = __p0; \
46228  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46229  int32_t __ret; \
46230  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46231  __ret; \
46232 })
46233 #endif
46234 
46235 #ifdef __LITTLE_ENDIAN__
46236 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46237  int64x1_t __s0 = __p0; \
46238  int64_t __ret; \
46239  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46240  __ret; \
46241 })
46242 #else
46243 #define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46244  int64x1_t __s0 = __p0; \
46245  int64_t __ret; \
46246  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46247  __ret; \
46248 })
46249 #endif
46250 
46251 #ifdef __LITTLE_ENDIAN__
46252 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46253  int16x4_t __s0 = __p0; \
46254  int16_t __ret; \
46255  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46256  __ret; \
46257 })
46258 #else
46259 #define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46260  int16x4_t __s0 = __p0; \
46261  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46262  int16_t __ret; \
46263  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46264  __ret; \
46265 })
46266 #endif
46267 
46268 #ifdef __LITTLE_ENDIAN__
46269 #define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46270  poly64x1_t __s0 = __p0; \
46271  poly64x1_t __ret; \
46272  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46273  __ret; \
46274 })
46275 #else
46276 #define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46277  poly64x1_t __s0 = __p0; \
46278  poly64x1_t __ret; \
46279  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46280  __ret; \
46281 })
46282 #endif
46283 
46284 #ifdef __LITTLE_ENDIAN__
46285 #define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46286  poly64x1_t __s0 = __p0; \
46287  poly64x2_t __ret; \
46288  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46289  __ret; \
46290 })
46291 #else
46292 #define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46293  poly64x1_t __s0 = __p0; \
46294  poly64x2_t __ret; \
46295  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46296  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46297  __ret; \
46298 })
46299 #endif
46300 
46301 #ifdef __LITTLE_ENDIAN__
46302 #define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46303  float64x1_t __s0 = __p0; \
46304  float64x2_t __ret; \
46305  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46306  __ret; \
46307 })
46308 #else
46309 #define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46310  float64x1_t __s0 = __p0; \
46311  float64x2_t __ret; \
46312  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46313  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46314  __ret; \
46315 })
46316 #endif
46317 
46318 #ifdef __LITTLE_ENDIAN__
46319 #define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46320  float16x4_t __s0 = __p0; \
46321  float16x8_t __ret; \
46322  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46323  __ret; \
46324 })
46325 #else
46326 #define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46327  float16x4_t __s0 = __p0; \
46328  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46329  float16x8_t __ret; \
46330  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46331  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46332  __ret; \
46333 })
46334 #endif
46335 
46336 #ifdef __LITTLE_ENDIAN__
46337 #define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46338  float64x1_t __s0 = __p0; \
46339  float64x1_t __ret; \
46340  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46341  __ret; \
46342 })
46343 #else
46344 #define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46345  float64x1_t __s0 = __p0; \
46346  float64x1_t __ret; \
46347  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46348  __ret; \
46349 })
46350 #endif
46351 
46352 #ifdef __LITTLE_ENDIAN__
46353 #define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46354  float16x4_t __s0 = __p0; \
46355  float16x4_t __ret; \
46356  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46357  __ret; \
46358 })
46359 #else
46360 #define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46361  float16x4_t __s0 = __p0; \
46362  float16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46363  float16x4_t __ret; \
46364  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46366  __ret; \
46367 })
46368 #endif
46369 
46370 #ifdef __LITTLE_ENDIAN__
46371 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46372  poly8x16_t __s0 = __p0; \
46373  poly8_t __ret; \
46374  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46375  __ret; \
46376 })
46377 #else
46378 #define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46379  poly8x16_t __s0 = __p0; \
46380  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46381  poly8_t __ret; \
46382  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46383  __ret; \
46384 })
46385 #endif
46386 
46387 #ifdef __LITTLE_ENDIAN__
46388 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46389  poly16x8_t __s0 = __p0; \
46390  poly16_t __ret; \
46391  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46392  __ret; \
46393 })
46394 #else
46395 #define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46396  poly16x8_t __s0 = __p0; \
46397  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46398  poly16_t __ret; \
46399  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46400  __ret; \
46401 })
46402 #endif
46403 
46404 #ifdef __LITTLE_ENDIAN__
46405 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46406  uint8x16_t __s0 = __p0; \
46407  uint8_t __ret; \
46408  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46409  __ret; \
46410 })
46411 #else
46412 #define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46413  uint8x16_t __s0 = __p0; \
46414  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46415  uint8_t __ret; \
46416  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46417  __ret; \
46418 })
46419 #endif
46420 
46421 #ifdef __LITTLE_ENDIAN__
46422 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46423  uint32x4_t __s0 = __p0; \
46424  uint32_t __ret; \
46425  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46426  __ret; \
46427 })
46428 #else
46429 #define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46430  uint32x4_t __s0 = __p0; \
46431  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46432  uint32_t __ret; \
46433  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46434  __ret; \
46435 })
46436 #endif
46437 
46438 #ifdef __LITTLE_ENDIAN__
46439 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46440  uint64x2_t __s0 = __p0; \
46441  uint64_t __ret; \
46442  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46443  __ret; \
46444 })
46445 #else
46446 #define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46447  uint64x2_t __s0 = __p0; \
46448  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46449  uint64_t __ret; \
46450  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46451  __ret; \
46452 })
46453 #endif
46454 
46455 #ifdef __LITTLE_ENDIAN__
46456 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46457  uint16x8_t __s0 = __p0; \
46458  uint16_t __ret; \
46459  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46460  __ret; \
46461 })
46462 #else
46463 #define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46464  uint16x8_t __s0 = __p0; \
46465  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46466  uint16_t __ret; \
46467  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46468  __ret; \
46469 })
46470 #endif
46471 
46472 #ifdef __LITTLE_ENDIAN__
46473 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46474  int8x16_t __s0 = __p0; \
46475  int8_t __ret; \
46476  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46477  __ret; \
46478 })
46479 #else
46480 #define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46481  int8x16_t __s0 = __p0; \
46482  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46483  int8_t __ret; \
46484  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46485  __ret; \
46486 })
46487 #endif
46488 
46489 #ifdef __LITTLE_ENDIAN__
46490 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46491  float64x2_t __s0 = __p0; \
46492  float64_t __ret; \
46493  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__s0, __p1); \
46494  __ret; \
46495 })
46496 #else
46497 #define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46498  float64x2_t __s0 = __p0; \
46499  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46500  float64_t __ret; \
46501  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__rev0, __p1); \
46502  __ret; \
46503 })
46504 #endif
46505 
46506 #ifdef __LITTLE_ENDIAN__
46507 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46508  float32x4_t __s0 = __p0; \
46509  float32_t __ret; \
46510  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__s0, __p1); \
46511  __ret; \
46512 })
46513 #else
46514 #define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46515  float32x4_t __s0 = __p0; \
46516  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46517  float32_t __ret; \
46518  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__rev0, __p1); \
46519  __ret; \
46520 })
46521 #endif
46522 
46523 #ifdef __LITTLE_ENDIAN__
46524 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46525  int32x4_t __s0 = __p0; \
46526  int32_t __ret; \
46527  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46528  __ret; \
46529 })
46530 #else
46531 #define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46532  int32x4_t __s0 = __p0; \
46533  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46534  int32_t __ret; \
46535  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46536  __ret; \
46537 })
46538 #endif
46539 
46540 #ifdef __LITTLE_ENDIAN__
46541 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46542  int64x2_t __s0 = __p0; \
46543  int64_t __ret; \
46544  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46545  __ret; \
46546 })
46547 #else
46548 #define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46549  int64x2_t __s0 = __p0; \
46550  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46551  int64_t __ret; \
46552  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46553  __ret; \
46554 })
46555 #endif
46556 
46557 #ifdef __LITTLE_ENDIAN__
46558 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46559  int16x8_t __s0 = __p0; \
46560  int16_t __ret; \
46561  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46562  __ret; \
46563 })
46564 #else
46565 #define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46566  int16x8_t __s0 = __p0; \
46567  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46568  int16_t __ret; \
46569  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46570  __ret; \
46571 })
46572 #endif
46573 
46574 #ifdef __LITTLE_ENDIAN__
46575 #define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46576  poly8x16_t __s0 = __p0; \
46577  poly8x8_t __ret; \
46578  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46579  __ret; \
46580 })
46581 #else
46582 #define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46583  poly8x16_t __s0 = __p0; \
46584  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46585  poly8x8_t __ret; \
46586  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46587  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46588  __ret; \
46589 })
46590 #endif
46591 
46592 #ifdef __LITTLE_ENDIAN__
46593 #define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46594  poly64x2_t __s0 = __p0; \
46595  poly64x1_t __ret; \
46596  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46597  __ret; \
46598 })
46599 #else
46600 #define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46601  poly64x2_t __s0 = __p0; \
46602  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46603  poly64x1_t __ret; \
46604  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46605  __ret; \
46606 })
46607 #endif
46608 
46609 #ifdef __LITTLE_ENDIAN__
46610 #define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46611  poly16x8_t __s0 = __p0; \
46612  poly16x4_t __ret; \
46613  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46614  __ret; \
46615 })
46616 #else
46617 #define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46618  poly16x8_t __s0 = __p0; \
46619  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46620  poly16x4_t __ret; \
46621  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46623  __ret; \
46624 })
46625 #endif
46626 
46627 #ifdef __LITTLE_ENDIAN__
46628 #define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46629  poly8x16_t __s0 = __p0; \
46630  poly8x16_t __ret; \
46631  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46632  __ret; \
46633 })
46634 #else
46635 #define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46636  poly8x16_t __s0 = __p0; \
46637  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46638  poly8x16_t __ret; \
46639  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46640  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46641  __ret; \
46642 })
46643 #endif
46644 
46645 #ifdef __LITTLE_ENDIAN__
46646 #define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46647  poly64x2_t __s0 = __p0; \
46648  poly64x2_t __ret; \
46649  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46650  __ret; \
46651 })
46652 #else
46653 #define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46654  poly64x2_t __s0 = __p0; \
46655  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46656  poly64x2_t __ret; \
46657  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46659  __ret; \
46660 })
46661 #endif
46662 
46663 #ifdef __LITTLE_ENDIAN__
46664 #define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46665  poly16x8_t __s0 = __p0; \
46666  poly16x8_t __ret; \
46667  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46668  __ret; \
46669 })
46670 #else
46671 #define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46672  poly16x8_t __s0 = __p0; \
46673  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46674  poly16x8_t __ret; \
46675  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46676  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46677  __ret; \
46678 })
46679 #endif
46680 
46681 #ifdef __LITTLE_ENDIAN__
46682 #define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46683  uint8x16_t __s0 = __p0; \
46684  uint8x16_t __ret; \
46685  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46686  __ret; \
46687 })
46688 #else
46689 #define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46690  uint8x16_t __s0 = __p0; \
46691  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46692  uint8x16_t __ret; \
46693  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46694  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46695  __ret; \
46696 })
46697 #endif
46698 
46699 #ifdef __LITTLE_ENDIAN__
46700 #define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46701  uint32x4_t __s0 = __p0; \
46702  uint32x4_t __ret; \
46703  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46704  __ret; \
46705 })
46706 #else
46707 #define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46708  uint32x4_t __s0 = __p0; \
46709  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46710  uint32x4_t __ret; \
46711  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46712  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46713  __ret; \
46714 })
46715 #endif
46716 
46717 #ifdef __LITTLE_ENDIAN__
46718 #define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46719  uint64x2_t __s0 = __p0; \
46720  uint64x2_t __ret; \
46721  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46722  __ret; \
46723 })
46724 #else
46725 #define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46726  uint64x2_t __s0 = __p0; \
46727  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46728  uint64x2_t __ret; \
46729  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46730  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46731  __ret; \
46732 })
46733 #endif
46734 
46735 #ifdef __LITTLE_ENDIAN__
46736 #define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46737  uint16x8_t __s0 = __p0; \
46738  uint16x8_t __ret; \
46739  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46740  __ret; \
46741 })
46742 #else
46743 #define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46744  uint16x8_t __s0 = __p0; \
46745  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46746  uint16x8_t __ret; \
46747  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46749  __ret; \
46750 })
46751 #endif
46752 
46753 #ifdef __LITTLE_ENDIAN__
46754 #define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46755  int8x16_t __s0 = __p0; \
46756  int8x16_t __ret; \
46757  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46758  __ret; \
46759 })
46760 #else
46761 #define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46762  int8x16_t __s0 = __p0; \
46763  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46764  int8x16_t __ret; \
46765  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46766  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46767  __ret; \
46768 })
46769 #endif
46770 
46771 #ifdef __LITTLE_ENDIAN__
46772 #define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46773  float64x2_t __s0 = __p0; \
46774  float64x2_t __ret; \
46775  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46776  __ret; \
46777 })
46778 #else
46779 #define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46780  float64x2_t __s0 = __p0; \
46781  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46782  float64x2_t __ret; \
46783  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46784  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46785  __ret; \
46786 })
46787 #endif
46788 
46789 #ifdef __LITTLE_ENDIAN__
46790 #define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46791  float32x4_t __s0 = __p0; \
46792  float32x4_t __ret; \
46793  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46794  __ret; \
46795 })
46796 #else
46797 #define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46798  float32x4_t __s0 = __p0; \
46799  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46800  float32x4_t __ret; \
46801  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46803  __ret; \
46804 })
46805 #endif
46806 
46807 #ifdef __LITTLE_ENDIAN__
46808 #define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46809  float16x8_t __s0 = __p0; \
46810  float16x8_t __ret; \
46811  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46812  __ret; \
46813 })
46814 #else
46815 #define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46816  float16x8_t __s0 = __p0; \
46817  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46818  float16x8_t __ret; \
46819  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46821  __ret; \
46822 })
46823 #endif
46824 
46825 #ifdef __LITTLE_ENDIAN__
46826 #define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46827  int32x4_t __s0 = __p0; \
46828  int32x4_t __ret; \
46829  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46830  __ret; \
46831 })
46832 #else
46833 #define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46834  int32x4_t __s0 = __p0; \
46835  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46836  int32x4_t __ret; \
46837  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46839  __ret; \
46840 })
46841 #endif
46842 
46843 #ifdef __LITTLE_ENDIAN__
46844 #define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46845  int64x2_t __s0 = __p0; \
46846  int64x2_t __ret; \
46847  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46848  __ret; \
46849 })
46850 #else
46851 #define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46852  int64x2_t __s0 = __p0; \
46853  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46854  int64x2_t __ret; \
46855  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46856  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46857  __ret; \
46858 })
46859 #endif
46860 
46861 #ifdef __LITTLE_ENDIAN__
46862 #define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46863  int16x8_t __s0 = __p0; \
46864  int16x8_t __ret; \
46865  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46866  __ret; \
46867 })
46868 #else
46869 #define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46870  int16x8_t __s0 = __p0; \
46871  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46872  int16x8_t __ret; \
46873  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46874  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46875  __ret; \
46876 })
46877 #endif
46878 
46879 #ifdef __LITTLE_ENDIAN__
46880 #define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46881  uint8x16_t __s0 = __p0; \
46882  uint8x8_t __ret; \
46883  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46884  __ret; \
46885 })
46886 #else
46887 #define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46888  uint8x16_t __s0 = __p0; \
46889  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46890  uint8x8_t __ret; \
46891  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46892  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46893  __ret; \
46894 })
46895 #endif
46896 
46897 #ifdef __LITTLE_ENDIAN__
46898 #define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46899  uint32x4_t __s0 = __p0; \
46900  uint32x2_t __ret; \
46901  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46902  __ret; \
46903 })
46904 #else
46905 #define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46906  uint32x4_t __s0 = __p0; \
46907  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46908  uint32x2_t __ret; \
46909  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46910  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46911  __ret; \
46912 })
46913 #endif
46914 
46915 #ifdef __LITTLE_ENDIAN__
46916 #define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46917  uint64x2_t __s0 = __p0; \
46918  uint64x1_t __ret; \
46919  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46920  __ret; \
46921 })
46922 #else
46923 #define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46924  uint64x2_t __s0 = __p0; \
46925  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46926  uint64x1_t __ret; \
46927  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46928  __ret; \
46929 })
46930 #endif
46931 
46932 #ifdef __LITTLE_ENDIAN__
46933 #define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46934  uint16x8_t __s0 = __p0; \
46935  uint16x4_t __ret; \
46936  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46937  __ret; \
46938 })
46939 #else
46940 #define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46941  uint16x8_t __s0 = __p0; \
46942  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46943  uint16x4_t __ret; \
46944  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46945  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46946  __ret; \
46947 })
46948 #endif
46949 
46950 #ifdef __LITTLE_ENDIAN__
46951 #define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46952  int8x16_t __s0 = __p0; \
46953  int8x8_t __ret; \
46954  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46955  __ret; \
46956 })
46957 #else
46958 #define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46959  int8x16_t __s0 = __p0; \
46960  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46961  int8x8_t __ret; \
46962  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46964  __ret; \
46965 })
46966 #endif
46967 
46968 #ifdef __LITTLE_ENDIAN__
46969 #define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
46970  float64x2_t __s0 = __p0; \
46971  float64x1_t __ret; \
46972  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46973  __ret; \
46974 })
46975 #else
46976 #define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
46977  float64x2_t __s0 = __p0; \
46978  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46979  float64x1_t __ret; \
46980  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46981  __ret; \
46982 })
46983 #endif
46984 
46985 #ifdef __LITTLE_ENDIAN__
46986 #define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
46987  float32x4_t __s0 = __p0; \
46988  float32x2_t __ret; \
46989  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46990  __ret; \
46991 })
46992 #else
46993 #define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
46994  float32x4_t __s0 = __p0; \
46995  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46996  float32x2_t __ret; \
46997  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46998  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46999  __ret; \
47000 })
47001 #endif
47002 
47003 #ifdef __LITTLE_ENDIAN__
47004 #define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47005  float16x8_t __s0 = __p0; \
47006  float16x4_t __ret; \
47007  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47008  __ret; \
47009 })
47010 #else
47011 #define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47012  float16x8_t __s0 = __p0; \
47013  float16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47014  float16x4_t __ret; \
47015  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47016  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47017  __ret; \
47018 })
47019 #endif
47020 
47021 #ifdef __LITTLE_ENDIAN__
47022 #define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47023  int32x4_t __s0 = __p0; \
47024  int32x2_t __ret; \
47025  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
47026  __ret; \
47027 })
47028 #else
47029 #define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47030  int32x4_t __s0 = __p0; \
47031  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47032  int32x2_t __ret; \
47033  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
47034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47035  __ret; \
47036 })
47037 #endif
47038 
47039 #ifdef __LITTLE_ENDIAN__
47040 #define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47041  int64x2_t __s0 = __p0; \
47042  int64x1_t __ret; \
47043  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
47044  __ret; \
47045 })
47046 #else
47047 #define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47048  int64x2_t __s0 = __p0; \
47049  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47050  int64x1_t __ret; \
47051  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
47052  __ret; \
47053 })
47054 #endif
47055 
47056 #ifdef __LITTLE_ENDIAN__
47057 #define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47058  int16x8_t __s0 = __p0; \
47059  int16x4_t __ret; \
47060  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47061  __ret; \
47062 })
47063 #else
47064 #define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47065  int16x8_t __s0 = __p0; \
47066  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47067  int16x4_t __ret; \
47068  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47069  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47070  __ret; \
47071 })
47072 #endif
47073 
47074 #ifdef __LITTLE_ENDIAN__
47075 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47076  poly64x1_t __ret;
47077  __ret = (poly64x1_t) {__p0};
47078  return __ret;
47079 }
47080 #else
47081 __ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47082  poly64x1_t __ret;
47083  __ret = (poly64x1_t) {__p0};
47084  return __ret;
47085 }
47086 #endif
47087 
47088 #ifdef __LITTLE_ENDIAN__
47089 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47090  poly64x2_t __ret;
47091  __ret = (poly64x2_t) {__p0, __p0};
47092  return __ret;
47093 }
47094 #else
47095 __ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47096  poly64x2_t __ret;
47097  __ret = (poly64x2_t) {__p0, __p0};
47098  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47099  return __ret;
47100 }
47101 #endif
47102 
47103 #ifdef __LITTLE_ENDIAN__
47104 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
47105  float64x2_t __ret;
47106  __ret = (float64x2_t) {__p0, __p0};
47107  return __ret;
47108 }
47109 #else
47110 __ai float64x2_t vdupq_n_f64(float64_t __p0) {
47111  float64x2_t __ret;
47112  __ret = (float64x2_t) {__p0, __p0};
47113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47114  return __ret;
47115 }
47116 #endif
47117 
47118 #ifdef __LITTLE_ENDIAN__
47119 __ai float64x1_t vdup_n_f64(float64_t __p0) {
47120  float64x1_t __ret;
47121  __ret = (float64x1_t) {__p0};
47122  return __ret;
47123 }
47124 #else
47125 __ai float64x1_t vdup_n_f64(float64_t __p0) {
47126  float64x1_t __ret;
47127  __ret = (float64x1_t) {__p0};
47128  return __ret;
47129 }
47130 #endif
47131 
47132 #ifdef __LITTLE_ENDIAN__
47133 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47134  poly64x1_t __s0 = __p0; \
47135  poly64x1_t __s1 = __p1; \
47136  poly64x1_t __ret; \
47137  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47138  __ret; \
47139 })
47140 #else
47141 #define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47142  poly64x1_t __s0 = __p0; \
47143  poly64x1_t __s1 = __p1; \
47144  poly64x1_t __ret; \
47145  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47146  __ret; \
47147 })
47148 #endif
47149 
47150 #ifdef __LITTLE_ENDIAN__
47151 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47152  poly64x2_t __s0 = __p0; \
47153  poly64x2_t __s1 = __p1; \
47154  poly64x2_t __ret; \
47155  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
47156  __ret; \
47157 })
47158 #else
47159 #define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47160  poly64x2_t __s0 = __p0; \
47161  poly64x2_t __s1 = __p1; \
47162  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47163  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47164  poly64x2_t __ret; \
47165  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
47166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47167  __ret; \
47168 })
47169 #endif
47170 
47171 #ifdef __LITTLE_ENDIAN__
47172 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47173  float64x2_t __s0 = __p0; \
47174  float64x2_t __s1 = __p1; \
47175  float64x2_t __ret; \
47176  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
47177  __ret; \
47178 })
47179 #else
47180 #define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47181  float64x2_t __s0 = __p0; \
47182  float64x2_t __s1 = __p1; \
47183  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47184  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47185  float64x2_t __ret; \
47186  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
47187  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47188  __ret; \
47189 })
47190 #endif
47191 
47192 #ifdef __LITTLE_ENDIAN__
47193 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47194  float64x1_t __s0 = __p0; \
47195  float64x1_t __s1 = __p1; \
47196  float64x1_t __ret; \
47197  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47198  __ret; \
47199 })
47200 #else
47201 #define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47202  float64x1_t __s0 = __p0; \
47203  float64x1_t __s1 = __p1; \
47204  float64x1_t __ret; \
47205  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47206  __ret; \
47207 })
47208 #endif
47209 
47210 #ifdef __LITTLE_ENDIAN__
47211 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47212  float64x2_t __ret;
47213  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47214  return __ret;
47215 }
47216 #else
47217 __ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47218  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47219  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47220  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47221  float64x2_t __ret;
47222  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
47223  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47224  return __ret;
47225 }
47226 __ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47227  float64x2_t __ret;
47228  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47229  return __ret;
47230 }
47231 #endif
47232 
47233 #ifdef __LITTLE_ENDIAN__
47234 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47235  float64x1_t __ret;
47236  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47237  return __ret;
47238 }
47239 #else
47240 __ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47241  float64x1_t __ret;
47242  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47243  return __ret;
47244 }
47245 #endif
47246 
47247 #ifdef __LITTLE_ENDIAN__
47248 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47249  float64_t __s0 = __p0; \
47250  float64_t __s1 = __p1; \
47251  float64x1_t __s2 = __p2; \
47252  float64_t __ret; \
47253  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47254  __ret; \
47255 })
47256 #else
47257 #define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47258  float64_t __s0 = __p0; \
47259  float64_t __s1 = __p1; \
47260  float64x1_t __s2 = __p2; \
47261  float64_t __ret; \
47262  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47263  __ret; \
47264 })
47265 #define __noswap_vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47266  float64_t __s0 = __p0; \
47267  float64_t __s1 = __p1; \
47268  float64x1_t __s2 = __p2; \
47269  float64_t __ret; \
47270  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47271  __ret; \
47272 })
47273 #endif
47274 
47275 #ifdef __LITTLE_ENDIAN__
47276 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47277  float32_t __s0 = __p0; \
47278  float32_t __s1 = __p1; \
47279  float32x2_t __s2 = __p2; \
47280  float32_t __ret; \
47281  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47282  __ret; \
47283 })
47284 #else
47285 #define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47286  float32_t __s0 = __p0; \
47287  float32_t __s1 = __p1; \
47288  float32x2_t __s2 = __p2; \
47289  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47290  float32_t __ret; \
47291  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__rev2, __p3); \
47292  __ret; \
47293 })
47294 #define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47295  float32_t __s0 = __p0; \
47296  float32_t __s1 = __p1; \
47297  float32x2_t __s2 = __p2; \
47298  float32_t __ret; \
47299  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47300  __ret; \
47301 })
47302 #endif
47303 
47304 #ifdef __LITTLE_ENDIAN__
47305 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47306  float64x2_t __s0 = __p0; \
47307  float64x2_t __s1 = __p1; \
47308  float64x1_t __s2 = __p2; \
47309  float64x2_t __ret; \
47310  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47311  __ret; \
47312 })
47313 #else
47314 #define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47315  float64x2_t __s0 = __p0; \
47316  float64x2_t __s1 = __p1; \
47317  float64x1_t __s2 = __p2; \
47318  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47319  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47320  float64x2_t __ret; \
47321  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
47322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47323  __ret; \
47324 })
47325 #define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47326  float64x2_t __s0 = __p0; \
47327  float64x2_t __s1 = __p1; \
47328  float64x1_t __s2 = __p2; \
47329  float64x2_t __ret; \
47330  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47331  __ret; \
47332 })
47333 #endif
47334 
47335 #ifdef __LITTLE_ENDIAN__
47336 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47337  float32x4_t __s0 = __p0; \
47338  float32x4_t __s1 = __p1; \
47339  float32x2_t __s2 = __p2; \
47340  float32x4_t __ret; \
47341  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47342  __ret; \
47343 })
47344 #else
47345 #define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47346  float32x4_t __s0 = __p0; \
47347  float32x4_t __s1 = __p1; \
47348  float32x2_t __s2 = __p2; \
47349  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47350  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47351  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47352  float32x4_t __ret; \
47353  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
47354  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47355  __ret; \
47356 })
47357 #define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47358  float32x4_t __s0 = __p0; \
47359  float32x4_t __s1 = __p1; \
47360  float32x2_t __s2 = __p2; \
47361  float32x4_t __ret; \
47362  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47363  __ret; \
47364 })
47365 #endif
47366 
47367 #ifdef __LITTLE_ENDIAN__
47368 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47369  float64x1_t __s0 = __p0; \
47370  float64x1_t __s1 = __p1; \
47371  float64x1_t __s2 = __p2; \
47372  float64x1_t __ret; \
47373  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47374  __ret; \
47375 })
47376 #else
47377 #define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47378  float64x1_t __s0 = __p0; \
47379  float64x1_t __s1 = __p1; \
47380  float64x1_t __s2 = __p2; \
47381  float64x1_t __ret; \
47382  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47383  __ret; \
47384 })
47385 #define __noswap_vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47386  float64x1_t __s0 = __p0; \
47387  float64x1_t __s1 = __p1; \
47388  float64x1_t __s2 = __p2; \
47389  float64x1_t __ret; \
47390  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47391  __ret; \
47392 })
47393 #endif
47394 
47395 #ifdef __LITTLE_ENDIAN__
47396 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47397  float32x2_t __s0 = __p0; \
47398  float32x2_t __s1 = __p1; \
47399  float32x2_t __s2 = __p2; \
47400  float32x2_t __ret; \
47401  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47402  __ret; \
47403 })
47404 #else
47405 #define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47406  float32x2_t __s0 = __p0; \
47407  float32x2_t __s1 = __p1; \
47408  float32x2_t __s2 = __p2; \
47409  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47410  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47411  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47412  float32x2_t __ret; \
47413  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
47414  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47415  __ret; \
47416 })
47417 #define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47418  float32x2_t __s0 = __p0; \
47419  float32x2_t __s1 = __p1; \
47420  float32x2_t __s2 = __p2; \
47421  float32x2_t __ret; \
47422  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47423  __ret; \
47424 })
47425 #endif
47426 
47427 #ifdef __LITTLE_ENDIAN__
47428 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47429  float64_t __s0 = __p0; \
47430  float64_t __s1 = __p1; \
47431  float64x2_t __s2 = __p2; \
47432  float64_t __ret; \
47433  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47434  __ret; \
47435 })
47436 #else
47437 #define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47438  float64_t __s0 = __p0; \
47439  float64_t __s1 = __p1; \
47440  float64x2_t __s2 = __p2; \
47441  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47442  float64_t __ret; \
47443  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__rev2, __p3); \
47444  __ret; \
47445 })
47446 #define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47447  float64_t __s0 = __p0; \
47448  float64_t __s1 = __p1; \
47449  float64x2_t __s2 = __p2; \
47450  float64_t __ret; \
47451  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47452  __ret; \
47453 })
47454 #endif
47455 
47456 #ifdef __LITTLE_ENDIAN__
47457 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47458  float32_t __s0 = __p0; \
47459  float32_t __s1 = __p1; \
47460  float32x4_t __s2 = __p2; \
47461  float32_t __ret; \
47462  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47463  __ret; \
47464 })
47465 #else
47466 #define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47467  float32_t __s0 = __p0; \
47468  float32_t __s1 = __p1; \
47469  float32x4_t __s2 = __p2; \
47470  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47471  float32_t __ret; \
47472  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__rev2, __p3); \
47473  __ret; \
47474 })
47475 #define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47476  float32_t __s0 = __p0; \
47477  float32_t __s1 = __p1; \
47478  float32x4_t __s2 = __p2; \
47479  float32_t __ret; \
47480  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47481  __ret; \
47482 })
47483 #endif
47484 
47485 #ifdef __LITTLE_ENDIAN__
47486 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47487  float64x2_t __s0 = __p0; \
47488  float64x2_t __s1 = __p1; \
47489  float64x2_t __s2 = __p2; \
47490  float64x2_t __ret; \
47491  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47492  __ret; \
47493 })
47494 #else
47495 #define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47496  float64x2_t __s0 = __p0; \
47497  float64x2_t __s1 = __p1; \
47498  float64x2_t __s2 = __p2; \
47499  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47500  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47501  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47502  float64x2_t __ret; \
47503  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
47504  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47505  __ret; \
47506 })
47507 #define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47508  float64x2_t __s0 = __p0; \
47509  float64x2_t __s1 = __p1; \
47510  float64x2_t __s2 = __p2; \
47511  float64x2_t __ret; \
47512  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47513  __ret; \
47514 })
47515 #endif
47516 
47517 #ifdef __LITTLE_ENDIAN__
47518 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47519  float32x4_t __s0 = __p0; \
47520  float32x4_t __s1 = __p1; \
47521  float32x4_t __s2 = __p2; \
47522  float32x4_t __ret; \
47523  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47524  __ret; \
47525 })
47526 #else
47527 #define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47528  float32x4_t __s0 = __p0; \
47529  float32x4_t __s1 = __p1; \
47530  float32x4_t __s2 = __p2; \
47531  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47532  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47533  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47534  float32x4_t __ret; \
47535  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
47536  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47537  __ret; \
47538 })
47539 #define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47540  float32x4_t __s0 = __p0; \
47541  float32x4_t __s1 = __p1; \
47542  float32x4_t __s2 = __p2; \
47543  float32x4_t __ret; \
47544  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47545  __ret; \
47546 })
47547 #endif
47548 
47549 #ifdef __LITTLE_ENDIAN__
47550 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47551  float64x1_t __s0 = __p0; \
47552  float64x1_t __s1 = __p1; \
47553  float64x2_t __s2 = __p2; \
47554  float64x1_t __ret; \
47555  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47556  __ret; \
47557 })
47558 #else
47559 #define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47560  float64x1_t __s0 = __p0; \
47561  float64x1_t __s1 = __p1; \
47562  float64x2_t __s2 = __p2; \
47563  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47564  float64x1_t __ret; \
47565  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
47566  __ret; \
47567 })
47568 #define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47569  float64x1_t __s0 = __p0; \
47570  float64x1_t __s1 = __p1; \
47571  float64x2_t __s2 = __p2; \
47572  float64x1_t __ret; \
47573  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47574  __ret; \
47575 })
47576 #endif
47577 
47578 #ifdef __LITTLE_ENDIAN__
47579 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47580  float32x2_t __s0 = __p0; \
47581  float32x2_t __s1 = __p1; \
47582  float32x4_t __s2 = __p2; \
47583  float32x2_t __ret; \
47584  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47585  __ret; \
47586 })
47587 #else
47588 #define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47589  float32x2_t __s0 = __p0; \
47590  float32x2_t __s1 = __p1; \
47591  float32x4_t __s2 = __p2; \
47592  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47593  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47594  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47595  float32x2_t __ret; \
47596  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
47597  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47598  __ret; \
47599 })
47600 #define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47601  float32x2_t __s0 = __p0; \
47602  float32x2_t __s1 = __p1; \
47603  float32x4_t __s2 = __p2; \
47604  float32x2_t __ret; \
47605  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47606  __ret; \
47607 })
47608 #endif
47609 
47610 #ifdef __LITTLE_ENDIAN__
47611 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47612  float64x2_t __ret;
47613  __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
47614  return __ret;
47615 }
47616 #else
47617 __ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47618  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47619  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47620  float64x2_t __ret;
47621  __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
47622  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47623  return __ret;
47624 }
47625 #endif
47626 
47627 #ifdef __LITTLE_ENDIAN__
47628 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47629  float32x4_t __ret;
47630  __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
47631  return __ret;
47632 }
47633 #else
47634 __ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47635  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47636  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47637  float32x4_t __ret;
47638  __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
47639  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47640  return __ret;
47641 }
47642 #endif
47643 
47644 #ifdef __LITTLE_ENDIAN__
47645 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47646  float32x2_t __ret;
47647  __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
47648  return __ret;
47649 }
47650 #else
47651 __ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47652  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47653  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47654  float32x2_t __ret;
47655  __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
47656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47657  return __ret;
47658 }
47659 #endif
47660 
47661 #ifdef __LITTLE_ENDIAN__
47662 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47663  float64x2_t __ret;
47664  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47665  return __ret;
47666 }
47667 #else
47668 __ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47669  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47670  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47671  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47672  float64x2_t __ret;
47673  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
47674  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47675  return __ret;
47676 }
47677 __ai float64x2_t __noswap_vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47678  float64x2_t __ret;
47679  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47680  return __ret;
47681 }
47682 #endif
47683 
47684 #ifdef __LITTLE_ENDIAN__
47685 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47686  float32x4_t __ret;
47687  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
47688  return __ret;
47689 }
47690 #else
47691 __ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47692  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47693  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47694  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
47695  float32x4_t __ret;
47696  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
47697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47698  return __ret;
47699 }
47700 __ai float32x4_t __noswap_vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47701  float32x4_t __ret;
47702  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
47703  return __ret;
47704 }
47705 #endif
47706 
47707 #ifdef __LITTLE_ENDIAN__
47708 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47709  float64x1_t __ret;
47710  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47711  return __ret;
47712 }
47713 #else
47714 __ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47715  float64x1_t __ret;
47716  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47717  return __ret;
47718 }
47719 #endif
47720 
47721 #ifdef __LITTLE_ENDIAN__
47722 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47723  float32x2_t __ret;
47724  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
47725  return __ret;
47726 }
47727 #else
47728 __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47729  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47730  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47731  float32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47732  float32x2_t __ret;
47733  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
47734  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47735  return __ret;
47736 }
47737 __ai float32x2_t __noswap_vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47738  float32x2_t __ret;
47739  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
47740  return __ret;
47741 }
47742 #endif
47743 
47744 #ifdef __LITTLE_ENDIAN__
47745 #define vfmsd_lane_f64(__p0_88, __p1_88, __p2_88, __p3_88) __extension__ ({ \
47746  float64_t __s0_88 = __p0_88; \
47747  float64_t __s1_88 = __p1_88; \
47748  float64x1_t __s2_88 = __p2_88; \
47749  float64_t __ret_88; \
47750  __ret_88 = vfmad_lane_f64(__s0_88, __s1_88, -__s2_88, __p3_88); \
47751  __ret_88; \
47752 })
47753 #else
47754 #define vfmsd_lane_f64(__p0_89, __p1_89, __p2_89, __p3_89) __extension__ ({ \
47755  float64_t __s0_89 = __p0_89; \
47756  float64_t __s1_89 = __p1_89; \
47757  float64x1_t __s2_89 = __p2_89; \
47758  float64_t __ret_89; \
47759  __ret_89 = __noswap_vfmad_lane_f64(__s0_89, __s1_89, -__s2_89, __p3_89); \
47760  __ret_89; \
47761 })
47762 #endif
47763 
47764 #ifdef __LITTLE_ENDIAN__
47765 #define vfmss_lane_f32(__p0_90, __p1_90, __p2_90, __p3_90) __extension__ ({ \
47766  float32_t __s0_90 = __p0_90; \
47767  float32_t __s1_90 = __p1_90; \
47768  float32x2_t __s2_90 = __p2_90; \
47769  float32_t __ret_90; \
47770  __ret_90 = vfmas_lane_f32(__s0_90, __s1_90, -__s2_90, __p3_90); \
47771  __ret_90; \
47772 })
47773 #else
47774 #define vfmss_lane_f32(__p0_91, __p1_91, __p2_91, __p3_91) __extension__ ({ \
47775  float32_t __s0_91 = __p0_91; \
47776  float32_t __s1_91 = __p1_91; \
47777  float32x2_t __s2_91 = __p2_91; \
47778  float32x2_t __rev2_91; __rev2_91 = __builtin_shufflevector(__s2_91, __s2_91, 1, 0); \
47779  float32_t __ret_91; \
47780  __ret_91 = __noswap_vfmas_lane_f32(__s0_91, __s1_91, -__rev2_91, __p3_91); \
47781  __ret_91; \
47782 })
47783 #endif
47784 
47785 #ifdef __LITTLE_ENDIAN__
47786 #define vfmsq_lane_f64(__p0_92, __p1_92, __p2_92, __p3_92) __extension__ ({ \
47787  float64x2_t __s0_92 = __p0_92; \
47788  float64x2_t __s1_92 = __p1_92; \
47789  float64x1_t __s2_92 = __p2_92; \
47790  float64x2_t __ret_92; \
47791  __ret_92 = vfmaq_lane_f64(__s0_92, __s1_92, -__s2_92, __p3_92); \
47792  __ret_92; \
47793 })
47794 #else
47795 #define vfmsq_lane_f64(__p0_93, __p1_93, __p2_93, __p3_93) __extension__ ({ \
47796  float64x2_t __s0_93 = __p0_93; \
47797  float64x2_t __s1_93 = __p1_93; \
47798  float64x1_t __s2_93 = __p2_93; \
47799  float64x2_t __rev0_93; __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
47800  float64x2_t __rev1_93; __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
47801  float64x2_t __ret_93; \
47802  __ret_93 = __noswap_vfmaq_lane_f64(__rev0_93, __rev1_93, -__s2_93, __p3_93); \
47803  __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
47804  __ret_93; \
47805 })
47806 #endif
47807 
47808 #ifdef __LITTLE_ENDIAN__
47809 #define vfmsq_lane_f32(__p0_94, __p1_94, __p2_94, __p3_94) __extension__ ({ \
47810  float32x4_t __s0_94 = __p0_94; \
47811  float32x4_t __s1_94 = __p1_94; \
47812  float32x2_t __s2_94 = __p2_94; \
47813  float32x4_t __ret_94; \
47814  __ret_94 = vfmaq_lane_f32(__s0_94, __s1_94, -__s2_94, __p3_94); \
47815  __ret_94; \
47816 })
47817 #else
47818 #define vfmsq_lane_f32(__p0_95, __p1_95, __p2_95, __p3_95) __extension__ ({ \
47819  float32x4_t __s0_95 = __p0_95; \
47820  float32x4_t __s1_95 = __p1_95; \
47821  float32x2_t __s2_95 = __p2_95; \
47822  float32x4_t __rev0_95; __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
47823  float32x4_t __rev1_95; __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
47824  float32x2_t __rev2_95; __rev2_95 = __builtin_shufflevector(__s2_95, __s2_95, 1, 0); \
47825  float32x4_t __ret_95; \
47826  __ret_95 = __noswap_vfmaq_lane_f32(__rev0_95, __rev1_95, -__rev2_95, __p3_95); \
47827  __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
47828  __ret_95; \
47829 })
47830 #endif
47831 
47832 #ifdef __LITTLE_ENDIAN__
47833 #define vfms_lane_f64(__p0_96, __p1_96, __p2_96, __p3_96) __extension__ ({ \
47834  float64x1_t __s0_96 = __p0_96; \
47835  float64x1_t __s1_96 = __p1_96; \
47836  float64x1_t __s2_96 = __p2_96; \
47837  float64x1_t __ret_96; \
47838  __ret_96 = vfma_lane_f64(__s0_96, __s1_96, -__s2_96, __p3_96); \
47839  __ret_96; \
47840 })
47841 #else
47842 #define vfms_lane_f64(__p0_97, __p1_97, __p2_97, __p3_97) __extension__ ({ \
47843  float64x1_t __s0_97 = __p0_97; \
47844  float64x1_t __s1_97 = __p1_97; \
47845  float64x1_t __s2_97 = __p2_97; \
47846  float64x1_t __ret_97; \
47847  __ret_97 = __noswap_vfma_lane_f64(__s0_97, __s1_97, -__s2_97, __p3_97); \
47848  __ret_97; \
47849 })
47850 #endif
47851 
47852 #ifdef __LITTLE_ENDIAN__
47853 #define vfms_lane_f32(__p0_98, __p1_98, __p2_98, __p3_98) __extension__ ({ \
47854  float32x2_t __s0_98 = __p0_98; \
47855  float32x2_t __s1_98 = __p1_98; \
47856  float32x2_t __s2_98 = __p2_98; \
47857  float32x2_t __ret_98; \
47858  __ret_98 = vfma_lane_f32(__s0_98, __s1_98, -__s2_98, __p3_98); \
47859  __ret_98; \
47860 })
47861 #else
47862 #define vfms_lane_f32(__p0_99, __p1_99, __p2_99, __p3_99) __extension__ ({ \
47863  float32x2_t __s0_99 = __p0_99; \
47864  float32x2_t __s1_99 = __p1_99; \
47865  float32x2_t __s2_99 = __p2_99; \
47866  float32x2_t __rev0_99; __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
47867  float32x2_t __rev1_99; __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
47868  float32x2_t __rev2_99; __rev2_99 = __builtin_shufflevector(__s2_99, __s2_99, 1, 0); \
47869  float32x2_t __ret_99; \
47870  __ret_99 = __noswap_vfma_lane_f32(__rev0_99, __rev1_99, -__rev2_99, __p3_99); \
47871  __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
47872  __ret_99; \
47873 })
47874 #endif
47875 
47876 #ifdef __LITTLE_ENDIAN__
47877 #define vfmsd_laneq_f64(__p0_100, __p1_100, __p2_100, __p3_100) __extension__ ({ \
47878  float64_t __s0_100 = __p0_100; \
47879  float64_t __s1_100 = __p1_100; \
47880  float64x2_t __s2_100 = __p2_100; \
47881  float64_t __ret_100; \
47882  __ret_100 = vfmad_laneq_f64(__s0_100, __s1_100, -__s2_100, __p3_100); \
47883  __ret_100; \
47884 })
47885 #else
47886 #define vfmsd_laneq_f64(__p0_101, __p1_101, __p2_101, __p3_101) __extension__ ({ \
47887  float64_t __s0_101 = __p0_101; \
47888  float64_t __s1_101 = __p1_101; \
47889  float64x2_t __s2_101 = __p2_101; \
47890  float64x2_t __rev2_101; __rev2_101 = __builtin_shufflevector(__s2_101, __s2_101, 1, 0); \
47891  float64_t __ret_101; \
47892  __ret_101 = __noswap_vfmad_laneq_f64(__s0_101, __s1_101, -__rev2_101, __p3_101); \
47893  __ret_101; \
47894 })
47895 #endif
47896 
47897 #ifdef __LITTLE_ENDIAN__
47898 #define vfmss_laneq_f32(__p0_102, __p1_102, __p2_102, __p3_102) __extension__ ({ \
47899  float32_t __s0_102 = __p0_102; \
47900  float32_t __s1_102 = __p1_102; \
47901  float32x4_t __s2_102 = __p2_102; \
47902  float32_t __ret_102; \
47903  __ret_102 = vfmas_laneq_f32(__s0_102, __s1_102, -__s2_102, __p3_102); \
47904  __ret_102; \
47905 })
47906 #else
47907 #define vfmss_laneq_f32(__p0_103, __p1_103, __p2_103, __p3_103) __extension__ ({ \
47908  float32_t __s0_103 = __p0_103; \
47909  float32_t __s1_103 = __p1_103; \
47910  float32x4_t __s2_103 = __p2_103; \
47911  float32x4_t __rev2_103; __rev2_103 = __builtin_shufflevector(__s2_103, __s2_103, 3, 2, 1, 0); \
47912  float32_t __ret_103; \
47913  __ret_103 = __noswap_vfmas_laneq_f32(__s0_103, __s1_103, -__rev2_103, __p3_103); \
47914  __ret_103; \
47915 })
47916 #endif
47917 
47918 #ifdef __LITTLE_ENDIAN__
47919 #define vfmsq_laneq_f64(__p0_104, __p1_104, __p2_104, __p3_104) __extension__ ({ \
47920  float64x2_t __s0_104 = __p0_104; \
47921  float64x2_t __s1_104 = __p1_104; \
47922  float64x2_t __s2_104 = __p2_104; \
47923  float64x2_t __ret_104; \
47924  __ret_104 = vfmaq_laneq_f64(__s0_104, __s1_104, -__s2_104, __p3_104); \
47925  __ret_104; \
47926 })
47927 #else
47928 #define vfmsq_laneq_f64(__p0_105, __p1_105, __p2_105, __p3_105) __extension__ ({ \
47929  float64x2_t __s0_105 = __p0_105; \
47930  float64x2_t __s1_105 = __p1_105; \
47931  float64x2_t __s2_105 = __p2_105; \
47932  float64x2_t __rev0_105; __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 1, 0); \
47933  float64x2_t __rev1_105; __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 1, 0); \
47934  float64x2_t __rev2_105; __rev2_105 = __builtin_shufflevector(__s2_105, __s2_105, 1, 0); \
47935  float64x2_t __ret_105; \
47936  __ret_105 = __noswap_vfmaq_laneq_f64(__rev0_105, __rev1_105, -__rev2_105, __p3_105); \
47937  __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 1, 0); \
47938  __ret_105; \
47939 })
47940 #endif
47941 
47942 #ifdef __LITTLE_ENDIAN__
47943 #define vfmsq_laneq_f32(__p0_106, __p1_106, __p2_106, __p3_106) __extension__ ({ \
47944  float32x4_t __s0_106 = __p0_106; \
47945  float32x4_t __s1_106 = __p1_106; \
47946  float32x4_t __s2_106 = __p2_106; \
47947  float32x4_t __ret_106; \
47948  __ret_106 = vfmaq_laneq_f32(__s0_106, __s1_106, -__s2_106, __p3_106); \
47949  __ret_106; \
47950 })
47951 #else
47952 #define vfmsq_laneq_f32(__p0_107, __p1_107, __p2_107, __p3_107) __extension__ ({ \
47953  float32x4_t __s0_107 = __p0_107; \
47954  float32x4_t __s1_107 = __p1_107; \
47955  float32x4_t __s2_107 = __p2_107; \
47956  float32x4_t __rev0_107; __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 3, 2, 1, 0); \
47957  float32x4_t __rev1_107; __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 3, 2, 1, 0); \
47958  float32x4_t __rev2_107; __rev2_107 = __builtin_shufflevector(__s2_107, __s2_107, 3, 2, 1, 0); \
47959  float32x4_t __ret_107; \
47960  __ret_107 = __noswap_vfmaq_laneq_f32(__rev0_107, __rev1_107, -__rev2_107, __p3_107); \
47961  __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 3, 2, 1, 0); \
47962  __ret_107; \
47963 })
47964 #endif
47965 
47966 #ifdef __LITTLE_ENDIAN__
47967 #define vfms_laneq_f64(__p0_108, __p1_108, __p2_108, __p3_108) __extension__ ({ \
47968  float64x1_t __s0_108 = __p0_108; \
47969  float64x1_t __s1_108 = __p1_108; \
47970  float64x2_t __s2_108 = __p2_108; \
47971  float64x1_t __ret_108; \
47972  __ret_108 = vfma_laneq_f64(__s0_108, __s1_108, -__s2_108, __p3_108); \
47973  __ret_108; \
47974 })
47975 #else
47976 #define vfms_laneq_f64(__p0_109, __p1_109, __p2_109, __p3_109) __extension__ ({ \
47977  float64x1_t __s0_109 = __p0_109; \
47978  float64x1_t __s1_109 = __p1_109; \
47979  float64x2_t __s2_109 = __p2_109; \
47980  float64x2_t __rev2_109; __rev2_109 = __builtin_shufflevector(__s2_109, __s2_109, 1, 0); \
47981  float64x1_t __ret_109; \
47982  __ret_109 = __noswap_vfma_laneq_f64(__s0_109, __s1_109, -__rev2_109, __p3_109); \
47983  __ret_109; \
47984 })
47985 #endif
47986 
47987 #ifdef __LITTLE_ENDIAN__
47988 #define vfms_laneq_f32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
47989  float32x2_t __s0_110 = __p0_110; \
47990  float32x2_t __s1_110 = __p1_110; \
47991  float32x4_t __s2_110 = __p2_110; \
47992  float32x2_t __ret_110; \
47993  __ret_110 = vfma_laneq_f32(__s0_110, __s1_110, -__s2_110, __p3_110); \
47994  __ret_110; \
47995 })
47996 #else
47997 #define vfms_laneq_f32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
47998  float32x2_t __s0_111 = __p0_111; \
47999  float32x2_t __s1_111 = __p1_111; \
48000  float32x4_t __s2_111 = __p2_111; \
48001  float32x2_t __rev0_111; __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
48002  float32x2_t __rev1_111; __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
48003  float32x4_t __rev2_111; __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 3, 2, 1, 0); \
48004  float32x2_t __ret_111; \
48005  __ret_111 = __noswap_vfma_laneq_f32(__rev0_111, __rev1_111, -__rev2_111, __p3_111); \
48006  __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
48007  __ret_111; \
48008 })
48009 #endif
48010 
48011 #ifdef __LITTLE_ENDIAN__
48012 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48013  float64x2_t __ret;
48014  __ret = vfmsq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
48015  return __ret;
48016 }
48017 #else
48018 __ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48019  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48020  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48021  float64x2_t __ret;
48022  __ret = __noswap_vfmsq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
48023  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48024  return __ret;
48025 }
48026 #endif
48027 
48028 #ifdef __LITTLE_ENDIAN__
48029 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48030  float32x4_t __ret;
48031  __ret = vfmsq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
48032  return __ret;
48033 }
48034 #else
48035 __ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48036  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48037  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
48038  float32x4_t __ret;
48039  __ret = __noswap_vfmsq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
48040  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48041  return __ret;
48042 }
48043 #endif
48044 
48045 #ifdef __LITTLE_ENDIAN__
48046 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48047  float32x2_t __ret;
48048  __ret = vfms_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
48049  return __ret;
48050 }
48051 #else
48052 __ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48053  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48054  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48055  float32x2_t __ret;
48056  __ret = __noswap_vfms_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
48057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48058  return __ret;
48059 }
48060 #endif
48061 
48062 #ifdef __LITTLE_ENDIAN__
48063 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48064  poly64x1_t __ret;
48065  __ret = __builtin_shufflevector(__p0, __p0, 1);
48066  return __ret;
48067 }
48068 #else
48069 __ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48070  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48071  poly64x1_t __ret;
48072  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48073  return __ret;
48074 }
48075 __ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
48076  poly64x1_t __ret;
48077  __ret = __builtin_shufflevector(__p0, __p0, 1);
48078  return __ret;
48079 }
48080 #endif
48081 
48082 #ifdef __LITTLE_ENDIAN__
48083 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
48084  float64x1_t __ret;
48085  __ret = __builtin_shufflevector(__p0, __p0, 1);
48086  return __ret;
48087 }
48088 #else
48089 __ai float64x1_t vget_high_f64(float64x2_t __p0) {
48090  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48091  float64x1_t __ret;
48092  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48093  return __ret;
48094 }
48095 #endif
48096 
48097 #ifdef __LITTLE_ENDIAN__
48098 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
48099  poly64x1_t __s0 = __p0; \
48100  poly64_t __ret; \
48101  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48102  __ret; \
48103 })
48104 #else
48105 #define vget_lane_p64(__p0, __p1) __extension__ ({ \
48106  poly64x1_t __s0 = __p0; \
48107  poly64_t __ret; \
48108  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48109  __ret; \
48110 })
48111 #define __noswap_vget_lane_p64(__p0, __p1) __extension__ ({ \
48112  poly64x1_t __s0 = __p0; \
48113  poly64_t __ret; \
48114  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48115  __ret; \
48116 })
48117 #endif
48118 
48119 #ifdef __LITTLE_ENDIAN__
48120 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48121  poly64x2_t __s0 = __p0; \
48122  poly64_t __ret; \
48123  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48124  __ret; \
48125 })
48126 #else
48127 #define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48128  poly64x2_t __s0 = __p0; \
48129  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48130  poly64_t __ret; \
48131  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
48132  __ret; \
48133 })
48134 #define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48135  poly64x2_t __s0 = __p0; \
48136  poly64_t __ret; \
48137  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48138  __ret; \
48139 })
48140 #endif
48141 
48142 #ifdef __LITTLE_ENDIAN__
48143 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48144  float64x2_t __s0 = __p0; \
48145  float64_t __ret; \
48146  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48147  __ret; \
48148 })
48149 #else
48150 #define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48151  float64x2_t __s0 = __p0; \
48152  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48153  float64_t __ret; \
48154  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__rev0, __p1); \
48155  __ret; \
48156 })
48157 #define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48158  float64x2_t __s0 = __p0; \
48159  float64_t __ret; \
48160  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48161  __ret; \
48162 })
48163 #endif
48164 
48165 #ifdef __LITTLE_ENDIAN__
48166 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
48167  float64x1_t __s0 = __p0; \
48168  float64_t __ret; \
48169  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48170  __ret; \
48171 })
48172 #else
48173 #define vget_lane_f64(__p0, __p1) __extension__ ({ \
48174  float64x1_t __s0 = __p0; \
48175  float64_t __ret; \
48176  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48177  __ret; \
48178 })
48179 #define __noswap_vget_lane_f64(__p0, __p1) __extension__ ({ \
48180  float64x1_t __s0 = __p0; \
48181  float64_t __ret; \
48182  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48183  __ret; \
48184 })
48185 #endif
48186 
48187 #ifdef __LITTLE_ENDIAN__
48188 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48189  poly64x1_t __ret;
48190  __ret = __builtin_shufflevector(__p0, __p0, 0);
48191  return __ret;
48192 }
48193 #else
48194 __ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48195  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48196  poly64x1_t __ret;
48197  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48198  return __ret;
48199 }
48200 #endif
48201 
48202 #ifdef __LITTLE_ENDIAN__
48203 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
48204  float64x1_t __ret;
48205  __ret = __builtin_shufflevector(__p0, __p0, 0);
48206  return __ret;
48207 }
48208 #else
48209 __ai float64x1_t vget_low_f64(float64x2_t __p0) {
48210  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48211  float64x1_t __ret;
48212  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48213  return __ret;
48214 }
48215 #endif
48216 
48217 #ifdef __LITTLE_ENDIAN__
48218 #define vld1_p64(__p0) __extension__ ({ \
48219  poly64x1_t __ret; \
48220  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48221  __ret; \
48222 })
48223 #else
48224 #define vld1_p64(__p0) __extension__ ({ \
48225  poly64x1_t __ret; \
48226  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48227  __ret; \
48228 })
48229 #endif
48230 
48231 #ifdef __LITTLE_ENDIAN__
48232 #define vld1q_p64(__p0) __extension__ ({ \
48233  poly64x2_t __ret; \
48234  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48235  __ret; \
48236 })
48237 #else
48238 #define vld1q_p64(__p0) __extension__ ({ \
48239  poly64x2_t __ret; \
48240  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48241  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48242  __ret; \
48243 })
48244 #endif
48245 
48246 #ifdef __LITTLE_ENDIAN__
48247 #define vld1q_f64(__p0) __extension__ ({ \
48248  float64x2_t __ret; \
48249  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48250  __ret; \
48251 })
48252 #else
48253 #define vld1q_f64(__p0) __extension__ ({ \
48254  float64x2_t __ret; \
48255  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48256  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48257  __ret; \
48258 })
48259 #endif
48260 
48261 #ifdef __LITTLE_ENDIAN__
48262 #define vld1_f64(__p0) __extension__ ({ \
48263  float64x1_t __ret; \
48264  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48265  __ret; \
48266 })
48267 #else
48268 #define vld1_f64(__p0) __extension__ ({ \
48269  float64x1_t __ret; \
48270  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48271  __ret; \
48272 })
48273 #endif
48274 
48275 #ifdef __LITTLE_ENDIAN__
48276 #define vld1_dup_p64(__p0) __extension__ ({ \
48277  poly64x1_t __ret; \
48278  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48279  __ret; \
48280 })
48281 #else
48282 #define vld1_dup_p64(__p0) __extension__ ({ \
48283  poly64x1_t __ret; \
48284  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48285  __ret; \
48286 })
48287 #endif
48288 
48289 #ifdef __LITTLE_ENDIAN__
48290 #define vld1q_dup_p64(__p0) __extension__ ({ \
48291  poly64x2_t __ret; \
48292  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48293  __ret; \
48294 })
48295 #else
48296 #define vld1q_dup_p64(__p0) __extension__ ({ \
48297  poly64x2_t __ret; \
48298  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48299  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48300  __ret; \
48301 })
48302 #endif
48303 
48304 #ifdef __LITTLE_ENDIAN__
48305 #define vld1q_dup_f64(__p0) __extension__ ({ \
48306  float64x2_t __ret; \
48307  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48308  __ret; \
48309 })
48310 #else
48311 #define vld1q_dup_f64(__p0) __extension__ ({ \
48312  float64x2_t __ret; \
48313  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48314  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48315  __ret; \
48316 })
48317 #endif
48318 
48319 #ifdef __LITTLE_ENDIAN__
48320 #define vld1_dup_f64(__p0) __extension__ ({ \
48321  float64x1_t __ret; \
48322  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48323  __ret; \
48324 })
48325 #else
48326 #define vld1_dup_f64(__p0) __extension__ ({ \
48327  float64x1_t __ret; \
48328  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48329  __ret; \
48330 })
48331 #endif
48332 
48333 #ifdef __LITTLE_ENDIAN__
48334 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48335  poly64x1_t __s1 = __p1; \
48336  poly64x1_t __ret; \
48337  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48338  __ret; \
48339 })
48340 #else
48341 #define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48342  poly64x1_t __s1 = __p1; \
48343  poly64x1_t __ret; \
48344  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48345  __ret; \
48346 })
48347 #endif
48348 
48349 #ifdef __LITTLE_ENDIAN__
48350 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48351  poly64x2_t __s1 = __p1; \
48352  poly64x2_t __ret; \
48353  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
48354  __ret; \
48355 })
48356 #else
48357 #define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48358  poly64x2_t __s1 = __p1; \
48359  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48360  poly64x2_t __ret; \
48361  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
48362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48363  __ret; \
48364 })
48365 #endif
48366 
48367 #ifdef __LITTLE_ENDIAN__
48368 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48369  float64x2_t __s1 = __p1; \
48370  float64x2_t __ret; \
48371  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
48372  __ret; \
48373 })
48374 #else
48375 #define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48376  float64x2_t __s1 = __p1; \
48377  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48378  float64x2_t __ret; \
48379  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
48380  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48381  __ret; \
48382 })
48383 #endif
48384 
48385 #ifdef __LITTLE_ENDIAN__
48386 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48387  float64x1_t __s1 = __p1; \
48388  float64x1_t __ret; \
48389  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48390  __ret; \
48391 })
48392 #else
48393 #define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48394  float64x1_t __s1 = __p1; \
48395  float64x1_t __ret; \
48396  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48397  __ret; \
48398 })
48399 #endif
48400 
48401 #ifdef __LITTLE_ENDIAN__
48402 #define vld1_p8_x2(__p0) __extension__ ({ \
48403  poly8x8x2_t __ret; \
48404  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48405  __ret; \
48406 })
48407 #else
48408 #define vld1_p8_x2(__p0) __extension__ ({ \
48409  poly8x8x2_t __ret; \
48410  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48411  \
48412  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48413  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48414  __ret; \
48415 })
48416 #endif
48417 
48418 #ifdef __LITTLE_ENDIAN__
48419 #define vld1_p64_x2(__p0) __extension__ ({ \
48420  poly64x1x2_t __ret; \
48421  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48422  __ret; \
48423 })
48424 #else
48425 #define vld1_p64_x2(__p0) __extension__ ({ \
48426  poly64x1x2_t __ret; \
48427  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48428  __ret; \
48429 })
48430 #endif
48431 
48432 #ifdef __LITTLE_ENDIAN__
48433 #define vld1_p16_x2(__p0) __extension__ ({ \
48434  poly16x4x2_t __ret; \
48435  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48436  __ret; \
48437 })
48438 #else
48439 #define vld1_p16_x2(__p0) __extension__ ({ \
48440  poly16x4x2_t __ret; \
48441  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48442  \
48443  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48444  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48445  __ret; \
48446 })
48447 #endif
48448 
48449 #ifdef __LITTLE_ENDIAN__
48450 #define vld1q_p8_x2(__p0) __extension__ ({ \
48451  poly8x16x2_t __ret; \
48452  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48453  __ret; \
48454 })
48455 #else
48456 #define vld1q_p8_x2(__p0) __extension__ ({ \
48457  poly8x16x2_t __ret; \
48458  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48459  \
48460  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48461  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48462  __ret; \
48463 })
48464 #endif
48465 
48466 #ifdef __LITTLE_ENDIAN__
48467 #define vld1q_p64_x2(__p0) __extension__ ({ \
48468  poly64x2x2_t __ret; \
48469  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48470  __ret; \
48471 })
48472 #else
48473 #define vld1q_p64_x2(__p0) __extension__ ({ \
48474  poly64x2x2_t __ret; \
48475  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48476  \
48477  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48478  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48479  __ret; \
48480 })
48481 #endif
48482 
48483 #ifdef __LITTLE_ENDIAN__
48484 #define vld1q_p16_x2(__p0) __extension__ ({ \
48485  poly16x8x2_t __ret; \
48486  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48487  __ret; \
48488 })
48489 #else
48490 #define vld1q_p16_x2(__p0) __extension__ ({ \
48491  poly16x8x2_t __ret; \
48492  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48493  \
48494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48496  __ret; \
48497 })
48498 #endif
48499 
48500 #ifdef __LITTLE_ENDIAN__
48501 #define vld1q_u8_x2(__p0) __extension__ ({ \
48502  uint8x16x2_t __ret; \
48503  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48504  __ret; \
48505 })
48506 #else
48507 #define vld1q_u8_x2(__p0) __extension__ ({ \
48508  uint8x16x2_t __ret; \
48509  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48510  \
48511  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48512  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48513  __ret; \
48514 })
48515 #endif
48516 
48517 #ifdef __LITTLE_ENDIAN__
48518 #define vld1q_u32_x2(__p0) __extension__ ({ \
48519  uint32x4x2_t __ret; \
48520  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48521  __ret; \
48522 })
48523 #else
48524 #define vld1q_u32_x2(__p0) __extension__ ({ \
48525  uint32x4x2_t __ret; \
48526  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48527  \
48528  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48529  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48530  __ret; \
48531 })
48532 #endif
48533 
48534 #ifdef __LITTLE_ENDIAN__
48535 #define vld1q_u64_x2(__p0) __extension__ ({ \
48536  uint64x2x2_t __ret; \
48537  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48538  __ret; \
48539 })
48540 #else
48541 #define vld1q_u64_x2(__p0) __extension__ ({ \
48542  uint64x2x2_t __ret; \
48543  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48544  \
48545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48547  __ret; \
48548 })
48549 #endif
48550 
48551 #ifdef __LITTLE_ENDIAN__
48552 #define vld1q_u16_x2(__p0) __extension__ ({ \
48553  uint16x8x2_t __ret; \
48554  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48555  __ret; \
48556 })
48557 #else
48558 #define vld1q_u16_x2(__p0) __extension__ ({ \
48559  uint16x8x2_t __ret; \
48560  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48561  \
48562  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48563  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48564  __ret; \
48565 })
48566 #endif
48567 
48568 #ifdef __LITTLE_ENDIAN__
48569 #define vld1q_s8_x2(__p0) __extension__ ({ \
48570  int8x16x2_t __ret; \
48571  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48572  __ret; \
48573 })
48574 #else
48575 #define vld1q_s8_x2(__p0) __extension__ ({ \
48576  int8x16x2_t __ret; \
48577  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48578  \
48579  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48580  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48581  __ret; \
48582 })
48583 #endif
48584 
48585 #ifdef __LITTLE_ENDIAN__
48586 #define vld1q_f64_x2(__p0) __extension__ ({ \
48587  float64x2x2_t __ret; \
48588  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48589  __ret; \
48590 })
48591 #else
48592 #define vld1q_f64_x2(__p0) __extension__ ({ \
48593  float64x2x2_t __ret; \
48594  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48595  \
48596  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48597  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48598  __ret; \
48599 })
48600 #endif
48601 
48602 #ifdef __LITTLE_ENDIAN__
48603 #define vld1q_f32_x2(__p0) __extension__ ({ \
48604  float32x4x2_t __ret; \
48605  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48606  __ret; \
48607 })
48608 #else
48609 #define vld1q_f32_x2(__p0) __extension__ ({ \
48610  float32x4x2_t __ret; \
48611  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48612  \
48613  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48614  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48615  __ret; \
48616 })
48617 #endif
48618 
48619 #ifdef __LITTLE_ENDIAN__
48620 #define vld1q_f16_x2(__p0) __extension__ ({ \
48621  float16x8x2_t __ret; \
48622  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48623  __ret; \
48624 })
48625 #else
48626 #define vld1q_f16_x2(__p0) __extension__ ({ \
48627  float16x8x2_t __ret; \
48628  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48629  \
48630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48632  __ret; \
48633 })
48634 #endif
48635 
48636 #ifdef __LITTLE_ENDIAN__
48637 #define vld1q_s32_x2(__p0) __extension__ ({ \
48638  int32x4x2_t __ret; \
48639  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48640  __ret; \
48641 })
48642 #else
48643 #define vld1q_s32_x2(__p0) __extension__ ({ \
48644  int32x4x2_t __ret; \
48645  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48646  \
48647  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48648  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48649  __ret; \
48650 })
48651 #endif
48652 
48653 #ifdef __LITTLE_ENDIAN__
48654 #define vld1q_s64_x2(__p0) __extension__ ({ \
48655  int64x2x2_t __ret; \
48656  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48657  __ret; \
48658 })
48659 #else
48660 #define vld1q_s64_x2(__p0) __extension__ ({ \
48661  int64x2x2_t __ret; \
48662  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48663  \
48664  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48665  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48666  __ret; \
48667 })
48668 #endif
48669 
48670 #ifdef __LITTLE_ENDIAN__
48671 #define vld1q_s16_x2(__p0) __extension__ ({ \
48672  int16x8x2_t __ret; \
48673  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48674  __ret; \
48675 })
48676 #else
48677 #define vld1q_s16_x2(__p0) __extension__ ({ \
48678  int16x8x2_t __ret; \
48679  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48680  \
48681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48683  __ret; \
48684 })
48685 #endif
48686 
48687 #ifdef __LITTLE_ENDIAN__
48688 #define vld1_u8_x2(__p0) __extension__ ({ \
48689  uint8x8x2_t __ret; \
48690  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48691  __ret; \
48692 })
48693 #else
48694 #define vld1_u8_x2(__p0) __extension__ ({ \
48695  uint8x8x2_t __ret; \
48696  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48697  \
48698  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48699  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48700  __ret; \
48701 })
48702 #endif
48703 
48704 #ifdef __LITTLE_ENDIAN__
48705 #define vld1_u32_x2(__p0) __extension__ ({ \
48706  uint32x2x2_t __ret; \
48707  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48708  __ret; \
48709 })
48710 #else
48711 #define vld1_u32_x2(__p0) __extension__ ({ \
48712  uint32x2x2_t __ret; \
48713  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48714  \
48715  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48716  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48717  __ret; \
48718 })
48719 #endif
48720 
48721 #ifdef __LITTLE_ENDIAN__
48722 #define vld1_u64_x2(__p0) __extension__ ({ \
48723  uint64x1x2_t __ret; \
48724  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48725  __ret; \
48726 })
48727 #else
48728 #define vld1_u64_x2(__p0) __extension__ ({ \
48729  uint64x1x2_t __ret; \
48730  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48731  __ret; \
48732 })
48733 #endif
48734 
48735 #ifdef __LITTLE_ENDIAN__
48736 #define vld1_u16_x2(__p0) __extension__ ({ \
48737  uint16x4x2_t __ret; \
48738  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48739  __ret; \
48740 })
48741 #else
48742 #define vld1_u16_x2(__p0) __extension__ ({ \
48743  uint16x4x2_t __ret; \
48744  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48745  \
48746  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48747  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48748  __ret; \
48749 })
48750 #endif
48751 
48752 #ifdef __LITTLE_ENDIAN__
48753 #define vld1_s8_x2(__p0) __extension__ ({ \
48754  int8x8x2_t __ret; \
48755  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48756  __ret; \
48757 })
48758 #else
48759 #define vld1_s8_x2(__p0) __extension__ ({ \
48760  int8x8x2_t __ret; \
48761  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48762  \
48763  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48764  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48765  __ret; \
48766 })
48767 #endif
48768 
48769 #ifdef __LITTLE_ENDIAN__
48770 #define vld1_f64_x2(__p0) __extension__ ({ \
48771  float64x1x2_t __ret; \
48772  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48773  __ret; \
48774 })
48775 #else
48776 #define vld1_f64_x2(__p0) __extension__ ({ \
48777  float64x1x2_t __ret; \
48778  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48779  __ret; \
48780 })
48781 #endif
48782 
48783 #ifdef __LITTLE_ENDIAN__
48784 #define vld1_f32_x2(__p0) __extension__ ({ \
48785  float32x2x2_t __ret; \
48786  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48787  __ret; \
48788 })
48789 #else
48790 #define vld1_f32_x2(__p0) __extension__ ({ \
48791  float32x2x2_t __ret; \
48792  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48793  \
48794  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48795  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48796  __ret; \
48797 })
48798 #endif
48799 
48800 #ifdef __LITTLE_ENDIAN__
48801 #define vld1_f16_x2(__p0) __extension__ ({ \
48802  float16x4x2_t __ret; \
48803  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48804  __ret; \
48805 })
48806 #else
48807 #define vld1_f16_x2(__p0) __extension__ ({ \
48808  float16x4x2_t __ret; \
48809  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48810  \
48811  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48812  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48813  __ret; \
48814 })
48815 #endif
48816 
48817 #ifdef __LITTLE_ENDIAN__
48818 #define vld1_s32_x2(__p0) __extension__ ({ \
48819  int32x2x2_t __ret; \
48820  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48821  __ret; \
48822 })
48823 #else
48824 #define vld1_s32_x2(__p0) __extension__ ({ \
48825  int32x2x2_t __ret; \
48826  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48827  \
48828  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48829  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48830  __ret; \
48831 })
48832 #endif
48833 
48834 #ifdef __LITTLE_ENDIAN__
48835 #define vld1_s64_x2(__p0) __extension__ ({ \
48836  int64x1x2_t __ret; \
48837  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48838  __ret; \
48839 })
48840 #else
48841 #define vld1_s64_x2(__p0) __extension__ ({ \
48842  int64x1x2_t __ret; \
48843  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48844  __ret; \
48845 })
48846 #endif
48847 
48848 #ifdef __LITTLE_ENDIAN__
48849 #define vld1_s16_x2(__p0) __extension__ ({ \
48850  int16x4x2_t __ret; \
48851  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48852  __ret; \
48853 })
48854 #else
48855 #define vld1_s16_x2(__p0) __extension__ ({ \
48856  int16x4x2_t __ret; \
48857  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48858  \
48859  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48860  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48861  __ret; \
48862 })
48863 #endif
48864 
48865 #ifdef __LITTLE_ENDIAN__
48866 #define vld1_p8_x3(__p0) __extension__ ({ \
48867  poly8x8x3_t __ret; \
48868  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48869  __ret; \
48870 })
48871 #else
48872 #define vld1_p8_x3(__p0) __extension__ ({ \
48873  poly8x8x3_t __ret; \
48874  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48875  \
48876  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48877  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48878  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48879  __ret; \
48880 })
48881 #endif
48882 
48883 #ifdef __LITTLE_ENDIAN__
48884 #define vld1_p64_x3(__p0) __extension__ ({ \
48885  poly64x1x3_t __ret; \
48886  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48887  __ret; \
48888 })
48889 #else
48890 #define vld1_p64_x3(__p0) __extension__ ({ \
48891  poly64x1x3_t __ret; \
48892  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48893  __ret; \
48894 })
48895 #endif
48896 
48897 #ifdef __LITTLE_ENDIAN__
48898 #define vld1_p16_x3(__p0) __extension__ ({ \
48899  poly16x4x3_t __ret; \
48900  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48901  __ret; \
48902 })
48903 #else
48904 #define vld1_p16_x3(__p0) __extension__ ({ \
48905  poly16x4x3_t __ret; \
48906  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48907  \
48908  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48909  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48910  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48911  __ret; \
48912 })
48913 #endif
48914 
48915 #ifdef __LITTLE_ENDIAN__
48916 #define vld1q_p8_x3(__p0) __extension__ ({ \
48917  poly8x16x3_t __ret; \
48918  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48919  __ret; \
48920 })
48921 #else
48922 #define vld1q_p8_x3(__p0) __extension__ ({ \
48923  poly8x16x3_t __ret; \
48924  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48925  \
48926  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48927  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48928  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48929  __ret; \
48930 })
48931 #endif
48932 
48933 #ifdef __LITTLE_ENDIAN__
48934 #define vld1q_p64_x3(__p0) __extension__ ({ \
48935  poly64x2x3_t __ret; \
48936  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48937  __ret; \
48938 })
48939 #else
48940 #define vld1q_p64_x3(__p0) __extension__ ({ \
48941  poly64x2x3_t __ret; \
48942  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48943  \
48944  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48945  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48946  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48947  __ret; \
48948 })
48949 #endif
48950 
48951 #ifdef __LITTLE_ENDIAN__
48952 #define vld1q_p16_x3(__p0) __extension__ ({ \
48953  poly16x8x3_t __ret; \
48954  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48955  __ret; \
48956 })
48957 #else
48958 #define vld1q_p16_x3(__p0) __extension__ ({ \
48959  poly16x8x3_t __ret; \
48960  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48961  \
48962  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48963  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48964  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48965  __ret; \
48966 })
48967 #endif
48968 
48969 #ifdef __LITTLE_ENDIAN__
48970 #define vld1q_u8_x3(__p0) __extension__ ({ \
48971  uint8x16x3_t __ret; \
48972  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48973  __ret; \
48974 })
48975 #else
48976 #define vld1q_u8_x3(__p0) __extension__ ({ \
48977  uint8x16x3_t __ret; \
48978  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48979  \
48980  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48981  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48982  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48983  __ret; \
48984 })
48985 #endif
48986 
48987 #ifdef __LITTLE_ENDIAN__
48988 #define vld1q_u32_x3(__p0) __extension__ ({ \
48989  uint32x4x3_t __ret; \
48990  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48991  __ret; \
48992 })
48993 #else
48994 #define vld1q_u32_x3(__p0) __extension__ ({ \
48995  uint32x4x3_t __ret; \
48996  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48997  \
48998  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48999  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49000  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49001  __ret; \
49002 })
49003 #endif
49004 
49005 #ifdef __LITTLE_ENDIAN__
49006 #define vld1q_u64_x3(__p0) __extension__ ({ \
49007  uint64x2x3_t __ret; \
49008  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49009  __ret; \
49010 })
49011 #else
49012 #define vld1q_u64_x3(__p0) __extension__ ({ \
49013  uint64x2x3_t __ret; \
49014  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49015  \
49016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49018  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49019  __ret; \
49020 })
49021 #endif
49022 
49023 #ifdef __LITTLE_ENDIAN__
49024 #define vld1q_u16_x3(__p0) __extension__ ({ \
49025  uint16x8x3_t __ret; \
49026  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49027  __ret; \
49028 })
49029 #else
49030 #define vld1q_u16_x3(__p0) __extension__ ({ \
49031  uint16x8x3_t __ret; \
49032  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49033  \
49034  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49035  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49036  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49037  __ret; \
49038 })
49039 #endif
49040 
49041 #ifdef __LITTLE_ENDIAN__
49042 #define vld1q_s8_x3(__p0) __extension__ ({ \
49043  int8x16x3_t __ret; \
49044  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49045  __ret; \
49046 })
49047 #else
49048 #define vld1q_s8_x3(__p0) __extension__ ({ \
49049  int8x16x3_t __ret; \
49050  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49051  \
49052  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49053  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49054  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49055  __ret; \
49056 })
49057 #endif
49058 
49059 #ifdef __LITTLE_ENDIAN__
49060 #define vld1q_f64_x3(__p0) __extension__ ({ \
49061  float64x2x3_t __ret; \
49062  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49063  __ret; \
49064 })
49065 #else
49066 #define vld1q_f64_x3(__p0) __extension__ ({ \
49067  float64x2x3_t __ret; \
49068  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49069  \
49070  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49071  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49072  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49073  __ret; \
49074 })
49075 #endif
49076 
49077 #ifdef __LITTLE_ENDIAN__
49078 #define vld1q_f32_x3(__p0) __extension__ ({ \
49079  float32x4x3_t __ret; \
49080  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49081  __ret; \
49082 })
49083 #else
49084 #define vld1q_f32_x3(__p0) __extension__ ({ \
49085  float32x4x3_t __ret; \
49086  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49087  \
49088  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49089  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49090  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49091  __ret; \
49092 })
49093 #endif
49094 
49095 #ifdef __LITTLE_ENDIAN__
49096 #define vld1q_f16_x3(__p0) __extension__ ({ \
49097  float16x8x3_t __ret; \
49098  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49099  __ret; \
49100 })
49101 #else
49102 #define vld1q_f16_x3(__p0) __extension__ ({ \
49103  float16x8x3_t __ret; \
49104  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49105  \
49106  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49107  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49108  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49109  __ret; \
49110 })
49111 #endif
49112 
49113 #ifdef __LITTLE_ENDIAN__
49114 #define vld1q_s32_x3(__p0) __extension__ ({ \
49115  int32x4x3_t __ret; \
49116  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49117  __ret; \
49118 })
49119 #else
49120 #define vld1q_s32_x3(__p0) __extension__ ({ \
49121  int32x4x3_t __ret; \
49122  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49123  \
49124  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49125  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49126  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49127  __ret; \
49128 })
49129 #endif
49130 
49131 #ifdef __LITTLE_ENDIAN__
49132 #define vld1q_s64_x3(__p0) __extension__ ({ \
49133  int64x2x3_t __ret; \
49134  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49135  __ret; \
49136 })
49137 #else
49138 #define vld1q_s64_x3(__p0) __extension__ ({ \
49139  int64x2x3_t __ret; \
49140  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49141  \
49142  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49143  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49144  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49145  __ret; \
49146 })
49147 #endif
49148 
49149 #ifdef __LITTLE_ENDIAN__
49150 #define vld1q_s16_x3(__p0) __extension__ ({ \
49151  int16x8x3_t __ret; \
49152  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49153  __ret; \
49154 })
49155 #else
49156 #define vld1q_s16_x3(__p0) __extension__ ({ \
49157  int16x8x3_t __ret; \
49158  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49159  \
49160  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49161  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49162  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49163  __ret; \
49164 })
49165 #endif
49166 
49167 #ifdef __LITTLE_ENDIAN__
49168 #define vld1_u8_x3(__p0) __extension__ ({ \
49169  uint8x8x3_t __ret; \
49170  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49171  __ret; \
49172 })
49173 #else
49174 #define vld1_u8_x3(__p0) __extension__ ({ \
49175  uint8x8x3_t __ret; \
49176  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49177  \
49178  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49179  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49180  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49181  __ret; \
49182 })
49183 #endif
49184 
49185 #ifdef __LITTLE_ENDIAN__
49186 #define vld1_u32_x3(__p0) __extension__ ({ \
49187  uint32x2x3_t __ret; \
49188  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49189  __ret; \
49190 })
49191 #else
49192 #define vld1_u32_x3(__p0) __extension__ ({ \
49193  uint32x2x3_t __ret; \
49194  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49195  \
49196  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49197  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49198  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49199  __ret; \
49200 })
49201 #endif
49202 
49203 #ifdef __LITTLE_ENDIAN__
49204 #define vld1_u64_x3(__p0) __extension__ ({ \
49205  uint64x1x3_t __ret; \
49206  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49207  __ret; \
49208 })
49209 #else
49210 #define vld1_u64_x3(__p0) __extension__ ({ \
49211  uint64x1x3_t __ret; \
49212  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49213  __ret; \
49214 })
49215 #endif
49216 
49217 #ifdef __LITTLE_ENDIAN__
49218 #define vld1_u16_x3(__p0) __extension__ ({ \
49219  uint16x4x3_t __ret; \
49220  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49221  __ret; \
49222 })
49223 #else
49224 #define vld1_u16_x3(__p0) __extension__ ({ \
49225  uint16x4x3_t __ret; \
49226  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49227  \
49228  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49229  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49230  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49231  __ret; \
49232 })
49233 #endif
49234 
49235 #ifdef __LITTLE_ENDIAN__
49236 #define vld1_s8_x3(__p0) __extension__ ({ \
49237  int8x8x3_t __ret; \
49238  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49239  __ret; \
49240 })
49241 #else
49242 #define vld1_s8_x3(__p0) __extension__ ({ \
49243  int8x8x3_t __ret; \
49244  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49245  \
49246  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49247  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49248  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49249  __ret; \
49250 })
49251 #endif
49252 
49253 #ifdef __LITTLE_ENDIAN__
49254 #define vld1_f64_x3(__p0) __extension__ ({ \
49255  float64x1x3_t __ret; \
49256  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49257  __ret; \
49258 })
49259 #else
49260 #define vld1_f64_x3(__p0) __extension__ ({ \
49261  float64x1x3_t __ret; \
49262  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49263  __ret; \
49264 })
49265 #endif
49266 
49267 #ifdef __LITTLE_ENDIAN__
49268 #define vld1_f32_x3(__p0) __extension__ ({ \
49269  float32x2x3_t __ret; \
49270  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49271  __ret; \
49272 })
49273 #else
49274 #define vld1_f32_x3(__p0) __extension__ ({ \
49275  float32x2x3_t __ret; \
49276  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49277  \
49278  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49279  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49280  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49281  __ret; \
49282 })
49283 #endif
49284 
49285 #ifdef __LITTLE_ENDIAN__
49286 #define vld1_f16_x3(__p0) __extension__ ({ \
49287  float16x4x3_t __ret; \
49288  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49289  __ret; \
49290 })
49291 #else
49292 #define vld1_f16_x3(__p0) __extension__ ({ \
49293  float16x4x3_t __ret; \
49294  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49295  \
49296  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49297  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49298  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49299  __ret; \
49300 })
49301 #endif
49302 
49303 #ifdef __LITTLE_ENDIAN__
49304 #define vld1_s32_x3(__p0) __extension__ ({ \
49305  int32x2x3_t __ret; \
49306  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49307  __ret; \
49308 })
49309 #else
49310 #define vld1_s32_x3(__p0) __extension__ ({ \
49311  int32x2x3_t __ret; \
49312  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49313  \
49314  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49315  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49316  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49317  __ret; \
49318 })
49319 #endif
49320 
49321 #ifdef __LITTLE_ENDIAN__
49322 #define vld1_s64_x3(__p0) __extension__ ({ \
49323  int64x1x3_t __ret; \
49324  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49325  __ret; \
49326 })
49327 #else
49328 #define vld1_s64_x3(__p0) __extension__ ({ \
49329  int64x1x3_t __ret; \
49330  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49331  __ret; \
49332 })
49333 #endif
49334 
49335 #ifdef __LITTLE_ENDIAN__
49336 #define vld1_s16_x3(__p0) __extension__ ({ \
49337  int16x4x3_t __ret; \
49338  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49339  __ret; \
49340 })
49341 #else
49342 #define vld1_s16_x3(__p0) __extension__ ({ \
49343  int16x4x3_t __ret; \
49344  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49345  \
49346  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49347  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49348  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49349  __ret; \
49350 })
49351 #endif
49352 
49353 #ifdef __LITTLE_ENDIAN__
49354 #define vld1_p8_x4(__p0) __extension__ ({ \
49355  poly8x8x4_t __ret; \
49356  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49357  __ret; \
49358 })
49359 #else
49360 #define vld1_p8_x4(__p0) __extension__ ({ \
49361  poly8x8x4_t __ret; \
49362  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49363  \
49364  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49365  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49366  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49367  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49368  __ret; \
49369 })
49370 #endif
49371 
49372 #ifdef __LITTLE_ENDIAN__
49373 #define vld1_p64_x4(__p0) __extension__ ({ \
49374  poly64x1x4_t __ret; \
49375  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49376  __ret; \
49377 })
49378 #else
49379 #define vld1_p64_x4(__p0) __extension__ ({ \
49380  poly64x1x4_t __ret; \
49381  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49382  __ret; \
49383 })
49384 #endif
49385 
49386 #ifdef __LITTLE_ENDIAN__
49387 #define vld1_p16_x4(__p0) __extension__ ({ \
49388  poly16x4x4_t __ret; \
49389  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49390  __ret; \
49391 })
49392 #else
49393 #define vld1_p16_x4(__p0) __extension__ ({ \
49394  poly16x4x4_t __ret; \
49395  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49396  \
49397  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49398  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49399  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49400  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49401  __ret; \
49402 })
49403 #endif
49404 
49405 #ifdef __LITTLE_ENDIAN__
49406 #define vld1q_p8_x4(__p0) __extension__ ({ \
49407  poly8x16x4_t __ret; \
49408  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49409  __ret; \
49410 })
49411 #else
49412 #define vld1q_p8_x4(__p0) __extension__ ({ \
49413  poly8x16x4_t __ret; \
49414  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49415  \
49416  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49417  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49418  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49419  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49420  __ret; \
49421 })
49422 #endif
49423 
49424 #ifdef __LITTLE_ENDIAN__
49425 #define vld1q_p64_x4(__p0) __extension__ ({ \
49426  poly64x2x4_t __ret; \
49427  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49428  __ret; \
49429 })
49430 #else
49431 #define vld1q_p64_x4(__p0) __extension__ ({ \
49432  poly64x2x4_t __ret; \
49433  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49434  \
49435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49437  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49438  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49439  __ret; \
49440 })
49441 #endif
49442 
49443 #ifdef __LITTLE_ENDIAN__
49444 #define vld1q_p16_x4(__p0) __extension__ ({ \
49445  poly16x8x4_t __ret; \
49446  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49447  __ret; \
49448 })
49449 #else
49450 #define vld1q_p16_x4(__p0) __extension__ ({ \
49451  poly16x8x4_t __ret; \
49452  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49453  \
49454  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49455  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49456  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49457  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49458  __ret; \
49459 })
49460 #endif
49461 
49462 #ifdef __LITTLE_ENDIAN__
49463 #define vld1q_u8_x4(__p0) __extension__ ({ \
49464  uint8x16x4_t __ret; \
49465  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49466  __ret; \
49467 })
49468 #else
49469 #define vld1q_u8_x4(__p0) __extension__ ({ \
49470  uint8x16x4_t __ret; \
49471  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49472  \
49473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49476  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49477  __ret; \
49478 })
49479 #endif
49480 
49481 #ifdef __LITTLE_ENDIAN__
49482 #define vld1q_u32_x4(__p0) __extension__ ({ \
49483  uint32x4x4_t __ret; \
49484  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49485  __ret; \
49486 })
49487 #else
49488 #define vld1q_u32_x4(__p0) __extension__ ({ \
49489  uint32x4x4_t __ret; \
49490  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49491  \
49492  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49493  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49494  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49495  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49496  __ret; \
49497 })
49498 #endif
49499 
49500 #ifdef __LITTLE_ENDIAN__
49501 #define vld1q_u64_x4(__p0) __extension__ ({ \
49502  uint64x2x4_t __ret; \
49503  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49504  __ret; \
49505 })
49506 #else
49507 #define vld1q_u64_x4(__p0) __extension__ ({ \
49508  uint64x2x4_t __ret; \
49509  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49510  \
49511  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49512  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49513  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49514  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49515  __ret; \
49516 })
49517 #endif
49518 
49519 #ifdef __LITTLE_ENDIAN__
49520 #define vld1q_u16_x4(__p0) __extension__ ({ \
49521  uint16x8x4_t __ret; \
49522  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49523  __ret; \
49524 })
49525 #else
49526 #define vld1q_u16_x4(__p0) __extension__ ({ \
49527  uint16x8x4_t __ret; \
49528  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49529  \
49530  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49531  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49532  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49533  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49534  __ret; \
49535 })
49536 #endif
49537 
49538 #ifdef __LITTLE_ENDIAN__
49539 #define vld1q_s8_x4(__p0) __extension__ ({ \
49540  int8x16x4_t __ret; \
49541  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49542  __ret; \
49543 })
49544 #else
49545 #define vld1q_s8_x4(__p0) __extension__ ({ \
49546  int8x16x4_t __ret; \
49547  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49548  \
49549  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49550  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49551  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49552  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49553  __ret; \
49554 })
49555 #endif
49556 
49557 #ifdef __LITTLE_ENDIAN__
49558 #define vld1q_f64_x4(__p0) __extension__ ({ \
49559  float64x2x4_t __ret; \
49560  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49561  __ret; \
49562 })
49563 #else
49564 #define vld1q_f64_x4(__p0) __extension__ ({ \
49565  float64x2x4_t __ret; \
49566  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49567  \
49568  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49569  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49570  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49571  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49572  __ret; \
49573 })
49574 #endif
49575 
49576 #ifdef __LITTLE_ENDIAN__
49577 #define vld1q_f32_x4(__p0) __extension__ ({ \
49578  float32x4x4_t __ret; \
49579  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49580  __ret; \
49581 })
49582 #else
49583 #define vld1q_f32_x4(__p0) __extension__ ({ \
49584  float32x4x4_t __ret; \
49585  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49586  \
49587  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49588  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49589  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49590  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49591  __ret; \
49592 })
49593 #endif
49594 
49595 #ifdef __LITTLE_ENDIAN__
49596 #define vld1q_f16_x4(__p0) __extension__ ({ \
49597  float16x8x4_t __ret; \
49598  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49599  __ret; \
49600 })
49601 #else
49602 #define vld1q_f16_x4(__p0) __extension__ ({ \
49603  float16x8x4_t __ret; \
49604  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49605  \
49606  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49607  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49608  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49609  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49610  __ret; \
49611 })
49612 #endif
49613 
49614 #ifdef __LITTLE_ENDIAN__
49615 #define vld1q_s32_x4(__p0) __extension__ ({ \
49616  int32x4x4_t __ret; \
49617  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49618  __ret; \
49619 })
49620 #else
49621 #define vld1q_s32_x4(__p0) __extension__ ({ \
49622  int32x4x4_t __ret; \
49623  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49624  \
49625  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49626  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49627  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49628  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49629  __ret; \
49630 })
49631 #endif
49632 
49633 #ifdef __LITTLE_ENDIAN__
49634 #define vld1q_s64_x4(__p0) __extension__ ({ \
49635  int64x2x4_t __ret; \
49636  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49637  __ret; \
49638 })
49639 #else
49640 #define vld1q_s64_x4(__p0) __extension__ ({ \
49641  int64x2x4_t __ret; \
49642  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49643  \
49644  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49645  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49646  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49647  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49648  __ret; \
49649 })
49650 #endif
49651 
49652 #ifdef __LITTLE_ENDIAN__
49653 #define vld1q_s16_x4(__p0) __extension__ ({ \
49654  int16x8x4_t __ret; \
49655  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49656  __ret; \
49657 })
49658 #else
49659 #define vld1q_s16_x4(__p0) __extension__ ({ \
49660  int16x8x4_t __ret; \
49661  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49662  \
49663  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49664  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49665  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49666  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49667  __ret; \
49668 })
49669 #endif
49670 
49671 #ifdef __LITTLE_ENDIAN__
49672 #define vld1_u8_x4(__p0) __extension__ ({ \
49673  uint8x8x4_t __ret; \
49674  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49675  __ret; \
49676 })
49677 #else
49678 #define vld1_u8_x4(__p0) __extension__ ({ \
49679  uint8x8x4_t __ret; \
49680  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49681  \
49682  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49683  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49684  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49685  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49686  __ret; \
49687 })
49688 #endif
49689 
49690 #ifdef __LITTLE_ENDIAN__
49691 #define vld1_u32_x4(__p0) __extension__ ({ \
49692  uint32x2x4_t __ret; \
49693  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49694  __ret; \
49695 })
49696 #else
49697 #define vld1_u32_x4(__p0) __extension__ ({ \
49698  uint32x2x4_t __ret; \
49699  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49700  \
49701  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49702  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49703  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49704  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49705  __ret; \
49706 })
49707 #endif
49708 
49709 #ifdef __LITTLE_ENDIAN__
49710 #define vld1_u64_x4(__p0) __extension__ ({ \
49711  uint64x1x4_t __ret; \
49712  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49713  __ret; \
49714 })
49715 #else
49716 #define vld1_u64_x4(__p0) __extension__ ({ \
49717  uint64x1x4_t __ret; \
49718  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49719  __ret; \
49720 })
49721 #endif
49722 
49723 #ifdef __LITTLE_ENDIAN__
49724 #define vld1_u16_x4(__p0) __extension__ ({ \
49725  uint16x4x4_t __ret; \
49726  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49727  __ret; \
49728 })
49729 #else
49730 #define vld1_u16_x4(__p0) __extension__ ({ \
49731  uint16x4x4_t __ret; \
49732  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49733  \
49734  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49735  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49736  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49737  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49738  __ret; \
49739 })
49740 #endif
49741 
49742 #ifdef __LITTLE_ENDIAN__
49743 #define vld1_s8_x4(__p0) __extension__ ({ \
49744  int8x8x4_t __ret; \
49745  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49746  __ret; \
49747 })
49748 #else
49749 #define vld1_s8_x4(__p0) __extension__ ({ \
49750  int8x8x4_t __ret; \
49751  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49752  \
49753  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49754  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49755  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49756  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49757  __ret; \
49758 })
49759 #endif
49760 
49761 #ifdef __LITTLE_ENDIAN__
49762 #define vld1_f64_x4(__p0) __extension__ ({ \
49763  float64x1x4_t __ret; \
49764  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49765  __ret; \
49766 })
49767 #else
49768 #define vld1_f64_x4(__p0) __extension__ ({ \
49769  float64x1x4_t __ret; \
49770  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49771  __ret; \
49772 })
49773 #endif
49774 
49775 #ifdef __LITTLE_ENDIAN__
49776 #define vld1_f32_x4(__p0) __extension__ ({ \
49777  float32x2x4_t __ret; \
49778  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49779  __ret; \
49780 })
49781 #else
49782 #define vld1_f32_x4(__p0) __extension__ ({ \
49783  float32x2x4_t __ret; \
49784  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49785  \
49786  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49787  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49788  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49789  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49790  __ret; \
49791 })
49792 #endif
49793 
49794 #ifdef __LITTLE_ENDIAN__
49795 #define vld1_f16_x4(__p0) __extension__ ({ \
49796  float16x4x4_t __ret; \
49797  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49798  __ret; \
49799 })
49800 #else
49801 #define vld1_f16_x4(__p0) __extension__ ({ \
49802  float16x4x4_t __ret; \
49803  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49804  \
49805  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49806  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49807  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49808  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49809  __ret; \
49810 })
49811 #endif
49812 
49813 #ifdef __LITTLE_ENDIAN__
49814 #define vld1_s32_x4(__p0) __extension__ ({ \
49815  int32x2x4_t __ret; \
49816  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49817  __ret; \
49818 })
49819 #else
49820 #define vld1_s32_x4(__p0) __extension__ ({ \
49821  int32x2x4_t __ret; \
49822  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49823  \
49824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49826  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49827  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49828  __ret; \
49829 })
49830 #endif
49831 
49832 #ifdef __LITTLE_ENDIAN__
49833 #define vld1_s64_x4(__p0) __extension__ ({ \
49834  int64x1x4_t __ret; \
49835  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49836  __ret; \
49837 })
49838 #else
49839 #define vld1_s64_x4(__p0) __extension__ ({ \
49840  int64x1x4_t __ret; \
49841  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49842  __ret; \
49843 })
49844 #endif
49845 
49846 #ifdef __LITTLE_ENDIAN__
49847 #define vld1_s16_x4(__p0) __extension__ ({ \
49848  int16x4x4_t __ret; \
49849  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49850  __ret; \
49851 })
49852 #else
49853 #define vld1_s16_x4(__p0) __extension__ ({ \
49854  int16x4x4_t __ret; \
49855  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49856  \
49857  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49858  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49859  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49860  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49861  __ret; \
49862 })
49863 #endif
49864 
49865 #ifdef __LITTLE_ENDIAN__
49866 #define vld2_p64(__p0) __extension__ ({ \
49867  poly64x1x2_t __ret; \
49868  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49869  __ret; \
49870 })
49871 #else
49872 #define vld2_p64(__p0) __extension__ ({ \
49873  poly64x1x2_t __ret; \
49874  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49875  __ret; \
49876 })
49877 #endif
49878 
49879 #ifdef __LITTLE_ENDIAN__
49880 #define vld2q_p64(__p0) __extension__ ({ \
49881  poly64x2x2_t __ret; \
49882  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49883  __ret; \
49884 })
49885 #else
49886 #define vld2q_p64(__p0) __extension__ ({ \
49887  poly64x2x2_t __ret; \
49888  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49889  \
49890  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49891  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49892  __ret; \
49893 })
49894 #endif
49895 
49896 #ifdef __LITTLE_ENDIAN__
49897 #define vld2q_u64(__p0) __extension__ ({ \
49898  uint64x2x2_t __ret; \
49899  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49900  __ret; \
49901 })
49902 #else
49903 #define vld2q_u64(__p0) __extension__ ({ \
49904  uint64x2x2_t __ret; \
49905  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49906  \
49907  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49908  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49909  __ret; \
49910 })
49911 #endif
49912 
49913 #ifdef __LITTLE_ENDIAN__
49914 #define vld2q_f64(__p0) __extension__ ({ \
49915  float64x2x2_t __ret; \
49916  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49917  __ret; \
49918 })
49919 #else
49920 #define vld2q_f64(__p0) __extension__ ({ \
49921  float64x2x2_t __ret; \
49922  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49923  \
49924  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49925  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49926  __ret; \
49927 })
49928 #endif
49929 
49930 #ifdef __LITTLE_ENDIAN__
49931 #define vld2q_s64(__p0) __extension__ ({ \
49932  int64x2x2_t __ret; \
49933  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49934  __ret; \
49935 })
49936 #else
49937 #define vld2q_s64(__p0) __extension__ ({ \
49938  int64x2x2_t __ret; \
49939  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49940  \
49941  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49942  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49943  __ret; \
49944 })
49945 #endif
49946 
49947 #ifdef __LITTLE_ENDIAN__
49948 #define vld2_f64(__p0) __extension__ ({ \
49949  float64x1x2_t __ret; \
49950  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49951  __ret; \
49952 })
49953 #else
49954 #define vld2_f64(__p0) __extension__ ({ \
49955  float64x1x2_t __ret; \
49956  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49957  __ret; \
49958 })
49959 #endif
49960 
49961 #ifdef __LITTLE_ENDIAN__
49962 #define vld2_dup_p64(__p0) __extension__ ({ \
49963  poly64x1x2_t __ret; \
49964  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49965  __ret; \
49966 })
49967 #else
49968 #define vld2_dup_p64(__p0) __extension__ ({ \
49969  poly64x1x2_t __ret; \
49970  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49971  __ret; \
49972 })
49973 #endif
49974 
49975 #ifdef __LITTLE_ENDIAN__
49976 #define vld2q_dup_p8(__p0) __extension__ ({ \
49977  poly8x16x2_t __ret; \
49978  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49979  __ret; \
49980 })
49981 #else
49982 #define vld2q_dup_p8(__p0) __extension__ ({ \
49983  poly8x16x2_t __ret; \
49984  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49985  \
49986  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49987  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49988  __ret; \
49989 })
49990 #endif
49991 
49992 #ifdef __LITTLE_ENDIAN__
49993 #define vld2q_dup_p64(__p0) __extension__ ({ \
49994  poly64x2x2_t __ret; \
49995  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
49996  __ret; \
49997 })
49998 #else
49999 #define vld2q_dup_p64(__p0) __extension__ ({ \
50000  poly64x2x2_t __ret; \
50001  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
50002  \
50003  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50004  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50005  __ret; \
50006 })
50007 #endif
50008 
50009 #ifdef __LITTLE_ENDIAN__
50010 #define vld2q_dup_p16(__p0) __extension__ ({ \
50011  poly16x8x2_t __ret; \
50012  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50013  __ret; \
50014 })
50015 #else
50016 #define vld2q_dup_p16(__p0) __extension__ ({ \
50017  poly16x8x2_t __ret; \
50018  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50019  \
50020  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50021  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50022  __ret; \
50023 })
50024 #endif
50025 
50026 #ifdef __LITTLE_ENDIAN__
50027 #define vld2q_dup_u8(__p0) __extension__ ({ \
50028  uint8x16x2_t __ret; \
50029  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50030  __ret; \
50031 })
50032 #else
50033 #define vld2q_dup_u8(__p0) __extension__ ({ \
50034  uint8x16x2_t __ret; \
50035  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50036  \
50037  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50038  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50039  __ret; \
50040 })
50041 #endif
50042 
50043 #ifdef __LITTLE_ENDIAN__
50044 #define vld2q_dup_u32(__p0) __extension__ ({ \
50045  uint32x4x2_t __ret; \
50046  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50047  __ret; \
50048 })
50049 #else
50050 #define vld2q_dup_u32(__p0) __extension__ ({ \
50051  uint32x4x2_t __ret; \
50052  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50053  \
50054  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50055  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50056  __ret; \
50057 })
50058 #endif
50059 
50060 #ifdef __LITTLE_ENDIAN__
50061 #define vld2q_dup_u64(__p0) __extension__ ({ \
50062  uint64x2x2_t __ret; \
50063  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50064  __ret; \
50065 })
50066 #else
50067 #define vld2q_dup_u64(__p0) __extension__ ({ \
50068  uint64x2x2_t __ret; \
50069  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50070  \
50071  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50072  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50073  __ret; \
50074 })
50075 #endif
50076 
50077 #ifdef __LITTLE_ENDIAN__
50078 #define vld2q_dup_u16(__p0) __extension__ ({ \
50079  uint16x8x2_t __ret; \
50080  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50081  __ret; \
50082 })
50083 #else
50084 #define vld2q_dup_u16(__p0) __extension__ ({ \
50085  uint16x8x2_t __ret; \
50086  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50087  \
50088  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50089  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50090  __ret; \
50091 })
50092 #endif
50093 
50094 #ifdef __LITTLE_ENDIAN__
50095 #define vld2q_dup_s8(__p0) __extension__ ({ \
50096  int8x16x2_t __ret; \
50097  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50098  __ret; \
50099 })
50100 #else
50101 #define vld2q_dup_s8(__p0) __extension__ ({ \
50102  int8x16x2_t __ret; \
50103  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50104  \
50105  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50106  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50107  __ret; \
50108 })
50109 #endif
50110 
50111 #ifdef __LITTLE_ENDIAN__
50112 #define vld2q_dup_f64(__p0) __extension__ ({ \
50113  float64x2x2_t __ret; \
50114  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50115  __ret; \
50116 })
50117 #else
50118 #define vld2q_dup_f64(__p0) __extension__ ({ \
50119  float64x2x2_t __ret; \
50120  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50121  \
50122  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50123  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50124  __ret; \
50125 })
50126 #endif
50127 
50128 #ifdef __LITTLE_ENDIAN__
50129 #define vld2q_dup_f32(__p0) __extension__ ({ \
50130  float32x4x2_t __ret; \
50131  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50132  __ret; \
50133 })
50134 #else
50135 #define vld2q_dup_f32(__p0) __extension__ ({ \
50136  float32x4x2_t __ret; \
50137  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50138  \
50139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50141  __ret; \
50142 })
50143 #endif
50144 
50145 #ifdef __LITTLE_ENDIAN__
50146 #define vld2q_dup_f16(__p0) __extension__ ({ \
50147  float16x8x2_t __ret; \
50148  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50149  __ret; \
50150 })
50151 #else
50152 #define vld2q_dup_f16(__p0) __extension__ ({ \
50153  float16x8x2_t __ret; \
50154  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50155  \
50156  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50157  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50158  __ret; \
50159 })
50160 #endif
50161 
50162 #ifdef __LITTLE_ENDIAN__
50163 #define vld2q_dup_s32(__p0) __extension__ ({ \
50164  int32x4x2_t __ret; \
50165  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50166  __ret; \
50167 })
50168 #else
50169 #define vld2q_dup_s32(__p0) __extension__ ({ \
50170  int32x4x2_t __ret; \
50171  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50172  \
50173  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50174  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50175  __ret; \
50176 })
50177 #endif
50178 
50179 #ifdef __LITTLE_ENDIAN__
50180 #define vld2q_dup_s64(__p0) __extension__ ({ \
50181  int64x2x2_t __ret; \
50182  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50183  __ret; \
50184 })
50185 #else
50186 #define vld2q_dup_s64(__p0) __extension__ ({ \
50187  int64x2x2_t __ret; \
50188  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50189  \
50190  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50191  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50192  __ret; \
50193 })
50194 #endif
50195 
50196 #ifdef __LITTLE_ENDIAN__
50197 #define vld2q_dup_s16(__p0) __extension__ ({ \
50198  int16x8x2_t __ret; \
50199  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50200  __ret; \
50201 })
50202 #else
50203 #define vld2q_dup_s16(__p0) __extension__ ({ \
50204  int16x8x2_t __ret; \
50205  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50206  \
50207  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50208  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50209  __ret; \
50210 })
50211 #endif
50212 
50213 #ifdef __LITTLE_ENDIAN__
50214 #define vld2_dup_f64(__p0) __extension__ ({ \
50215  float64x1x2_t __ret; \
50216  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50217  __ret; \
50218 })
50219 #else
50220 #define vld2_dup_f64(__p0) __extension__ ({ \
50221  float64x1x2_t __ret; \
50222  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50223  __ret; \
50224 })
50225 #endif
50226 
50227 #ifdef __LITTLE_ENDIAN__
50228 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50229  poly64x1x2_t __s1 = __p1; \
50230  poly64x1x2_t __ret; \
50231  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50232  __ret; \
50233 })
50234 #else
50235 #define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50236  poly64x1x2_t __s1 = __p1; \
50237  poly64x1x2_t __ret; \
50238  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50239  __ret; \
50240 })
50241 #endif
50242 
50243 #ifdef __LITTLE_ENDIAN__
50244 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50245  poly8x16x2_t __s1 = __p1; \
50246  poly8x16x2_t __ret; \
50247  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
50248  __ret; \
50249 })
50250 #else
50251 #define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50252  poly8x16x2_t __s1 = __p1; \
50253  poly8x16x2_t __rev1; \
50254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50256  poly8x16x2_t __ret; \
50257  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
50258  \
50259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50261  __ret; \
50262 })
50263 #endif
50264 
50265 #ifdef __LITTLE_ENDIAN__
50266 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50267  poly64x2x2_t __s1 = __p1; \
50268  poly64x2x2_t __ret; \
50269  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
50270  __ret; \
50271 })
50272 #else
50273 #define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50274  poly64x2x2_t __s1 = __p1; \
50275  poly64x2x2_t __rev1; \
50276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50278  poly64x2x2_t __ret; \
50279  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
50280  \
50281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50283  __ret; \
50284 })
50285 #endif
50286 
50287 #ifdef __LITTLE_ENDIAN__
50288 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50289  uint8x16x2_t __s1 = __p1; \
50290  uint8x16x2_t __ret; \
50291  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
50292  __ret; \
50293 })
50294 #else
50295 #define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50296  uint8x16x2_t __s1 = __p1; \
50297  uint8x16x2_t __rev1; \
50298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50300  uint8x16x2_t __ret; \
50301  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
50302  \
50303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50305  __ret; \
50306 })
50307 #endif
50308 
50309 #ifdef __LITTLE_ENDIAN__
50310 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50311  uint64x2x2_t __s1 = __p1; \
50312  uint64x2x2_t __ret; \
50313  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
50314  __ret; \
50315 })
50316 #else
50317 #define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50318  uint64x2x2_t __s1 = __p1; \
50319  uint64x2x2_t __rev1; \
50320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50322  uint64x2x2_t __ret; \
50323  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
50324  \
50325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50327  __ret; \
50328 })
50329 #endif
50330 
50331 #ifdef __LITTLE_ENDIAN__
50332 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50333  int8x16x2_t __s1 = __p1; \
50334  int8x16x2_t __ret; \
50335  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
50336  __ret; \
50337 })
50338 #else
50339 #define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50340  int8x16x2_t __s1 = __p1; \
50341  int8x16x2_t __rev1; \
50342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50344  int8x16x2_t __ret; \
50345  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
50346  \
50347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50349  __ret; \
50350 })
50351 #endif
50352 
50353 #ifdef __LITTLE_ENDIAN__
50354 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50355  float64x2x2_t __s1 = __p1; \
50356  float64x2x2_t __ret; \
50357  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 42); \
50358  __ret; \
50359 })
50360 #else
50361 #define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50362  float64x2x2_t __s1 = __p1; \
50363  float64x2x2_t __rev1; \
50364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50366  float64x2x2_t __ret; \
50367  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
50368  \
50369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50371  __ret; \
50372 })
50373 #endif
50374 
50375 #ifdef __LITTLE_ENDIAN__
50376 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50377  int64x2x2_t __s1 = __p1; \
50378  int64x2x2_t __ret; \
50379  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 35); \
50380  __ret; \
50381 })
50382 #else
50383 #define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50384  int64x2x2_t __s1 = __p1; \
50385  int64x2x2_t __rev1; \
50386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50388  int64x2x2_t __ret; \
50389  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
50390  \
50391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50393  __ret; \
50394 })
50395 #endif
50396 
50397 #ifdef __LITTLE_ENDIAN__
50398 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50399  uint64x1x2_t __s1 = __p1; \
50400  uint64x1x2_t __ret; \
50401  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50402  __ret; \
50403 })
50404 #else
50405 #define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50406  uint64x1x2_t __s1 = __p1; \
50407  uint64x1x2_t __ret; \
50408  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50409  __ret; \
50410 })
50411 #endif
50412 
50413 #ifdef __LITTLE_ENDIAN__
50414 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50415  float64x1x2_t __s1 = __p1; \
50416  float64x1x2_t __ret; \
50417  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50418  __ret; \
50419 })
50420 #else
50421 #define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50422  float64x1x2_t __s1 = __p1; \
50423  float64x1x2_t __ret; \
50424  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50425  __ret; \
50426 })
50427 #endif
50428 
50429 #ifdef __LITTLE_ENDIAN__
50430 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50431  int64x1x2_t __s1 = __p1; \
50432  int64x1x2_t __ret; \
50433  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50434  __ret; \
50435 })
50436 #else
50437 #define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50438  int64x1x2_t __s1 = __p1; \
50439  int64x1x2_t __ret; \
50440  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50441  __ret; \
50442 })
50443 #endif
50444 
50445 #ifdef __LITTLE_ENDIAN__
50446 #define vld3_p64(__p0) __extension__ ({ \
50447  poly64x1x3_t __ret; \
50448  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50449  __ret; \
50450 })
50451 #else
50452 #define vld3_p64(__p0) __extension__ ({ \
50453  poly64x1x3_t __ret; \
50454  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50455  __ret; \
50456 })
50457 #endif
50458 
50459 #ifdef __LITTLE_ENDIAN__
50460 #define vld3q_p64(__p0) __extension__ ({ \
50461  poly64x2x3_t __ret; \
50462  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50463  __ret; \
50464 })
50465 #else
50466 #define vld3q_p64(__p0) __extension__ ({ \
50467  poly64x2x3_t __ret; \
50468  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50469  \
50470  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50471  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50472  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50473  __ret; \
50474 })
50475 #endif
50476 
50477 #ifdef __LITTLE_ENDIAN__
50478 #define vld3q_u64(__p0) __extension__ ({ \
50479  uint64x2x3_t __ret; \
50480  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50481  __ret; \
50482 })
50483 #else
50484 #define vld3q_u64(__p0) __extension__ ({ \
50485  uint64x2x3_t __ret; \
50486  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50487  \
50488  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50489  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50490  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50491  __ret; \
50492 })
50493 #endif
50494 
50495 #ifdef __LITTLE_ENDIAN__
50496 #define vld3q_f64(__p0) __extension__ ({ \
50497  float64x2x3_t __ret; \
50498  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50499  __ret; \
50500 })
50501 #else
50502 #define vld3q_f64(__p0) __extension__ ({ \
50503  float64x2x3_t __ret; \
50504  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50505  \
50506  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50507  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50508  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50509  __ret; \
50510 })
50511 #endif
50512 
50513 #ifdef __LITTLE_ENDIAN__
50514 #define vld3q_s64(__p0) __extension__ ({ \
50515  int64x2x3_t __ret; \
50516  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50517  __ret; \
50518 })
50519 #else
50520 #define vld3q_s64(__p0) __extension__ ({ \
50521  int64x2x3_t __ret; \
50522  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50523  \
50524  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50525  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50526  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50527  __ret; \
50528 })
50529 #endif
50530 
50531 #ifdef __LITTLE_ENDIAN__
50532 #define vld3_f64(__p0) __extension__ ({ \
50533  float64x1x3_t __ret; \
50534  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50535  __ret; \
50536 })
50537 #else
50538 #define vld3_f64(__p0) __extension__ ({ \
50539  float64x1x3_t __ret; \
50540  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50541  __ret; \
50542 })
50543 #endif
50544 
50545 #ifdef __LITTLE_ENDIAN__
50546 #define vld3_dup_p64(__p0) __extension__ ({ \
50547  poly64x1x3_t __ret; \
50548  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50549  __ret; \
50550 })
50551 #else
50552 #define vld3_dup_p64(__p0) __extension__ ({ \
50553  poly64x1x3_t __ret; \
50554  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50555  __ret; \
50556 })
50557 #endif
50558 
50559 #ifdef __LITTLE_ENDIAN__
50560 #define vld3q_dup_p8(__p0) __extension__ ({ \
50561  poly8x16x3_t __ret; \
50562  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50563  __ret; \
50564 })
50565 #else
50566 #define vld3q_dup_p8(__p0) __extension__ ({ \
50567  poly8x16x3_t __ret; \
50568  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50569  \
50570  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50571  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50572  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50573  __ret; \
50574 })
50575 #endif
50576 
50577 #ifdef __LITTLE_ENDIAN__
50578 #define vld3q_dup_p64(__p0) __extension__ ({ \
50579  poly64x2x3_t __ret; \
50580  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50581  __ret; \
50582 })
50583 #else
50584 #define vld3q_dup_p64(__p0) __extension__ ({ \
50585  poly64x2x3_t __ret; \
50586  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50587  \
50588  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50589  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50590  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50591  __ret; \
50592 })
50593 #endif
50594 
50595 #ifdef __LITTLE_ENDIAN__
50596 #define vld3q_dup_p16(__p0) __extension__ ({ \
50597  poly16x8x3_t __ret; \
50598  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50599  __ret; \
50600 })
50601 #else
50602 #define vld3q_dup_p16(__p0) __extension__ ({ \
50603  poly16x8x3_t __ret; \
50604  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50605  \
50606  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50607  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50608  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50609  __ret; \
50610 })
50611 #endif
50612 
50613 #ifdef __LITTLE_ENDIAN__
50614 #define vld3q_dup_u8(__p0) __extension__ ({ \
50615  uint8x16x3_t __ret; \
50616  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50617  __ret; \
50618 })
50619 #else
50620 #define vld3q_dup_u8(__p0) __extension__ ({ \
50621  uint8x16x3_t __ret; \
50622  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50623  \
50624  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50625  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50626  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50627  __ret; \
50628 })
50629 #endif
50630 
50631 #ifdef __LITTLE_ENDIAN__
50632 #define vld3q_dup_u32(__p0) __extension__ ({ \
50633  uint32x4x3_t __ret; \
50634  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50635  __ret; \
50636 })
50637 #else
50638 #define vld3q_dup_u32(__p0) __extension__ ({ \
50639  uint32x4x3_t __ret; \
50640  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50641  \
50642  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50643  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50644  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50645  __ret; \
50646 })
50647 #endif
50648 
50649 #ifdef __LITTLE_ENDIAN__
50650 #define vld3q_dup_u64(__p0) __extension__ ({ \
50651  uint64x2x3_t __ret; \
50652  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50653  __ret; \
50654 })
50655 #else
50656 #define vld3q_dup_u64(__p0) __extension__ ({ \
50657  uint64x2x3_t __ret; \
50658  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50659  \
50660  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50661  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50662  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50663  __ret; \
50664 })
50665 #endif
50666 
50667 #ifdef __LITTLE_ENDIAN__
50668 #define vld3q_dup_u16(__p0) __extension__ ({ \
50669  uint16x8x3_t __ret; \
50670  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50671  __ret; \
50672 })
50673 #else
50674 #define vld3q_dup_u16(__p0) __extension__ ({ \
50675  uint16x8x3_t __ret; \
50676  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50677  \
50678  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50679  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50680  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50681  __ret; \
50682 })
50683 #endif
50684 
50685 #ifdef __LITTLE_ENDIAN__
50686 #define vld3q_dup_s8(__p0) __extension__ ({ \
50687  int8x16x3_t __ret; \
50688  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50689  __ret; \
50690 })
50691 #else
50692 #define vld3q_dup_s8(__p0) __extension__ ({ \
50693  int8x16x3_t __ret; \
50694  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50695  \
50696  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50697  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50698  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50699  __ret; \
50700 })
50701 #endif
50702 
50703 #ifdef __LITTLE_ENDIAN__
50704 #define vld3q_dup_f64(__p0) __extension__ ({ \
50705  float64x2x3_t __ret; \
50706  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50707  __ret; \
50708 })
50709 #else
50710 #define vld3q_dup_f64(__p0) __extension__ ({ \
50711  float64x2x3_t __ret; \
50712  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50713  \
50714  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50715  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50716  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50717  __ret; \
50718 })
50719 #endif
50720 
50721 #ifdef __LITTLE_ENDIAN__
50722 #define vld3q_dup_f32(__p0) __extension__ ({ \
50723  float32x4x3_t __ret; \
50724  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50725  __ret; \
50726 })
50727 #else
50728 #define vld3q_dup_f32(__p0) __extension__ ({ \
50729  float32x4x3_t __ret; \
50730  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50731  \
50732  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50733  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50734  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50735  __ret; \
50736 })
50737 #endif
50738 
50739 #ifdef __LITTLE_ENDIAN__
50740 #define vld3q_dup_f16(__p0) __extension__ ({ \
50741  float16x8x3_t __ret; \
50742  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50743  __ret; \
50744 })
50745 #else
50746 #define vld3q_dup_f16(__p0) __extension__ ({ \
50747  float16x8x3_t __ret; \
50748  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50749  \
50750  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50751  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50752  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50753  __ret; \
50754 })
50755 #endif
50756 
50757 #ifdef __LITTLE_ENDIAN__
50758 #define vld3q_dup_s32(__p0) __extension__ ({ \
50759  int32x4x3_t __ret; \
50760  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50761  __ret; \
50762 })
50763 #else
50764 #define vld3q_dup_s32(__p0) __extension__ ({ \
50765  int32x4x3_t __ret; \
50766  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50767  \
50768  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50769  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50770  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50771  __ret; \
50772 })
50773 #endif
50774 
50775 #ifdef __LITTLE_ENDIAN__
50776 #define vld3q_dup_s64(__p0) __extension__ ({ \
50777  int64x2x3_t __ret; \
50778  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50779  __ret; \
50780 })
50781 #else
50782 #define vld3q_dup_s64(__p0) __extension__ ({ \
50783  int64x2x3_t __ret; \
50784  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50785  \
50786  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50787  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50788  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50789  __ret; \
50790 })
50791 #endif
50792 
50793 #ifdef __LITTLE_ENDIAN__
50794 #define vld3q_dup_s16(__p0) __extension__ ({ \
50795  int16x8x3_t __ret; \
50796  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50797  __ret; \
50798 })
50799 #else
50800 #define vld3q_dup_s16(__p0) __extension__ ({ \
50801  int16x8x3_t __ret; \
50802  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50803  \
50804  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50805  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50806  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50807  __ret; \
50808 })
50809 #endif
50810 
50811 #ifdef __LITTLE_ENDIAN__
50812 #define vld3_dup_f64(__p0) __extension__ ({ \
50813  float64x1x3_t __ret; \
50814  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50815  __ret; \
50816 })
50817 #else
50818 #define vld3_dup_f64(__p0) __extension__ ({ \
50819  float64x1x3_t __ret; \
50820  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50821  __ret; \
50822 })
50823 #endif
50824 
50825 #ifdef __LITTLE_ENDIAN__
50826 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50827  poly64x1x3_t __s1 = __p1; \
50828  poly64x1x3_t __ret; \
50829  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50830  __ret; \
50831 })
50832 #else
50833 #define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50834  poly64x1x3_t __s1 = __p1; \
50835  poly64x1x3_t __ret; \
50836  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50837  __ret; \
50838 })
50839 #endif
50840 
50841 #ifdef __LITTLE_ENDIAN__
50842 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50843  poly8x16x3_t __s1 = __p1; \
50844  poly8x16x3_t __ret; \
50845  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
50846  __ret; \
50847 })
50848 #else
50849 #define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50850  poly8x16x3_t __s1 = __p1; \
50851  poly8x16x3_t __rev1; \
50852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50854  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50855  poly8x16x3_t __ret; \
50856  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
50857  \
50858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50860  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50861  __ret; \
50862 })
50863 #endif
50864 
50865 #ifdef __LITTLE_ENDIAN__
50866 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50867  poly64x2x3_t __s1 = __p1; \
50868  poly64x2x3_t __ret; \
50869  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
50870  __ret; \
50871 })
50872 #else
50873 #define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50874  poly64x2x3_t __s1 = __p1; \
50875  poly64x2x3_t __rev1; \
50876  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50877  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50878  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50879  poly64x2x3_t __ret; \
50880  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
50881  \
50882  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50883  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50884  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50885  __ret; \
50886 })
50887 #endif
50888 
50889 #ifdef __LITTLE_ENDIAN__
50890 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50891  uint8x16x3_t __s1 = __p1; \
50892  uint8x16x3_t __ret; \
50893  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
50894  __ret; \
50895 })
50896 #else
50897 #define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50898  uint8x16x3_t __s1 = __p1; \
50899  uint8x16x3_t __rev1; \
50900  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50901  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50902  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50903  uint8x16x3_t __ret; \
50904  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
50905  \
50906  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50907  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50908  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50909  __ret; \
50910 })
50911 #endif
50912 
50913 #ifdef __LITTLE_ENDIAN__
50914 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50915  uint64x2x3_t __s1 = __p1; \
50916  uint64x2x3_t __ret; \
50917  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
50918  __ret; \
50919 })
50920 #else
50921 #define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50922  uint64x2x3_t __s1 = __p1; \
50923  uint64x2x3_t __rev1; \
50924  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50925  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50926  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50927  uint64x2x3_t __ret; \
50928  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
50929  \
50930  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50931  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50932  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50933  __ret; \
50934 })
50935 #endif
50936 
50937 #ifdef __LITTLE_ENDIAN__
50938 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50939  int8x16x3_t __s1 = __p1; \
50940  int8x16x3_t __ret; \
50941  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
50942  __ret; \
50943 })
50944 #else
50945 #define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50946  int8x16x3_t __s1 = __p1; \
50947  int8x16x3_t __rev1; \
50948  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50949  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50950  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50951  int8x16x3_t __ret; \
50952  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
50953  \
50954  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50955  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50956  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50957  __ret; \
50958 })
50959 #endif
50960 
50961 #ifdef __LITTLE_ENDIAN__
50962 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50963  float64x2x3_t __s1 = __p1; \
50964  float64x2x3_t __ret; \
50965  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
50966  __ret; \
50967 })
50968 #else
50969 #define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50970  float64x2x3_t __s1 = __p1; \
50971  float64x2x3_t __rev1; \
50972  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50973  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50974  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50975  float64x2x3_t __ret; \
50976  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
50977  \
50978  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50979  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50980  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50981  __ret; \
50982 })
50983 #endif
50984 
50985 #ifdef __LITTLE_ENDIAN__
50986 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50987  int64x2x3_t __s1 = __p1; \
50988  int64x2x3_t __ret; \
50989  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
50990  __ret; \
50991 })
50992 #else
50993 #define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50994  int64x2x3_t __s1 = __p1; \
50995  int64x2x3_t __rev1; \
50996  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50997  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50998  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50999  int64x2x3_t __ret; \
51000  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
51001  \
51002  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51003  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51004  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51005  __ret; \
51006 })
51007 #endif
51008 
51009 #ifdef __LITTLE_ENDIAN__
51010 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51011  uint64x1x3_t __s1 = __p1; \
51012  uint64x1x3_t __ret; \
51013  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51014  __ret; \
51015 })
51016 #else
51017 #define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51018  uint64x1x3_t __s1 = __p1; \
51019  uint64x1x3_t __ret; \
51020  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51021  __ret; \
51022 })
51023 #endif
51024 
51025 #ifdef __LITTLE_ENDIAN__
51026 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51027  float64x1x3_t __s1 = __p1; \
51028  float64x1x3_t __ret; \
51029  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51030  __ret; \
51031 })
51032 #else
51033 #define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51034  float64x1x3_t __s1 = __p1; \
51035  float64x1x3_t __ret; \
51036  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51037  __ret; \
51038 })
51039 #endif
51040 
51041 #ifdef __LITTLE_ENDIAN__
51042 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51043  int64x1x3_t __s1 = __p1; \
51044  int64x1x3_t __ret; \
51045  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51046  __ret; \
51047 })
51048 #else
51049 #define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51050  int64x1x3_t __s1 = __p1; \
51051  int64x1x3_t __ret; \
51052  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51053  __ret; \
51054 })
51055 #endif
51056 
51057 #ifdef __LITTLE_ENDIAN__
51058 #define vld4_p64(__p0) __extension__ ({ \
51059  poly64x1x4_t __ret; \
51060  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51061  __ret; \
51062 })
51063 #else
51064 #define vld4_p64(__p0) __extension__ ({ \
51065  poly64x1x4_t __ret; \
51066  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51067  __ret; \
51068 })
51069 #endif
51070 
51071 #ifdef __LITTLE_ENDIAN__
51072 #define vld4q_p64(__p0) __extension__ ({ \
51073  poly64x2x4_t __ret; \
51074  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51075  __ret; \
51076 })
51077 #else
51078 #define vld4q_p64(__p0) __extension__ ({ \
51079  poly64x2x4_t __ret; \
51080  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51081  \
51082  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51083  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51084  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51085  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51086  __ret; \
51087 })
51088 #endif
51089 
51090 #ifdef __LITTLE_ENDIAN__
51091 #define vld4q_u64(__p0) __extension__ ({ \
51092  uint64x2x4_t __ret; \
51093  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51094  __ret; \
51095 })
51096 #else
51097 #define vld4q_u64(__p0) __extension__ ({ \
51098  uint64x2x4_t __ret; \
51099  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51100  \
51101  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51102  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51103  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51104  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51105  __ret; \
51106 })
51107 #endif
51108 
51109 #ifdef __LITTLE_ENDIAN__
51110 #define vld4q_f64(__p0) __extension__ ({ \
51111  float64x2x4_t __ret; \
51112  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51113  __ret; \
51114 })
51115 #else
51116 #define vld4q_f64(__p0) __extension__ ({ \
51117  float64x2x4_t __ret; \
51118  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51119  \
51120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51123  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51124  __ret; \
51125 })
51126 #endif
51127 
51128 #ifdef __LITTLE_ENDIAN__
51129 #define vld4q_s64(__p0) __extension__ ({ \
51130  int64x2x4_t __ret; \
51131  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51132  __ret; \
51133 })
51134 #else
51135 #define vld4q_s64(__p0) __extension__ ({ \
51136  int64x2x4_t __ret; \
51137  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51138  \
51139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51141  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51142  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51143  __ret; \
51144 })
51145 #endif
51146 
51147 #ifdef __LITTLE_ENDIAN__
51148 #define vld4_f64(__p0) __extension__ ({ \
51149  float64x1x4_t __ret; \
51150  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51151  __ret; \
51152 })
51153 #else
51154 #define vld4_f64(__p0) __extension__ ({ \
51155  float64x1x4_t __ret; \
51156  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51157  __ret; \
51158 })
51159 #endif
51160 
51161 #ifdef __LITTLE_ENDIAN__
51162 #define vld4_dup_p64(__p0) __extension__ ({ \
51163  poly64x1x4_t __ret; \
51164  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51165  __ret; \
51166 })
51167 #else
51168 #define vld4_dup_p64(__p0) __extension__ ({ \
51169  poly64x1x4_t __ret; \
51170  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51171  __ret; \
51172 })
51173 #endif
51174 
51175 #ifdef __LITTLE_ENDIAN__
51176 #define vld4q_dup_p8(__p0) __extension__ ({ \
51177  poly8x16x4_t __ret; \
51178  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51179  __ret; \
51180 })
51181 #else
51182 #define vld4q_dup_p8(__p0) __extension__ ({ \
51183  poly8x16x4_t __ret; \
51184  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51185  \
51186  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51187  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51188  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51189  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51190  __ret; \
51191 })
51192 #endif
51193 
51194 #ifdef __LITTLE_ENDIAN__
51195 #define vld4q_dup_p64(__p0) __extension__ ({ \
51196  poly64x2x4_t __ret; \
51197  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51198  __ret; \
51199 })
51200 #else
51201 #define vld4q_dup_p64(__p0) __extension__ ({ \
51202  poly64x2x4_t __ret; \
51203  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51204  \
51205  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51206  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51207  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51208  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51209  __ret; \
51210 })
51211 #endif
51212 
51213 #ifdef __LITTLE_ENDIAN__
51214 #define vld4q_dup_p16(__p0) __extension__ ({ \
51215  poly16x8x4_t __ret; \
51216  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51217  __ret; \
51218 })
51219 #else
51220 #define vld4q_dup_p16(__p0) __extension__ ({ \
51221  poly16x8x4_t __ret; \
51222  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51223  \
51224  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51225  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51226  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51227  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51228  __ret; \
51229 })
51230 #endif
51231 
51232 #ifdef __LITTLE_ENDIAN__
51233 #define vld4q_dup_u8(__p0) __extension__ ({ \
51234  uint8x16x4_t __ret; \
51235  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51236  __ret; \
51237 })
51238 #else
51239 #define vld4q_dup_u8(__p0) __extension__ ({ \
51240  uint8x16x4_t __ret; \
51241  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51242  \
51243  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51244  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51245  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51246  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51247  __ret; \
51248 })
51249 #endif
51250 
51251 #ifdef __LITTLE_ENDIAN__
51252 #define vld4q_dup_u32(__p0) __extension__ ({ \
51253  uint32x4x4_t __ret; \
51254  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51255  __ret; \
51256 })
51257 #else
51258 #define vld4q_dup_u32(__p0) __extension__ ({ \
51259  uint32x4x4_t __ret; \
51260  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51261  \
51262  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51263  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51264  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51265  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51266  __ret; \
51267 })
51268 #endif
51269 
51270 #ifdef __LITTLE_ENDIAN__
51271 #define vld4q_dup_u64(__p0) __extension__ ({ \
51272  uint64x2x4_t __ret; \
51273  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51274  __ret; \
51275 })
51276 #else
51277 #define vld4q_dup_u64(__p0) __extension__ ({ \
51278  uint64x2x4_t __ret; \
51279  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51280  \
51281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51283  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51284  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51285  __ret; \
51286 })
51287 #endif
51288 
51289 #ifdef __LITTLE_ENDIAN__
51290 #define vld4q_dup_u16(__p0) __extension__ ({ \
51291  uint16x8x4_t __ret; \
51292  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51293  __ret; \
51294 })
51295 #else
51296 #define vld4q_dup_u16(__p0) __extension__ ({ \
51297  uint16x8x4_t __ret; \
51298  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51299  \
51300  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51301  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51302  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51303  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51304  __ret; \
51305 })
51306 #endif
51307 
51308 #ifdef __LITTLE_ENDIAN__
51309 #define vld4q_dup_s8(__p0) __extension__ ({ \
51310  int8x16x4_t __ret; \
51311  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51312  __ret; \
51313 })
51314 #else
51315 #define vld4q_dup_s8(__p0) __extension__ ({ \
51316  int8x16x4_t __ret; \
51317  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51318  \
51319  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51320  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51321  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51322  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51323  __ret; \
51324 })
51325 #endif
51326 
51327 #ifdef __LITTLE_ENDIAN__
51328 #define vld4q_dup_f64(__p0) __extension__ ({ \
51329  float64x2x4_t __ret; \
51330  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51331  __ret; \
51332 })
51333 #else
51334 #define vld4q_dup_f64(__p0) __extension__ ({ \
51335  float64x2x4_t __ret; \
51336  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51337  \
51338  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51339  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51340  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51341  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51342  __ret; \
51343 })
51344 #endif
51345 
51346 #ifdef __LITTLE_ENDIAN__
51347 #define vld4q_dup_f32(__p0) __extension__ ({ \
51348  float32x4x4_t __ret; \
51349  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51350  __ret; \
51351 })
51352 #else
51353 #define vld4q_dup_f32(__p0) __extension__ ({ \
51354  float32x4x4_t __ret; \
51355  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51356  \
51357  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51358  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51359  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51360  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51361  __ret; \
51362 })
51363 #endif
51364 
51365 #ifdef __LITTLE_ENDIAN__
51366 #define vld4q_dup_f16(__p0) __extension__ ({ \
51367  float16x8x4_t __ret; \
51368  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51369  __ret; \
51370 })
51371 #else
51372 #define vld4q_dup_f16(__p0) __extension__ ({ \
51373  float16x8x4_t __ret; \
51374  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51375  \
51376  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51377  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51378  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51379  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51380  __ret; \
51381 })
51382 #endif
51383 
51384 #ifdef __LITTLE_ENDIAN__
51385 #define vld4q_dup_s32(__p0) __extension__ ({ \
51386  int32x4x4_t __ret; \
51387  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51388  __ret; \
51389 })
51390 #else
51391 #define vld4q_dup_s32(__p0) __extension__ ({ \
51392  int32x4x4_t __ret; \
51393  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51394  \
51395  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51396  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51397  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51398  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51399  __ret; \
51400 })
51401 #endif
51402 
51403 #ifdef __LITTLE_ENDIAN__
51404 #define vld4q_dup_s64(__p0) __extension__ ({ \
51405  int64x2x4_t __ret; \
51406  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51407  __ret; \
51408 })
51409 #else
51410 #define vld4q_dup_s64(__p0) __extension__ ({ \
51411  int64x2x4_t __ret; \
51412  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51413  \
51414  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51415  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51416  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51417  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51418  __ret; \
51419 })
51420 #endif
51421 
51422 #ifdef __LITTLE_ENDIAN__
51423 #define vld4q_dup_s16(__p0) __extension__ ({ \
51424  int16x8x4_t __ret; \
51425  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51426  __ret; \
51427 })
51428 #else
51429 #define vld4q_dup_s16(__p0) __extension__ ({ \
51430  int16x8x4_t __ret; \
51431  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51432  \
51433  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51434  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51435  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51436  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51437  __ret; \
51438 })
51439 #endif
51440 
51441 #ifdef __LITTLE_ENDIAN__
51442 #define vld4_dup_f64(__p0) __extension__ ({ \
51443  float64x1x4_t __ret; \
51444  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51445  __ret; \
51446 })
51447 #else
51448 #define vld4_dup_f64(__p0) __extension__ ({ \
51449  float64x1x4_t __ret; \
51450  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51451  __ret; \
51452 })
51453 #endif
51454 
51455 #ifdef __LITTLE_ENDIAN__
51456 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51457  poly64x1x4_t __s1 = __p1; \
51458  poly64x1x4_t __ret; \
51459  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51460  __ret; \
51461 })
51462 #else
51463 #define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51464  poly64x1x4_t __s1 = __p1; \
51465  poly64x1x4_t __ret; \
51466  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51467  __ret; \
51468 })
51469 #endif
51470 
51471 #ifdef __LITTLE_ENDIAN__
51472 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51473  poly8x16x4_t __s1 = __p1; \
51474  poly8x16x4_t __ret; \
51475  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
51476  __ret; \
51477 })
51478 #else
51479 #define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51480  poly8x16x4_t __s1 = __p1; \
51481  poly8x16x4_t __rev1; \
51482  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51483  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51484  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51485  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51486  poly8x16x4_t __ret; \
51487  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
51488  \
51489  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51490  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51491  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51492  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51493  __ret; \
51494 })
51495 #endif
51496 
51497 #ifdef __LITTLE_ENDIAN__
51498 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51499  poly64x2x4_t __s1 = __p1; \
51500  poly64x2x4_t __ret; \
51501  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
51502  __ret; \
51503 })
51504 #else
51505 #define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51506  poly64x2x4_t __s1 = __p1; \
51507  poly64x2x4_t __rev1; \
51508  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51509  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51510  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51511  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51512  poly64x2x4_t __ret; \
51513  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
51514  \
51515  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51516  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51517  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51518  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51519  __ret; \
51520 })
51521 #endif
51522 
51523 #ifdef __LITTLE_ENDIAN__
51524 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51525  uint8x16x4_t __s1 = __p1; \
51526  uint8x16x4_t __ret; \
51527  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
51528  __ret; \
51529 })
51530 #else
51531 #define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51532  uint8x16x4_t __s1 = __p1; \
51533  uint8x16x4_t __rev1; \
51534  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51535  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51536  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51537  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51538  uint8x16x4_t __ret; \
51539  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
51540  \
51541  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51542  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51543  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51544  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51545  __ret; \
51546 })
51547 #endif
51548 
51549 #ifdef __LITTLE_ENDIAN__
51550 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51551  uint64x2x4_t __s1 = __p1; \
51552  uint64x2x4_t __ret; \
51553  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
51554  __ret; \
51555 })
51556 #else
51557 #define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51558  uint64x2x4_t __s1 = __p1; \
51559  uint64x2x4_t __rev1; \
51560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51563  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51564  uint64x2x4_t __ret; \
51565  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
51566  \
51567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51569  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51570  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51571  __ret; \
51572 })
51573 #endif
51574 
51575 #ifdef __LITTLE_ENDIAN__
51576 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51577  int8x16x4_t __s1 = __p1; \
51578  int8x16x4_t __ret; \
51579  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
51580  __ret; \
51581 })
51582 #else
51583 #define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51584  int8x16x4_t __s1 = __p1; \
51585  int8x16x4_t __rev1; \
51586  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51587  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51588  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51589  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51590  int8x16x4_t __ret; \
51591  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
51592  \
51593  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51594  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51595  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51596  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51597  __ret; \
51598 })
51599 #endif
51600 
51601 #ifdef __LITTLE_ENDIAN__
51602 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51603  float64x2x4_t __s1 = __p1; \
51604  float64x2x4_t __ret; \
51605  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
51606  __ret; \
51607 })
51608 #else
51609 #define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51610  float64x2x4_t __s1 = __p1; \
51611  float64x2x4_t __rev1; \
51612  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51613  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51614  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51615  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51616  float64x2x4_t __ret; \
51617  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
51618  \
51619  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51620  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51621  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51622  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51623  __ret; \
51624 })
51625 #endif
51626 
51627 #ifdef __LITTLE_ENDIAN__
51628 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51629  int64x2x4_t __s1 = __p1; \
51630  int64x2x4_t __ret; \
51631  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
51632  __ret; \
51633 })
51634 #else
51635 #define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51636  int64x2x4_t __s1 = __p1; \
51637  int64x2x4_t __rev1; \
51638  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51639  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51640  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51641  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51642  int64x2x4_t __ret; \
51643  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
51644  \
51645  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51646  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51647  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51648  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51649  __ret; \
51650 })
51651 #endif
51652 
51653 #ifdef __LITTLE_ENDIAN__
51654 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51655  uint64x1x4_t __s1 = __p1; \
51656  uint64x1x4_t __ret; \
51657  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51658  __ret; \
51659 })
51660 #else
51661 #define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51662  uint64x1x4_t __s1 = __p1; \
51663  uint64x1x4_t __ret; \
51664  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51665  __ret; \
51666 })
51667 #endif
51668 
51669 #ifdef __LITTLE_ENDIAN__
51670 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51671  float64x1x4_t __s1 = __p1; \
51672  float64x1x4_t __ret; \
51673  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51674  __ret; \
51675 })
51676 #else
51677 #define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51678  float64x1x4_t __s1 = __p1; \
51679  float64x1x4_t __ret; \
51680  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51681  __ret; \
51682 })
51683 #endif
51684 
51685 #ifdef __LITTLE_ENDIAN__
51686 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51687  int64x1x4_t __s1 = __p1; \
51688  int64x1x4_t __ret; \
51689  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51690  __ret; \
51691 })
51692 #else
51693 #define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51694  int64x1x4_t __s1 = __p1; \
51695  int64x1x4_t __ret; \
51696  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51697  __ret; \
51698 })
51699 #endif
51700 
51701 #ifdef __LITTLE_ENDIAN__
51702 #define vldrq_p128(__p0) __extension__ ({ \
51703  poly128_t __ret; \
51704  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51705  __ret; \
51706 })
51707 #else
51708 #define vldrq_p128(__p0) __extension__ ({ \
51709  poly128_t __ret; \
51710  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51711  __ret; \
51712 })
51713 #endif
51714 
51715 #ifdef __LITTLE_ENDIAN__
51716 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51717  float64x2_t __ret;
51718  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51719  return __ret;
51720 }
51721 #else
51722 __ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51723  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51724  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51725  float64x2_t __ret;
51726  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51728  return __ret;
51729 }
51730 #endif
51731 
51732 #ifdef __LITTLE_ENDIAN__
51733 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51734  float64x1_t __ret;
51735  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51736  return __ret;
51737 }
51738 #else
51739 __ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51740  float64x1_t __ret;
51741  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51742  return __ret;
51743 }
51744 #endif
51745 
51746 #ifdef __LITTLE_ENDIAN__
51747 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51748  float64_t __ret;
51749  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__p0);
51750  return __ret;
51751 }
51752 #else
51753 __ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51754  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51755  float64_t __ret;
51756  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__rev0);
51757  return __ret;
51758 }
51759 #endif
51760 
51761 #ifdef __LITTLE_ENDIAN__
51762 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51763  float32_t __ret;
51764  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__p0);
51765  return __ret;
51766 }
51767 #else
51768 __ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51769  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51770  float32_t __ret;
51771  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__rev0);
51772  return __ret;
51773 }
51774 #endif
51775 
51776 #ifdef __LITTLE_ENDIAN__
51777 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51778  float32_t __ret;
51779  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__p0);
51780  return __ret;
51781 }
51782 #else
51783 __ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51784  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51785  float32_t __ret;
51786  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__rev0);
51787  return __ret;
51788 }
51789 #endif
51790 
51791 #ifdef __LITTLE_ENDIAN__
51792 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51793  uint8_t __ret;
51794  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__p0);
51795  return __ret;
51796 }
51797 #else
51798 __ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51799  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51800  uint8_t __ret;
51801  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__rev0);
51802  return __ret;
51803 }
51804 #endif
51805 
51806 #ifdef __LITTLE_ENDIAN__
51807 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51808  uint32_t __ret;
51809  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__p0);
51810  return __ret;
51811 }
51812 #else
51813 __ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51814  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51815  uint32_t __ret;
51816  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__rev0);
51817  return __ret;
51818 }
51819 #endif
51820 
51821 #ifdef __LITTLE_ENDIAN__
51822 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51823  uint16_t __ret;
51824  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__p0);
51825  return __ret;
51826 }
51827 #else
51828 __ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51829  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51830  uint16_t __ret;
51831  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__rev0);
51832  return __ret;
51833 }
51834 #endif
51835 
51836 #ifdef __LITTLE_ENDIAN__
51837 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
51838  int8_t __ret;
51839  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__p0);
51840  return __ret;
51841 }
51842 #else
51843 __ai int8_t vmaxvq_s8(int8x16_t __p0) {
51844  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51845  int8_t __ret;
51846  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__rev0);
51847  return __ret;
51848 }
51849 #endif
51850 
51851 #ifdef __LITTLE_ENDIAN__
51852 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
51853  float64_t __ret;
51854  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__p0);
51855  return __ret;
51856 }
51857 #else
51858 __ai float64_t vmaxvq_f64(float64x2_t __p0) {
51859  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51860  float64_t __ret;
51861  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__rev0);
51862  return __ret;
51863 }
51864 #endif
51865 
51866 #ifdef __LITTLE_ENDIAN__
51867 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
51868  float32_t __ret;
51869  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__p0);
51870  return __ret;
51871 }
51872 #else
51873 __ai float32_t vmaxvq_f32(float32x4_t __p0) {
51874  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51875  float32_t __ret;
51876  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__rev0);
51877  return __ret;
51878 }
51879 #endif
51880 
51881 #ifdef __LITTLE_ENDIAN__
51882 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
51883  int32_t __ret;
51884  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__p0);
51885  return __ret;
51886 }
51887 #else
51888 __ai int32_t vmaxvq_s32(int32x4_t __p0) {
51889  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51890  int32_t __ret;
51891  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__rev0);
51892  return __ret;
51893 }
51894 #endif
51895 
51896 #ifdef __LITTLE_ENDIAN__
51897 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
51898  int16_t __ret;
51899  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__p0);
51900  return __ret;
51901 }
51902 #else
51903 __ai int16_t vmaxvq_s16(int16x8_t __p0) {
51904  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51905  int16_t __ret;
51906  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__rev0);
51907  return __ret;
51908 }
51909 #endif
51910 
51911 #ifdef __LITTLE_ENDIAN__
51912 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51913  uint8_t __ret;
51914  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__p0);
51915  return __ret;
51916 }
51917 #else
51918 __ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51919  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51920  uint8_t __ret;
51921  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__rev0);
51922  return __ret;
51923 }
51924 #endif
51925 
51926 #ifdef __LITTLE_ENDIAN__
51927 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51928  uint32_t __ret;
51929  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__p0);
51930  return __ret;
51931 }
51932 #else
51933 __ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51934  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51935  uint32_t __ret;
51936  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__rev0);
51937  return __ret;
51938 }
51939 #endif
51940 
51941 #ifdef __LITTLE_ENDIAN__
51942 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51943  uint16_t __ret;
51944  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__p0);
51945  return __ret;
51946 }
51947 #else
51948 __ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51949  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51950  uint16_t __ret;
51951  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__rev0);
51952  return __ret;
51953 }
51954 #endif
51955 
51956 #ifdef __LITTLE_ENDIAN__
51957 __ai int8_t vmaxv_s8(int8x8_t __p0) {
51958  int8_t __ret;
51959  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__p0);
51960  return __ret;
51961 }
51962 #else
51963 __ai int8_t vmaxv_s8(int8x8_t __p0) {
51964  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51965  int8_t __ret;
51966  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__rev0);
51967  return __ret;
51968 }
51969 #endif
51970 
51971 #ifdef __LITTLE_ENDIAN__
51972 __ai float32_t vmaxv_f32(float32x2_t __p0) {
51973  float32_t __ret;
51974  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__p0);
51975  return __ret;
51976 }
51977 #else
51978 __ai float32_t vmaxv_f32(float32x2_t __p0) {
51979  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51980  float32_t __ret;
51981  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__rev0);
51982  return __ret;
51983 }
51984 #endif
51985 
51986 #ifdef __LITTLE_ENDIAN__
51987 __ai int32_t vmaxv_s32(int32x2_t __p0) {
51988  int32_t __ret;
51989  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__p0);
51990  return __ret;
51991 }
51992 #else
51993 __ai int32_t vmaxv_s32(int32x2_t __p0) {
51994  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51995  int32_t __ret;
51996  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__rev0);
51997  return __ret;
51998 }
51999 #endif
52000 
52001 #ifdef __LITTLE_ENDIAN__
52002 __ai int16_t vmaxv_s16(int16x4_t __p0) {
52003  int16_t __ret;
52004  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__p0);
52005  return __ret;
52006 }
52007 #else
52008 __ai int16_t vmaxv_s16(int16x4_t __p0) {
52009  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52010  int16_t __ret;
52011  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__rev0);
52012  return __ret;
52013 }
52014 #endif
52015 
52016 #ifdef __LITTLE_ENDIAN__
52017 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52018  float64x2_t __ret;
52019  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52020  return __ret;
52021 }
52022 #else
52023 __ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52024  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52025  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52026  float64x2_t __ret;
52027  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52028  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52029  return __ret;
52030 }
52031 #endif
52032 
52033 #ifdef __LITTLE_ENDIAN__
52034 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52035  float64x1_t __ret;
52036  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52037  return __ret;
52038 }
52039 #else
52040 __ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52041  float64x1_t __ret;
52042  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52043  return __ret;
52044 }
52045 #endif
52046 
52047 #ifdef __LITTLE_ENDIAN__
52048 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
52049  float64_t __ret;
52050  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__p0);
52051  return __ret;
52052 }
52053 #else
52054 __ai float64_t vminnmvq_f64(float64x2_t __p0) {
52055  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52056  float64_t __ret;
52057  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__rev0);
52058  return __ret;
52059 }
52060 #endif
52061 
52062 #ifdef __LITTLE_ENDIAN__
52063 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
52064  float32_t __ret;
52065  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__p0);
52066  return __ret;
52067 }
52068 #else
52069 __ai float32_t vminnmvq_f32(float32x4_t __p0) {
52070  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52071  float32_t __ret;
52072  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__rev0);
52073  return __ret;
52074 }
52075 #endif
52076 
52077 #ifdef __LITTLE_ENDIAN__
52078 __ai float32_t vminnmv_f32(float32x2_t __p0) {
52079  float32_t __ret;
52080  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__p0);
52081  return __ret;
52082 }
52083 #else
52084 __ai float32_t vminnmv_f32(float32x2_t __p0) {
52085  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52086  float32_t __ret;
52087  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__rev0);
52088  return __ret;
52089 }
52090 #endif
52091 
52092 #ifdef __LITTLE_ENDIAN__
52093 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
52094  uint8_t __ret;
52095  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__p0);
52096  return __ret;
52097 }
52098 #else
52099 __ai uint8_t vminvq_u8(uint8x16_t __p0) {
52100  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52101  uint8_t __ret;
52102  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__rev0);
52103  return __ret;
52104 }
52105 #endif
52106 
52107 #ifdef __LITTLE_ENDIAN__
52108 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
52109  uint32_t __ret;
52110  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__p0);
52111  return __ret;
52112 }
52113 #else
52114 __ai uint32_t vminvq_u32(uint32x4_t __p0) {
52115  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52116  uint32_t __ret;
52117  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__rev0);
52118  return __ret;
52119 }
52120 #endif
52121 
52122 #ifdef __LITTLE_ENDIAN__
52123 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
52124  uint16_t __ret;
52125  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__p0);
52126  return __ret;
52127 }
52128 #else
52129 __ai uint16_t vminvq_u16(uint16x8_t __p0) {
52130  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52131  uint16_t __ret;
52132  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__rev0);
52133  return __ret;
52134 }
52135 #endif
52136 
52137 #ifdef __LITTLE_ENDIAN__
52138 __ai int8_t vminvq_s8(int8x16_t __p0) {
52139  int8_t __ret;
52140  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__p0);
52141  return __ret;
52142 }
52143 #else
52144 __ai int8_t vminvq_s8(int8x16_t __p0) {
52145  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52146  int8_t __ret;
52147  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__rev0);
52148  return __ret;
52149 }
52150 #endif
52151 
52152 #ifdef __LITTLE_ENDIAN__
52153 __ai float64_t vminvq_f64(float64x2_t __p0) {
52154  float64_t __ret;
52155  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__p0);
52156  return __ret;
52157 }
52158 #else
52159 __ai float64_t vminvq_f64(float64x2_t __p0) {
52160  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52161  float64_t __ret;
52162  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__rev0);
52163  return __ret;
52164 }
52165 #endif
52166 
52167 #ifdef __LITTLE_ENDIAN__
52168 __ai float32_t vminvq_f32(float32x4_t __p0) {
52169  float32_t __ret;
52170  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__p0);
52171  return __ret;
52172 }
52173 #else
52174 __ai float32_t vminvq_f32(float32x4_t __p0) {
52175  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52176  float32_t __ret;
52177  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__rev0);
52178  return __ret;
52179 }
52180 #endif
52181 
52182 #ifdef __LITTLE_ENDIAN__
52183 __ai int32_t vminvq_s32(int32x4_t __p0) {
52184  int32_t __ret;
52185  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__p0);
52186  return __ret;
52187 }
52188 #else
52189 __ai int32_t vminvq_s32(int32x4_t __p0) {
52190  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52191  int32_t __ret;
52192  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__rev0);
52193  return __ret;
52194 }
52195 #endif
52196 
52197 #ifdef __LITTLE_ENDIAN__
52198 __ai int16_t vminvq_s16(int16x8_t __p0) {
52199  int16_t __ret;
52200  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__p0);
52201  return __ret;
52202 }
52203 #else
52204 __ai int16_t vminvq_s16(int16x8_t __p0) {
52205  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52206  int16_t __ret;
52207  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__rev0);
52208  return __ret;
52209 }
52210 #endif
52211 
52212 #ifdef __LITTLE_ENDIAN__
52213 __ai uint8_t vminv_u8(uint8x8_t __p0) {
52214  uint8_t __ret;
52215  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__p0);
52216  return __ret;
52217 }
52218 #else
52219 __ai uint8_t vminv_u8(uint8x8_t __p0) {
52220  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52221  uint8_t __ret;
52222  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__rev0);
52223  return __ret;
52224 }
52225 #endif
52226 
52227 #ifdef __LITTLE_ENDIAN__
52228 __ai uint32_t vminv_u32(uint32x2_t __p0) {
52229  uint32_t __ret;
52230  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__p0);
52231  return __ret;
52232 }
52233 #else
52234 __ai uint32_t vminv_u32(uint32x2_t __p0) {
52235  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52236  uint32_t __ret;
52237  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__rev0);
52238  return __ret;
52239 }
52240 #endif
52241 
52242 #ifdef __LITTLE_ENDIAN__
52243 __ai uint16_t vminv_u16(uint16x4_t __p0) {
52244  uint16_t __ret;
52245  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__p0);
52246  return __ret;
52247 }
52248 #else
52249 __ai uint16_t vminv_u16(uint16x4_t __p0) {
52250  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52251  uint16_t __ret;
52252  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__rev0);
52253  return __ret;
52254 }
52255 #endif
52256 
52257 #ifdef __LITTLE_ENDIAN__
52258 __ai int8_t vminv_s8(int8x8_t __p0) {
52259  int8_t __ret;
52260  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__p0);
52261  return __ret;
52262 }
52263 #else
52264 __ai int8_t vminv_s8(int8x8_t __p0) {
52265  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52266  int8_t __ret;
52267  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__rev0);
52268  return __ret;
52269 }
52270 #endif
52271 
52272 #ifdef __LITTLE_ENDIAN__
52273 __ai float32_t vminv_f32(float32x2_t __p0) {
52274  float32_t __ret;
52275  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__p0);
52276  return __ret;
52277 }
52278 #else
52279 __ai float32_t vminv_f32(float32x2_t __p0) {
52280  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52281  float32_t __ret;
52282  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__rev0);
52283  return __ret;
52284 }
52285 #endif
52286 
52287 #ifdef __LITTLE_ENDIAN__
52288 __ai int32_t vminv_s32(int32x2_t __p0) {
52289  int32_t __ret;
52290  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__p0);
52291  return __ret;
52292 }
52293 #else
52294 __ai int32_t vminv_s32(int32x2_t __p0) {
52295  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52296  int32_t __ret;
52297  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__rev0);
52298  return __ret;
52299 }
52300 #endif
52301 
52302 #ifdef __LITTLE_ENDIAN__
52303 __ai int16_t vminv_s16(int16x4_t __p0) {
52304  int16_t __ret;
52305  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__p0);
52306  return __ret;
52307 }
52308 #else
52309 __ai int16_t vminv_s16(int16x4_t __p0) {
52310  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52311  int16_t __ret;
52312  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__rev0);
52313  return __ret;
52314 }
52315 #endif
52316 
52317 #ifdef __LITTLE_ENDIAN__
52318 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52319  float64x2_t __ret;
52320  __ret = __p0 + __p1 * __p2;
52321  return __ret;
52322 }
52323 #else
52324 __ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52325  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52326  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52327  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52328  float64x2_t __ret;
52329  __ret = __rev0 + __rev1 * __rev2;
52330  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52331  return __ret;
52332 }
52333 #endif
52334 
52335 #ifdef __LITTLE_ENDIAN__
52336 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52337  float64x1_t __ret;
52338  __ret = __p0 + __p1 * __p2;
52339  return __ret;
52340 }
52341 #else
52342 __ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52343  float64x1_t __ret;
52344  __ret = __p0 + __p1 * __p2;
52345  return __ret;
52346 }
52347 #endif
52348 
52349 #ifdef __LITTLE_ENDIAN__
52350 #define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52351  uint32x4_t __s0 = __p0; \
52352  uint32x4_t __s1 = __p1; \
52353  uint32x4_t __s2 = __p2; \
52354  uint32x4_t __ret; \
52355  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52356  __ret; \
52357 })
52358 #else
52359 #define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52360  uint32x4_t __s0 = __p0; \
52361  uint32x4_t __s1 = __p1; \
52362  uint32x4_t __s2 = __p2; \
52363  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52364  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52365  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52366  uint32x4_t __ret; \
52367  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52368  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52369  __ret; \
52370 })
52371 #endif
52372 
52373 #ifdef __LITTLE_ENDIAN__
52374 #define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52375  uint16x8_t __s0 = __p0; \
52376  uint16x8_t __s1 = __p1; \
52377  uint16x8_t __s2 = __p2; \
52378  uint16x8_t __ret; \
52379  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52380  __ret; \
52381 })
52382 #else
52383 #define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52384  uint16x8_t __s0 = __p0; \
52385  uint16x8_t __s1 = __p1; \
52386  uint16x8_t __s2 = __p2; \
52387  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52388  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52389  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52390  uint16x8_t __ret; \
52391  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52392  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52393  __ret; \
52394 })
52395 #endif
52396 
52397 #ifdef __LITTLE_ENDIAN__
52398 #define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52399  float32x4_t __s0 = __p0; \
52400  float32x4_t __s1 = __p1; \
52401  float32x4_t __s2 = __p2; \
52402  float32x4_t __ret; \
52403  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52404  __ret; \
52405 })
52406 #else
52407 #define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52408  float32x4_t __s0 = __p0; \
52409  float32x4_t __s1 = __p1; \
52410  float32x4_t __s2 = __p2; \
52411  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52412  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52413  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52414  float32x4_t __ret; \
52415  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52416  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52417  __ret; \
52418 })
52419 #endif
52420 
52421 #ifdef __LITTLE_ENDIAN__
52422 #define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52423  int32x4_t __s0 = __p0; \
52424  int32x4_t __s1 = __p1; \
52425  int32x4_t __s2 = __p2; \
52426  int32x4_t __ret; \
52427  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52428  __ret; \
52429 })
52430 #else
52431 #define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52432  int32x4_t __s0 = __p0; \
52433  int32x4_t __s1 = __p1; \
52434  int32x4_t __s2 = __p2; \
52435  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52436  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52437  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52438  int32x4_t __ret; \
52439  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52441  __ret; \
52442 })
52443 #endif
52444 
52445 #ifdef __LITTLE_ENDIAN__
52446 #define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52447  int16x8_t __s0 = __p0; \
52448  int16x8_t __s1 = __p1; \
52449  int16x8_t __s2 = __p2; \
52450  int16x8_t __ret; \
52451  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52452  __ret; \
52453 })
52454 #else
52455 #define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52456  int16x8_t __s0 = __p0; \
52457  int16x8_t __s1 = __p1; \
52458  int16x8_t __s2 = __p2; \
52459  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52460  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52461  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52462  int16x8_t __ret; \
52463  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52464  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52465  __ret; \
52466 })
52467 #endif
52468 
52469 #ifdef __LITTLE_ENDIAN__
52470 #define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52471  uint32x2_t __s0 = __p0; \
52472  uint32x2_t __s1 = __p1; \
52473  uint32x4_t __s2 = __p2; \
52474  uint32x2_t __ret; \
52475  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52476  __ret; \
52477 })
52478 #else
52479 #define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52480  uint32x2_t __s0 = __p0; \
52481  uint32x2_t __s1 = __p1; \
52482  uint32x4_t __s2 = __p2; \
52483  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52484  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52485  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52486  uint32x2_t __ret; \
52487  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52488  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52489  __ret; \
52490 })
52491 #endif
52492 
52493 #ifdef __LITTLE_ENDIAN__
52494 #define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52495  uint16x4_t __s0 = __p0; \
52496  uint16x4_t __s1 = __p1; \
52497  uint16x8_t __s2 = __p2; \
52498  uint16x4_t __ret; \
52499  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52500  __ret; \
52501 })
52502 #else
52503 #define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52504  uint16x4_t __s0 = __p0; \
52505  uint16x4_t __s1 = __p1; \
52506  uint16x8_t __s2 = __p2; \
52507  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52508  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52509  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52510  uint16x4_t __ret; \
52511  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52512  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52513  __ret; \
52514 })
52515 #endif
52516 
52517 #ifdef __LITTLE_ENDIAN__
52518 #define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52519  float32x2_t __s0 = __p0; \
52520  float32x2_t __s1 = __p1; \
52521  float32x4_t __s2 = __p2; \
52522  float32x2_t __ret; \
52523  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52524  __ret; \
52525 })
52526 #else
52527 #define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52528  float32x2_t __s0 = __p0; \
52529  float32x2_t __s1 = __p1; \
52530  float32x4_t __s2 = __p2; \
52531  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52532  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52533  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52534  float32x2_t __ret; \
52535  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52536  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52537  __ret; \
52538 })
52539 #endif
52540 
52541 #ifdef __LITTLE_ENDIAN__
52542 #define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52543  int32x2_t __s0 = __p0; \
52544  int32x2_t __s1 = __p1; \
52545  int32x4_t __s2 = __p2; \
52546  int32x2_t __ret; \
52547  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52548  __ret; \
52549 })
52550 #else
52551 #define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52552  int32x2_t __s0 = __p0; \
52553  int32x2_t __s1 = __p1; \
52554  int32x4_t __s2 = __p2; \
52555  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52556  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52557  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52558  int32x2_t __ret; \
52559  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52560  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52561  __ret; \
52562 })
52563 #endif
52564 
52565 #ifdef __LITTLE_ENDIAN__
52566 #define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52567  int16x4_t __s0 = __p0; \
52568  int16x4_t __s1 = __p1; \
52569  int16x8_t __s2 = __p2; \
52570  int16x4_t __ret; \
52571  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52572  __ret; \
52573 })
52574 #else
52575 #define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52576  int16x4_t __s0 = __p0; \
52577  int16x4_t __s1 = __p1; \
52578  int16x8_t __s2 = __p2; \
52579  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52580  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52581  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52582  int16x4_t __ret; \
52583  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52584  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52585  __ret; \
52586 })
52587 #endif
52588 
52589 #ifdef __LITTLE_ENDIAN__
52590 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52591  float64x2_t __ret;
52592  __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
52593  return __ret;
52594 }
52595 #else
52596 __ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52597  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52598  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52599  float64x2_t __ret;
52600  __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
52601  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52602  return __ret;
52603 }
52604 #endif
52605 
52606 #ifdef __LITTLE_ENDIAN__
52607 #define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52608  uint64x2_t __s0 = __p0; \
52609  uint32x4_t __s1 = __p1; \
52610  uint32x2_t __s2 = __p2; \
52611  uint64x2_t __ret; \
52612  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52613  __ret; \
52614 })
52615 #else
52616 #define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52617  uint64x2_t __s0 = __p0; \
52618  uint32x4_t __s1 = __p1; \
52619  uint32x2_t __s2 = __p2; \
52620  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52621  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52622  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52623  uint64x2_t __ret; \
52624  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52625  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52626  __ret; \
52627 })
52628 #endif
52629 
52630 #ifdef __LITTLE_ENDIAN__
52631 #define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52632  uint32x4_t __s0 = __p0; \
52633  uint16x8_t __s1 = __p1; \
52634  uint16x4_t __s2 = __p2; \
52635  uint32x4_t __ret; \
52636  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52637  __ret; \
52638 })
52639 #else
52640 #define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52641  uint32x4_t __s0 = __p0; \
52642  uint16x8_t __s1 = __p1; \
52643  uint16x4_t __s2 = __p2; \
52644  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52645  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52646  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52647  uint32x4_t __ret; \
52648  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52649  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52650  __ret; \
52651 })
52652 #endif
52653 
52654 #ifdef __LITTLE_ENDIAN__
52655 #define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52656  int64x2_t __s0 = __p0; \
52657  int32x4_t __s1 = __p1; \
52658  int32x2_t __s2 = __p2; \
52659  int64x2_t __ret; \
52660  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52661  __ret; \
52662 })
52663 #else
52664 #define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52665  int64x2_t __s0 = __p0; \
52666  int32x4_t __s1 = __p1; \
52667  int32x2_t __s2 = __p2; \
52668  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52669  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52670  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52671  int64x2_t __ret; \
52672  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52673  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52674  __ret; \
52675 })
52676 #endif
52677 
52678 #ifdef __LITTLE_ENDIAN__
52679 #define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52680  int32x4_t __s0 = __p0; \
52681  int16x8_t __s1 = __p1; \
52682  int16x4_t __s2 = __p2; \
52683  int32x4_t __ret; \
52684  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52685  __ret; \
52686 })
52687 #else
52688 #define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52689  int32x4_t __s0 = __p0; \
52690  int16x8_t __s1 = __p1; \
52691  int16x4_t __s2 = __p2; \
52692  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52693  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52694  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52695  int32x4_t __ret; \
52696  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52698  __ret; \
52699 })
52700 #endif
52701 
52702 #ifdef __LITTLE_ENDIAN__
52703 #define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52704  uint64x2_t __s0 = __p0; \
52705  uint32x4_t __s1 = __p1; \
52706  uint32x4_t __s2 = __p2; \
52707  uint64x2_t __ret; \
52708  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52709  __ret; \
52710 })
52711 #else
52712 #define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52713  uint64x2_t __s0 = __p0; \
52714  uint32x4_t __s1 = __p1; \
52715  uint32x4_t __s2 = __p2; \
52716  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52717  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52718  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52719  uint64x2_t __ret; \
52720  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52722  __ret; \
52723 })
52724 #endif
52725 
52726 #ifdef __LITTLE_ENDIAN__
52727 #define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52728  uint32x4_t __s0 = __p0; \
52729  uint16x8_t __s1 = __p1; \
52730  uint16x8_t __s2 = __p2; \
52731  uint32x4_t __ret; \
52732  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52733  __ret; \
52734 })
52735 #else
52736 #define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52737  uint32x4_t __s0 = __p0; \
52738  uint16x8_t __s1 = __p1; \
52739  uint16x8_t __s2 = __p2; \
52740  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52741  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52742  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52743  uint32x4_t __ret; \
52744  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52745  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52746  __ret; \
52747 })
52748 #endif
52749 
52750 #ifdef __LITTLE_ENDIAN__
52751 #define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52752  int64x2_t __s0 = __p0; \
52753  int32x4_t __s1 = __p1; \
52754  int32x4_t __s2 = __p2; \
52755  int64x2_t __ret; \
52756  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52757  __ret; \
52758 })
52759 #else
52760 #define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52761  int64x2_t __s0 = __p0; \
52762  int32x4_t __s1 = __p1; \
52763  int32x4_t __s2 = __p2; \
52764  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52765  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52766  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52767  int64x2_t __ret; \
52768  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52770  __ret; \
52771 })
52772 #endif
52773 
52774 #ifdef __LITTLE_ENDIAN__
52775 #define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52776  int32x4_t __s0 = __p0; \
52777  int16x8_t __s1 = __p1; \
52778  int16x8_t __s2 = __p2; \
52779  int32x4_t __ret; \
52780  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52781  __ret; \
52782 })
52783 #else
52784 #define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52785  int32x4_t __s0 = __p0; \
52786  int16x8_t __s1 = __p1; \
52787  int16x8_t __s2 = __p2; \
52788  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52789  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52790  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52791  int32x4_t __ret; \
52792  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52794  __ret; \
52795 })
52796 #endif
52797 
52798 #ifdef __LITTLE_ENDIAN__
52799 #define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52800  uint64x2_t __s0 = __p0; \
52801  uint32x2_t __s1 = __p1; \
52802  uint32x4_t __s2 = __p2; \
52803  uint64x2_t __ret; \
52804  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52805  __ret; \
52806 })
52807 #else
52808 #define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52809  uint64x2_t __s0 = __p0; \
52810  uint32x2_t __s1 = __p1; \
52811  uint32x4_t __s2 = __p2; \
52812  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52813  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52814  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52815  uint64x2_t __ret; \
52816  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52818  __ret; \
52819 })
52820 #endif
52821 
52822 #ifdef __LITTLE_ENDIAN__
52823 #define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52824  uint32x4_t __s0 = __p0; \
52825  uint16x4_t __s1 = __p1; \
52826  uint16x8_t __s2 = __p2; \
52827  uint32x4_t __ret; \
52828  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52829  __ret; \
52830 })
52831 #else
52832 #define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52833  uint32x4_t __s0 = __p0; \
52834  uint16x4_t __s1 = __p1; \
52835  uint16x8_t __s2 = __p2; \
52836  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52837  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52838  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52839  uint32x4_t __ret; \
52840  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52841  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52842  __ret; \
52843 })
52844 #endif
52845 
52846 #ifdef __LITTLE_ENDIAN__
52847 #define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52848  int64x2_t __s0 = __p0; \
52849  int32x2_t __s1 = __p1; \
52850  int32x4_t __s2 = __p2; \
52851  int64x2_t __ret; \
52852  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52853  __ret; \
52854 })
52855 #else
52856 #define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52857  int64x2_t __s0 = __p0; \
52858  int32x2_t __s1 = __p1; \
52859  int32x4_t __s2 = __p2; \
52860  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52861  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52862  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52863  int64x2_t __ret; \
52864  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52865  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52866  __ret; \
52867 })
52868 #endif
52869 
52870 #ifdef __LITTLE_ENDIAN__
52871 #define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52872  int32x4_t __s0 = __p0; \
52873  int16x4_t __s1 = __p1; \
52874  int16x8_t __s2 = __p2; \
52875  int32x4_t __ret; \
52876  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52877  __ret; \
52878 })
52879 #else
52880 #define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52881  int32x4_t __s0 = __p0; \
52882  int16x4_t __s1 = __p1; \
52883  int16x8_t __s2 = __p2; \
52884  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52885  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52886  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52887  int32x4_t __ret; \
52888  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52890  __ret; \
52891 })
52892 #endif
52893 
52894 #ifdef __LITTLE_ENDIAN__
52895 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52896  float64x2_t __ret;
52897  __ret = __p0 - __p1 * __p2;
52898  return __ret;
52899 }
52900 #else
52901 __ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52902  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52903  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52904  float64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52905  float64x2_t __ret;
52906  __ret = __rev0 - __rev1 * __rev2;
52907  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52908  return __ret;
52909 }
52910 #endif
52911 
52912 #ifdef __LITTLE_ENDIAN__
52913 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52914  float64x1_t __ret;
52915  __ret = __p0 - __p1 * __p2;
52916  return __ret;
52917 }
52918 #else
52919 __ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52920  float64x1_t __ret;
52921  __ret = __p0 - __p1 * __p2;
52922  return __ret;
52923 }
52924 #endif
52925 
52926 #ifdef __LITTLE_ENDIAN__
52927 #define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52928  uint32x4_t __s0 = __p0; \
52929  uint32x4_t __s1 = __p1; \
52930  uint32x4_t __s2 = __p2; \
52931  uint32x4_t __ret; \
52932  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52933  __ret; \
52934 })
52935 #else
52936 #define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52937  uint32x4_t __s0 = __p0; \
52938  uint32x4_t __s1 = __p1; \
52939  uint32x4_t __s2 = __p2; \
52940  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52941  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52942  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52943  uint32x4_t __ret; \
52944  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52945  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52946  __ret; \
52947 })
52948 #endif
52949 
52950 #ifdef __LITTLE_ENDIAN__
52951 #define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52952  uint16x8_t __s0 = __p0; \
52953  uint16x8_t __s1 = __p1; \
52954  uint16x8_t __s2 = __p2; \
52955  uint16x8_t __ret; \
52956  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52957  __ret; \
52958 })
52959 #else
52960 #define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52961  uint16x8_t __s0 = __p0; \
52962  uint16x8_t __s1 = __p1; \
52963  uint16x8_t __s2 = __p2; \
52964  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52965  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52966  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52967  uint16x8_t __ret; \
52968  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52969  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52970  __ret; \
52971 })
52972 #endif
52973 
52974 #ifdef __LITTLE_ENDIAN__
52975 #define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52976  float32x4_t __s0 = __p0; \
52977  float32x4_t __s1 = __p1; \
52978  float32x4_t __s2 = __p2; \
52979  float32x4_t __ret; \
52980  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52981  __ret; \
52982 })
52983 #else
52984 #define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52985  float32x4_t __s0 = __p0; \
52986  float32x4_t __s1 = __p1; \
52987  float32x4_t __s2 = __p2; \
52988  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52989  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52990  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52991  float32x4_t __ret; \
52992  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52993  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52994  __ret; \
52995 })
52996 #endif
52997 
52998 #ifdef __LITTLE_ENDIAN__
52999 #define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53000  int32x4_t __s0 = __p0; \
53001  int32x4_t __s1 = __p1; \
53002  int32x4_t __s2 = __p2; \
53003  int32x4_t __ret; \
53004  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53005  __ret; \
53006 })
53007 #else
53008 #define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53009  int32x4_t __s0 = __p0; \
53010  int32x4_t __s1 = __p1; \
53011  int32x4_t __s2 = __p2; \
53012  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53013  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53014  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53015  int32x4_t __ret; \
53016  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53018  __ret; \
53019 })
53020 #endif
53021 
53022 #ifdef __LITTLE_ENDIAN__
53023 #define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53024  int16x8_t __s0 = __p0; \
53025  int16x8_t __s1 = __p1; \
53026  int16x8_t __s2 = __p2; \
53027  int16x8_t __ret; \
53028  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53029  __ret; \
53030 })
53031 #else
53032 #define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53033  int16x8_t __s0 = __p0; \
53034  int16x8_t __s1 = __p1; \
53035  int16x8_t __s2 = __p2; \
53036  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53037  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53038  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53039  int16x8_t __ret; \
53040  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53042  __ret; \
53043 })
53044 #endif
53045 
53046 #ifdef __LITTLE_ENDIAN__
53047 #define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53048  uint32x2_t __s0 = __p0; \
53049  uint32x2_t __s1 = __p1; \
53050  uint32x4_t __s2 = __p2; \
53051  uint32x2_t __ret; \
53052  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53053  __ret; \
53054 })
53055 #else
53056 #define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53057  uint32x2_t __s0 = __p0; \
53058  uint32x2_t __s1 = __p1; \
53059  uint32x4_t __s2 = __p2; \
53060  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53061  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53062  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53063  uint32x2_t __ret; \
53064  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53065  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53066  __ret; \
53067 })
53068 #endif
53069 
53070 #ifdef __LITTLE_ENDIAN__
53071 #define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53072  uint16x4_t __s0 = __p0; \
53073  uint16x4_t __s1 = __p1; \
53074  uint16x8_t __s2 = __p2; \
53075  uint16x4_t __ret; \
53076  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53077  __ret; \
53078 })
53079 #else
53080 #define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53081  uint16x4_t __s0 = __p0; \
53082  uint16x4_t __s1 = __p1; \
53083  uint16x8_t __s2 = __p2; \
53084  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53085  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53086  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53087  uint16x4_t __ret; \
53088  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53090  __ret; \
53091 })
53092 #endif
53093 
53094 #ifdef __LITTLE_ENDIAN__
53095 #define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53096  float32x2_t __s0 = __p0; \
53097  float32x2_t __s1 = __p1; \
53098  float32x4_t __s2 = __p2; \
53099  float32x2_t __ret; \
53100  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53101  __ret; \
53102 })
53103 #else
53104 #define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53105  float32x2_t __s0 = __p0; \
53106  float32x2_t __s1 = __p1; \
53107  float32x4_t __s2 = __p2; \
53108  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53109  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53110  float32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53111  float32x2_t __ret; \
53112  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53114  __ret; \
53115 })
53116 #endif
53117 
53118 #ifdef __LITTLE_ENDIAN__
53119 #define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53120  int32x2_t __s0 = __p0; \
53121  int32x2_t __s1 = __p1; \
53122  int32x4_t __s2 = __p2; \
53123  int32x2_t __ret; \
53124  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53125  __ret; \
53126 })
53127 #else
53128 #define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53129  int32x2_t __s0 = __p0; \
53130  int32x2_t __s1 = __p1; \
53131  int32x4_t __s2 = __p2; \
53132  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53133  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53134  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53135  int32x2_t __ret; \
53136  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53137  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53138  __ret; \
53139 })
53140 #endif
53141 
53142 #ifdef __LITTLE_ENDIAN__
53143 #define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53144  int16x4_t __s0 = __p0; \
53145  int16x4_t __s1 = __p1; \
53146  int16x8_t __s2 = __p2; \
53147  int16x4_t __ret; \
53148  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53149  __ret; \
53150 })
53151 #else
53152 #define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53153  int16x4_t __s0 = __p0; \
53154  int16x4_t __s1 = __p1; \
53155  int16x8_t __s2 = __p2; \
53156  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53157  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53158  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53159  int16x4_t __ret; \
53160  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53161  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53162  __ret; \
53163 })
53164 #endif
53165 
53166 #ifdef __LITTLE_ENDIAN__
53167 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53168  float64x2_t __ret;
53169  __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
53170  return __ret;
53171 }
53172 #else
53173 __ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53174  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53175  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53176  float64x2_t __ret;
53177  __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
53178  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53179  return __ret;
53180 }
53181 #endif
53182 
53183 #ifdef __LITTLE_ENDIAN__
53184 #define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53185  uint64x2_t __s0 = __p0; \
53186  uint32x4_t __s1 = __p1; \
53187  uint32x2_t __s2 = __p2; \
53188  uint64x2_t __ret; \
53189  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53190  __ret; \
53191 })
53192 #else
53193 #define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53194  uint64x2_t __s0 = __p0; \
53195  uint32x4_t __s1 = __p1; \
53196  uint32x2_t __s2 = __p2; \
53197  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53198  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53199  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53200  uint64x2_t __ret; \
53201  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53203  __ret; \
53204 })
53205 #endif
53206 
53207 #ifdef __LITTLE_ENDIAN__
53208 #define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53209  uint32x4_t __s0 = __p0; \
53210  uint16x8_t __s1 = __p1; \
53211  uint16x4_t __s2 = __p2; \
53212  uint32x4_t __ret; \
53213  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53214  __ret; \
53215 })
53216 #else
53217 #define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53218  uint32x4_t __s0 = __p0; \
53219  uint16x8_t __s1 = __p1; \
53220  uint16x4_t __s2 = __p2; \
53221  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53222  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53223  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53224  uint32x4_t __ret; \
53225  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53226  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53227  __ret; \
53228 })
53229 #endif
53230 
53231 #ifdef __LITTLE_ENDIAN__
53232 #define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53233  int64x2_t __s0 = __p0; \
53234  int32x4_t __s1 = __p1; \
53235  int32x2_t __s2 = __p2; \
53236  int64x2_t __ret; \
53237  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53238  __ret; \
53239 })
53240 #else
53241 #define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53242  int64x2_t __s0 = __p0; \
53243  int32x4_t __s1 = __p1; \
53244  int32x2_t __s2 = __p2; \
53245  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53246  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53247  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53248  int64x2_t __ret; \
53249  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53250  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53251  __ret; \
53252 })
53253 #endif
53254 
53255 #ifdef __LITTLE_ENDIAN__
53256 #define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53257  int32x4_t __s0 = __p0; \
53258  int16x8_t __s1 = __p1; \
53259  int16x4_t __s2 = __p2; \
53260  int32x4_t __ret; \
53261  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53262  __ret; \
53263 })
53264 #else
53265 #define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53266  int32x4_t __s0 = __p0; \
53267  int16x8_t __s1 = __p1; \
53268  int16x4_t __s2 = __p2; \
53269  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53270  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53271  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53272  int32x4_t __ret; \
53273  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53274  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53275  __ret; \
53276 })
53277 #endif
53278 
53279 #ifdef __LITTLE_ENDIAN__
53280 #define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53281  uint64x2_t __s0 = __p0; \
53282  uint32x4_t __s1 = __p1; \
53283  uint32x4_t __s2 = __p2; \
53284  uint64x2_t __ret; \
53285  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53286  __ret; \
53287 })
53288 #else
53289 #define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53290  uint64x2_t __s0 = __p0; \
53291  uint32x4_t __s1 = __p1; \
53292  uint32x4_t __s2 = __p2; \
53293  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53294  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53295  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53296  uint64x2_t __ret; \
53297  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53298  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53299  __ret; \
53300 })
53301 #endif
53302 
53303 #ifdef __LITTLE_ENDIAN__
53304 #define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53305  uint32x4_t __s0 = __p0; \
53306  uint16x8_t __s1 = __p1; \
53307  uint16x8_t __s2 = __p2; \
53308  uint32x4_t __ret; \
53309  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53310  __ret; \
53311 })
53312 #else
53313 #define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53314  uint32x4_t __s0 = __p0; \
53315  uint16x8_t __s1 = __p1; \
53316  uint16x8_t __s2 = __p2; \
53317  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53318  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53319  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53320  uint32x4_t __ret; \
53321  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53322  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53323  __ret; \
53324 })
53325 #endif
53326 
53327 #ifdef __LITTLE_ENDIAN__
53328 #define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53329  int64x2_t __s0 = __p0; \
53330  int32x4_t __s1 = __p1; \
53331  int32x4_t __s2 = __p2; \
53332  int64x2_t __ret; \
53333  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53334  __ret; \
53335 })
53336 #else
53337 #define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53338  int64x2_t __s0 = __p0; \
53339  int32x4_t __s1 = __p1; \
53340  int32x4_t __s2 = __p2; \
53341  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53342  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53343  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53344  int64x2_t __ret; \
53345  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53346  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53347  __ret; \
53348 })
53349 #endif
53350 
53351 #ifdef __LITTLE_ENDIAN__
53352 #define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53353  int32x4_t __s0 = __p0; \
53354  int16x8_t __s1 = __p1; \
53355  int16x8_t __s2 = __p2; \
53356  int32x4_t __ret; \
53357  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53358  __ret; \
53359 })
53360 #else
53361 #define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53362  int32x4_t __s0 = __p0; \
53363  int16x8_t __s1 = __p1; \
53364  int16x8_t __s2 = __p2; \
53365  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53366  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53367  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53368  int32x4_t __ret; \
53369  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53370  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53371  __ret; \
53372 })
53373 #endif
53374 
53375 #ifdef __LITTLE_ENDIAN__
53376 #define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53377  uint64x2_t __s0 = __p0; \
53378  uint32x2_t __s1 = __p1; \
53379  uint32x4_t __s2 = __p2; \
53380  uint64x2_t __ret; \
53381  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53382  __ret; \
53383 })
53384 #else
53385 #define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53386  uint64x2_t __s0 = __p0; \
53387  uint32x2_t __s1 = __p1; \
53388  uint32x4_t __s2 = __p2; \
53389  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53390  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53391  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53392  uint64x2_t __ret; \
53393  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53395  __ret; \
53396 })
53397 #endif
53398 
53399 #ifdef __LITTLE_ENDIAN__
53400 #define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53401  uint32x4_t __s0 = __p0; \
53402  uint16x4_t __s1 = __p1; \
53403  uint16x8_t __s2 = __p2; \
53404  uint32x4_t __ret; \
53405  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53406  __ret; \
53407 })
53408 #else
53409 #define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53410  uint32x4_t __s0 = __p0; \
53411  uint16x4_t __s1 = __p1; \
53412  uint16x8_t __s2 = __p2; \
53413  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53414  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53415  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53416  uint32x4_t __ret; \
53417  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53419  __ret; \
53420 })
53421 #endif
53422 
53423 #ifdef __LITTLE_ENDIAN__
53424 #define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53425  int64x2_t __s0 = __p0; \
53426  int32x2_t __s1 = __p1; \
53427  int32x4_t __s2 = __p2; \
53428  int64x2_t __ret; \
53429  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53430  __ret; \
53431 })
53432 #else
53433 #define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53434  int64x2_t __s0 = __p0; \
53435  int32x2_t __s1 = __p1; \
53436  int32x4_t __s2 = __p2; \
53437  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53438  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53439  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53440  int64x2_t __ret; \
53441  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53443  __ret; \
53444 })
53445 #endif
53446 
53447 #ifdef __LITTLE_ENDIAN__
53448 #define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53449  int32x4_t __s0 = __p0; \
53450  int16x4_t __s1 = __p1; \
53451  int16x8_t __s2 = __p2; \
53452  int32x4_t __ret; \
53453  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53454  __ret; \
53455 })
53456 #else
53457 #define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53458  int32x4_t __s0 = __p0; \
53459  int16x4_t __s1 = __p1; \
53460  int16x8_t __s2 = __p2; \
53461  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53462  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53463  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53464  int32x4_t __ret; \
53465  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53467  __ret; \
53468 })
53469 #endif
53470 
53471 #ifdef __LITTLE_ENDIAN__
53472 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53473  poly64x1_t __ret;
53474  __ret = (poly64x1_t) {__p0};
53475  return __ret;
53476 }
53477 #else
53478 __ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53479  poly64x1_t __ret;
53480  __ret = (poly64x1_t) {__p0};
53481  return __ret;
53482 }
53483 #endif
53484 
53485 #ifdef __LITTLE_ENDIAN__
53486 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53487  poly64x2_t __ret;
53488  __ret = (poly64x2_t) {__p0, __p0};
53489  return __ret;
53490 }
53491 #else
53492 __ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53493  poly64x2_t __ret;
53494  __ret = (poly64x2_t) {__p0, __p0};
53495  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53496  return __ret;
53497 }
53498 #endif
53499 
53500 #ifdef __LITTLE_ENDIAN__
53501 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
53502  float64x2_t __ret;
53503  __ret = (float64x2_t) {__p0, __p0};
53504  return __ret;
53505 }
53506 #else
53507 __ai float64x2_t vmovq_n_f64(float64_t __p0) {
53508  float64x2_t __ret;
53509  __ret = (float64x2_t) {__p0, __p0};
53510  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53511  return __ret;
53512 }
53513 #endif
53514 
53515 #ifdef __LITTLE_ENDIAN__
53516 __ai float64x1_t vmov_n_f64(float64_t __p0) {
53517  float64x1_t __ret;
53518  __ret = (float64x1_t) {__p0};
53519  return __ret;
53520 }
53521 #else
53522 __ai float64x1_t vmov_n_f64(float64_t __p0) {
53523  float64x1_t __ret;
53524  __ret = (float64x1_t) {__p0};
53525  return __ret;
53526 }
53527 #endif
53528 
53529 #ifdef __LITTLE_ENDIAN__
53530 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_112) {
53531  uint16x8_t __ret_112;
53532  uint8x8_t __a1_112 = vget_high_u8(__p0_112);
53533  __ret_112 = (uint16x8_t)(vshll_n_u8(__a1_112, 0));
53534  return __ret_112;
53535 }
53536 #else
53537 __ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_113) {
53538  uint8x16_t __rev0_113; __rev0_113 = __builtin_shufflevector(__p0_113, __p0_113, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53539  uint16x8_t __ret_113;
53540  uint8x8_t __a1_113 = __noswap_vget_high_u8(__rev0_113);
53541  __ret_113 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_113, 0));
53542  __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 7, 6, 5, 4, 3, 2, 1, 0);
53543  return __ret_113;
53544 }
53545 __ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_114) {
53546  uint16x8_t __ret_114;
53547  uint8x8_t __a1_114 = __noswap_vget_high_u8(__p0_114);
53548  __ret_114 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_114, 0));
53549  return __ret_114;
53550 }
53551 #endif
53552 
53553 #ifdef __LITTLE_ENDIAN__
53554 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_115) {
53555  uint64x2_t __ret_115;
53556  uint32x2_t __a1_115 = vget_high_u32(__p0_115);
53557  __ret_115 = (uint64x2_t)(vshll_n_u32(__a1_115, 0));
53558  return __ret_115;
53559 }
53560 #else
53561 __ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_116) {
53562  uint32x4_t __rev0_116; __rev0_116 = __builtin_shufflevector(__p0_116, __p0_116, 3, 2, 1, 0);
53563  uint64x2_t __ret_116;
53564  uint32x2_t __a1_116 = __noswap_vget_high_u32(__rev0_116);
53565  __ret_116 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_116, 0));
53566  __ret_116 = __builtin_shufflevector(__ret_116, __ret_116, 1, 0);
53567  return __ret_116;
53568 }
53569 __ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_117) {
53570  uint64x2_t __ret_117;
53571  uint32x2_t __a1_117 = __noswap_vget_high_u32(__p0_117);
53572  __ret_117 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_117, 0));
53573  return __ret_117;
53574 }
53575 #endif
53576 
53577 #ifdef __LITTLE_ENDIAN__
53578 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_118) {
53579  uint32x4_t __ret_118;
53580  uint16x4_t __a1_118 = vget_high_u16(__p0_118);
53581  __ret_118 = (uint32x4_t)(vshll_n_u16(__a1_118, 0));
53582  return __ret_118;
53583 }
53584 #else
53585 __ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_119) {
53586  uint16x8_t __rev0_119; __rev0_119 = __builtin_shufflevector(__p0_119, __p0_119, 7, 6, 5, 4, 3, 2, 1, 0);
53587  uint32x4_t __ret_119;
53588  uint16x4_t __a1_119 = __noswap_vget_high_u16(__rev0_119);
53589  __ret_119 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_119, 0));
53590  __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 3, 2, 1, 0);
53591  return __ret_119;
53592 }
53593 __ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_120) {
53594  uint32x4_t __ret_120;
53595  uint16x4_t __a1_120 = __noswap_vget_high_u16(__p0_120);
53596  __ret_120 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_120, 0));
53597  return __ret_120;
53598 }
53599 #endif
53600 
53601 #ifdef __LITTLE_ENDIAN__
53602 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_121) {
53603  int16x8_t __ret_121;
53604  int8x8_t __a1_121 = vget_high_s8(__p0_121);
53605  __ret_121 = (int16x8_t)(vshll_n_s8(__a1_121, 0));
53606  return __ret_121;
53607 }
53608 #else
53609 __ai int16x8_t vmovl_high_s8(int8x16_t __p0_122) {
53610  int8x16_t __rev0_122; __rev0_122 = __builtin_shufflevector(__p0_122, __p0_122, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53611  int16x8_t __ret_122;
53612  int8x8_t __a1_122 = __noswap_vget_high_s8(__rev0_122);
53613  __ret_122 = (int16x8_t)(__noswap_vshll_n_s8(__a1_122, 0));
53614  __ret_122 = __builtin_shufflevector(__ret_122, __ret_122, 7, 6, 5, 4, 3, 2, 1, 0);
53615  return __ret_122;
53616 }
53617 __ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_123) {
53618  int16x8_t __ret_123;
53619  int8x8_t __a1_123 = __noswap_vget_high_s8(__p0_123);
53620  __ret_123 = (int16x8_t)(__noswap_vshll_n_s8(__a1_123, 0));
53621  return __ret_123;
53622 }
53623 #endif
53624 
53625 #ifdef __LITTLE_ENDIAN__
53626 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_124) {
53627  int64x2_t __ret_124;
53628  int32x2_t __a1_124 = vget_high_s32(__p0_124);
53629  __ret_124 = (int64x2_t)(vshll_n_s32(__a1_124, 0));
53630  return __ret_124;
53631 }
53632 #else
53633 __ai int64x2_t vmovl_high_s32(int32x4_t __p0_125) {
53634  int32x4_t __rev0_125; __rev0_125 = __builtin_shufflevector(__p0_125, __p0_125, 3, 2, 1, 0);
53635  int64x2_t __ret_125;
53636  int32x2_t __a1_125 = __noswap_vget_high_s32(__rev0_125);
53637  __ret_125 = (int64x2_t)(__noswap_vshll_n_s32(__a1_125, 0));
53638  __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 1, 0);
53639  return __ret_125;
53640 }
53641 __ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_126) {
53642  int64x2_t __ret_126;
53643  int32x2_t __a1_126 = __noswap_vget_high_s32(__p0_126);
53644  __ret_126 = (int64x2_t)(__noswap_vshll_n_s32(__a1_126, 0));
53645  return __ret_126;
53646 }
53647 #endif
53648 
53649 #ifdef __LITTLE_ENDIAN__
53650 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_127) {
53651  int32x4_t __ret_127;
53652  int16x4_t __a1_127 = vget_high_s16(__p0_127);
53653  __ret_127 = (int32x4_t)(vshll_n_s16(__a1_127, 0));
53654  return __ret_127;
53655 }
53656 #else
53657 __ai int32x4_t vmovl_high_s16(int16x8_t __p0_128) {
53658  int16x8_t __rev0_128; __rev0_128 = __builtin_shufflevector(__p0_128, __p0_128, 7, 6, 5, 4, 3, 2, 1, 0);
53659  int32x4_t __ret_128;
53660  int16x4_t __a1_128 = __noswap_vget_high_s16(__rev0_128);
53661  __ret_128 = (int32x4_t)(__noswap_vshll_n_s16(__a1_128, 0));
53662  __ret_128 = __builtin_shufflevector(__ret_128, __ret_128, 3, 2, 1, 0);
53663  return __ret_128;
53664 }
53665 __ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_129) {
53666  int32x4_t __ret_129;
53667  int16x4_t __a1_129 = __noswap_vget_high_s16(__p0_129);
53668  __ret_129 = (int32x4_t)(__noswap_vshll_n_s16(__a1_129, 0));
53669  return __ret_129;
53670 }
53671 #endif
53672 
53673 #ifdef __LITTLE_ENDIAN__
53674 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53675  uint16x8_t __ret;
53676  __ret = vcombine_u16(__p0, vmovn_u32(__p1));
53677  return __ret;
53678 }
53679 #else
53680 __ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53681  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53682  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53683  uint16x8_t __ret;
53684  __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
53685  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53686  return __ret;
53687 }
53688 #endif
53689 
53690 #ifdef __LITTLE_ENDIAN__
53691 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53692  uint32x4_t __ret;
53693  __ret = vcombine_u32(__p0, vmovn_u64(__p1));
53694  return __ret;
53695 }
53696 #else
53697 __ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53698  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53699  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53700  uint32x4_t __ret;
53701  __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
53702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53703  return __ret;
53704 }
53705 #endif
53706 
53707 #ifdef __LITTLE_ENDIAN__
53708 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53709  uint8x16_t __ret;
53710  __ret = vcombine_u8(__p0, vmovn_u16(__p1));
53711  return __ret;
53712 }
53713 #else
53714 __ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53715  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53716  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53717  uint8x16_t __ret;
53718  __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
53719  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53720  return __ret;
53721 }
53722 #endif
53723 
53724 #ifdef __LITTLE_ENDIAN__
53725 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53726  int16x8_t __ret;
53727  __ret = vcombine_s16(__p0, vmovn_s32(__p1));
53728  return __ret;
53729 }
53730 #else
53731 __ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53732  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53733  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53734  int16x8_t __ret;
53735  __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
53736  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53737  return __ret;
53738 }
53739 #endif
53740 
53741 #ifdef __LITTLE_ENDIAN__
53742 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53743  int32x4_t __ret;
53744  __ret = vcombine_s32(__p0, vmovn_s64(__p1));
53745  return __ret;
53746 }
53747 #else
53748 __ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53749  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53750  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53751  int32x4_t __ret;
53752  __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
53753  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53754  return __ret;
53755 }
53756 #endif
53757 
53758 #ifdef __LITTLE_ENDIAN__
53759 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53760  int8x16_t __ret;
53761  __ret = vcombine_s8(__p0, vmovn_s16(__p1));
53762  return __ret;
53763 }
53764 #else
53765 __ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53766  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53767  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53768  int8x16_t __ret;
53769  __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
53770  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53771  return __ret;
53772 }
53773 #endif
53774 
53775 #ifdef __LITTLE_ENDIAN__
53776 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53777  float64x2_t __ret;
53778  __ret = __p0 * __p1;
53779  return __ret;
53780 }
53781 #else
53782 __ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53783  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53784  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53785  float64x2_t __ret;
53786  __ret = __rev0 * __rev1;
53787  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53788  return __ret;
53789 }
53790 #endif
53791 
53792 #ifdef __LITTLE_ENDIAN__
53793 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53794  float64x1_t __ret;
53795  __ret = __p0 * __p1;
53796  return __ret;
53797 }
53798 #else
53799 __ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53800  float64x1_t __ret;
53801  __ret = __p0 * __p1;
53802  return __ret;
53803 }
53804 #endif
53805 
53806 #ifdef __LITTLE_ENDIAN__
53807 #define vmuld_lane_f64(__p0_130, __p1_130, __p2_130) __extension__ ({ \
53808  float64_t __s0_130 = __p0_130; \
53809  float64x1_t __s1_130 = __p1_130; \
53810  float64_t __ret_130; \
53811  __ret_130 = __s0_130 * vget_lane_f64(__s1_130, __p2_130); \
53812  __ret_130; \
53813 })
53814 #else
53815 #define vmuld_lane_f64(__p0_131, __p1_131, __p2_131) __extension__ ({ \
53816  float64_t __s0_131 = __p0_131; \
53817  float64x1_t __s1_131 = __p1_131; \
53818  float64_t __ret_131; \
53819  __ret_131 = __s0_131 * __noswap_vget_lane_f64(__s1_131, __p2_131); \
53820  __ret_131; \
53821 })
53822 #endif
53823 
53824 #ifdef __LITTLE_ENDIAN__
53825 #define vmuls_lane_f32(__p0_132, __p1_132, __p2_132) __extension__ ({ \
53826  float32_t __s0_132 = __p0_132; \
53827  float32x2_t __s1_132 = __p1_132; \
53828  float32_t __ret_132; \
53829  __ret_132 = __s0_132 * vget_lane_f32(__s1_132, __p2_132); \
53830  __ret_132; \
53831 })
53832 #else
53833 #define vmuls_lane_f32(__p0_133, __p1_133, __p2_133) __extension__ ({ \
53834  float32_t __s0_133 = __p0_133; \
53835  float32x2_t __s1_133 = __p1_133; \
53836  float32x2_t __rev1_133; __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 1, 0); \
53837  float32_t __ret_133; \
53838  __ret_133 = __s0_133 * __noswap_vget_lane_f32(__rev1_133, __p2_133); \
53839  __ret_133; \
53840 })
53841 #endif
53842 
53843 #ifdef __LITTLE_ENDIAN__
53844 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53845  float64x1_t __s0 = __p0; \
53846  float64x1_t __s1 = __p1; \
53847  float64x1_t __ret; \
53848  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53849  __ret; \
53850 })
53851 #else
53852 #define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53853  float64x1_t __s0 = __p0; \
53854  float64x1_t __s1 = __p1; \
53855  float64x1_t __ret; \
53856  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53857  __ret; \
53858 })
53859 #endif
53860 
53861 #ifdef __LITTLE_ENDIAN__
53862 #define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53863  float64x2_t __s0 = __p0; \
53864  float64x1_t __s1 = __p1; \
53865  float64x2_t __ret; \
53866  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53867  __ret; \
53868 })
53869 #else
53870 #define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53871  float64x2_t __s0 = __p0; \
53872  float64x1_t __s1 = __p1; \
53873  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53874  float64x2_t __ret; \
53875  __ret = __rev0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53876  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53877  __ret; \
53878 })
53879 #endif
53880 
53881 #ifdef __LITTLE_ENDIAN__
53882 #define vmuld_laneq_f64(__p0_134, __p1_134, __p2_134) __extension__ ({ \
53883  float64_t __s0_134 = __p0_134; \
53884  float64x2_t __s1_134 = __p1_134; \
53885  float64_t __ret_134; \
53886  __ret_134 = __s0_134 * vgetq_lane_f64(__s1_134, __p2_134); \
53887  __ret_134; \
53888 })
53889 #else
53890 #define vmuld_laneq_f64(__p0_135, __p1_135, __p2_135) __extension__ ({ \
53891  float64_t __s0_135 = __p0_135; \
53892  float64x2_t __s1_135 = __p1_135; \
53893  float64x2_t __rev1_135; __rev1_135 = __builtin_shufflevector(__s1_135, __s1_135, 1, 0); \
53894  float64_t __ret_135; \
53895  __ret_135 = __s0_135 * __noswap_vgetq_lane_f64(__rev1_135, __p2_135); \
53896  __ret_135; \
53897 })
53898 #endif
53899 
53900 #ifdef __LITTLE_ENDIAN__
53901 #define vmuls_laneq_f32(__p0_136, __p1_136, __p2_136) __extension__ ({ \
53902  float32_t __s0_136 = __p0_136; \
53903  float32x4_t __s1_136 = __p1_136; \
53904  float32_t __ret_136; \
53905  __ret_136 = __s0_136 * vgetq_lane_f32(__s1_136, __p2_136); \
53906  __ret_136; \
53907 })
53908 #else
53909 #define vmuls_laneq_f32(__p0_137, __p1_137, __p2_137) __extension__ ({ \
53910  float32_t __s0_137 = __p0_137; \
53911  float32x4_t __s1_137 = __p1_137; \
53912  float32x4_t __rev1_137; __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 3, 2, 1, 0); \
53913  float32_t __ret_137; \
53914  __ret_137 = __s0_137 * __noswap_vgetq_lane_f32(__rev1_137, __p2_137); \
53915  __ret_137; \
53916 })
53917 #endif
53918 
53919 #ifdef __LITTLE_ENDIAN__
53920 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53921  float64x1_t __s0 = __p0; \
53922  float64x2_t __s1 = __p1; \
53923  float64x1_t __ret; \
53924  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
53925  __ret; \
53926 })
53927 #else
53928 #define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53929  float64x1_t __s0 = __p0; \
53930  float64x2_t __s1 = __p1; \
53931  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53932  float64x1_t __ret; \
53933  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
53934  __ret; \
53935 })
53936 #endif
53937 
53938 #ifdef __LITTLE_ENDIAN__
53939 #define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53940  uint32x4_t __s0 = __p0; \
53941  uint32x4_t __s1 = __p1; \
53942  uint32x4_t __ret; \
53943  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53944  __ret; \
53945 })
53946 #else
53947 #define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53948  uint32x4_t __s0 = __p0; \
53949  uint32x4_t __s1 = __p1; \
53950  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53951  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53952  uint32x4_t __ret; \
53953  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53954  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53955  __ret; \
53956 })
53957 #endif
53958 
53959 #ifdef __LITTLE_ENDIAN__
53960 #define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53961  uint16x8_t __s0 = __p0; \
53962  uint16x8_t __s1 = __p1; \
53963  uint16x8_t __ret; \
53964  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53965  __ret; \
53966 })
53967 #else
53968 #define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53969  uint16x8_t __s0 = __p0; \
53970  uint16x8_t __s1 = __p1; \
53971  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53972  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53973  uint16x8_t __ret; \
53974  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53975  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53976  __ret; \
53977 })
53978 #endif
53979 
53980 #ifdef __LITTLE_ENDIAN__
53981 #define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53982  float64x2_t __s0 = __p0; \
53983  float64x2_t __s1 = __p1; \
53984  float64x2_t __ret; \
53985  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53986  __ret; \
53987 })
53988 #else
53989 #define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53990  float64x2_t __s0 = __p0; \
53991  float64x2_t __s1 = __p1; \
53992  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53993  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53994  float64x2_t __ret; \
53995  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53996  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53997  __ret; \
53998 })
53999 #endif
54000 
54001 #ifdef __LITTLE_ENDIAN__
54002 #define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54003  float32x4_t __s0 = __p0; \
54004  float32x4_t __s1 = __p1; \
54005  float32x4_t __ret; \
54006  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54007  __ret; \
54008 })
54009 #else
54010 #define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54011  float32x4_t __s0 = __p0; \
54012  float32x4_t __s1 = __p1; \
54013  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54014  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54015  float32x4_t __ret; \
54016  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54018  __ret; \
54019 })
54020 #endif
54021 
54022 #ifdef __LITTLE_ENDIAN__
54023 #define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54024  int32x4_t __s0 = __p0; \
54025  int32x4_t __s1 = __p1; \
54026  int32x4_t __ret; \
54027  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54028  __ret; \
54029 })
54030 #else
54031 #define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54032  int32x4_t __s0 = __p0; \
54033  int32x4_t __s1 = __p1; \
54034  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54035  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54036  int32x4_t __ret; \
54037  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54038  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54039  __ret; \
54040 })
54041 #endif
54042 
54043 #ifdef __LITTLE_ENDIAN__
54044 #define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54045  int16x8_t __s0 = __p0; \
54046  int16x8_t __s1 = __p1; \
54047  int16x8_t __ret; \
54048  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54049  __ret; \
54050 })
54051 #else
54052 #define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54053  int16x8_t __s0 = __p0; \
54054  int16x8_t __s1 = __p1; \
54055  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54056  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54057  int16x8_t __ret; \
54058  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
54060  __ret; \
54061 })
54062 #endif
54063 
54064 #ifdef __LITTLE_ENDIAN__
54065 #define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54066  uint32x2_t __s0 = __p0; \
54067  uint32x4_t __s1 = __p1; \
54068  uint32x2_t __ret; \
54069  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54070  __ret; \
54071 })
54072 #else
54073 #define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54074  uint32x2_t __s0 = __p0; \
54075  uint32x4_t __s1 = __p1; \
54076  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54077  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54078  uint32x2_t __ret; \
54079  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54080  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54081  __ret; \
54082 })
54083 #endif
54084 
54085 #ifdef __LITTLE_ENDIAN__
54086 #define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54087  uint16x4_t __s0 = __p0; \
54088  uint16x8_t __s1 = __p1; \
54089  uint16x4_t __ret; \
54090  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54091  __ret; \
54092 })
54093 #else
54094 #define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54095  uint16x4_t __s0 = __p0; \
54096  uint16x8_t __s1 = __p1; \
54097  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54098  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54099  uint16x4_t __ret; \
54100  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54101  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54102  __ret; \
54103 })
54104 #endif
54105 
54106 #ifdef __LITTLE_ENDIAN__
54107 #define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54108  float32x2_t __s0 = __p0; \
54109  float32x4_t __s1 = __p1; \
54110  float32x2_t __ret; \
54111  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54112  __ret; \
54113 })
54114 #else
54115 #define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54116  float32x2_t __s0 = __p0; \
54117  float32x4_t __s1 = __p1; \
54118  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54119  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54120  float32x2_t __ret; \
54121  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54123  __ret; \
54124 })
54125 #endif
54126 
54127 #ifdef __LITTLE_ENDIAN__
54128 #define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54129  int32x2_t __s0 = __p0; \
54130  int32x4_t __s1 = __p1; \
54131  int32x2_t __ret; \
54132  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54133  __ret; \
54134 })
54135 #else
54136 #define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54137  int32x2_t __s0 = __p0; \
54138  int32x4_t __s1 = __p1; \
54139  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54140  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54141  int32x2_t __ret; \
54142  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54144  __ret; \
54145 })
54146 #endif
54147 
54148 #ifdef __LITTLE_ENDIAN__
54149 #define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54150  int16x4_t __s0 = __p0; \
54151  int16x8_t __s1 = __p1; \
54152  int16x4_t __ret; \
54153  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54154  __ret; \
54155 })
54156 #else
54157 #define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54158  int16x4_t __s0 = __p0; \
54159  int16x8_t __s1 = __p1; \
54160  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54161  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54162  int16x4_t __ret; \
54163  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54164  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54165  __ret; \
54166 })
54167 #endif
54168 
54169 #ifdef __LITTLE_ENDIAN__
54170 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54171  float64x1_t __ret;
54172  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54173  return __ret;
54174 }
54175 #else
54176 __ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54177  float64x1_t __ret;
54178  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54179  return __ret;
54180 }
54181 #endif
54182 
54183 #ifdef __LITTLE_ENDIAN__
54184 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54185  float64x2_t __ret;
54186  __ret = __p0 * (float64x2_t) {__p1, __p1};
54187  return __ret;
54188 }
54189 #else
54190 __ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54191  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54192  float64x2_t __ret;
54193  __ret = __rev0 * (float64x2_t) {__p1, __p1};
54194  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54195  return __ret;
54196 }
54197 #endif
54198 
54199 #ifdef __LITTLE_ENDIAN__
54200 __ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54201  poly128_t __ret;
54202  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54203  return __ret;
54204 }
54205 #else
54206 __ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54207  poly128_t __ret;
54208  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54209  return __ret;
54210 }
54211 __ai poly128_t __noswap_vmull_p64(poly64_t __p0, poly64_t __p1) {
54212  poly128_t __ret;
54213  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54214  return __ret;
54215 }
54216 #endif
54217 
54218 #ifdef __LITTLE_ENDIAN__
54219 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54220  poly16x8_t __ret;
54221  __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
54222  return __ret;
54223 }
54224 #else
54225 __ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54226  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54227  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54228  poly16x8_t __ret;
54230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54231  return __ret;
54232 }
54233 #endif
54234 
54235 #ifdef __LITTLE_ENDIAN__
54236 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54237  uint16x8_t __ret;
54238  __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
54239  return __ret;
54240 }
54241 #else
54242 __ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54243  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54244  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54245  uint16x8_t __ret;
54247  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54248  return __ret;
54249 }
54250 #endif
54251 
54252 #ifdef __LITTLE_ENDIAN__
54253 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54254  uint64x2_t __ret;
54255  __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
54256  return __ret;
54257 }
54258 #else
54259 __ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54260  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54261  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54262  uint64x2_t __ret;
54264  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54265  return __ret;
54266 }
54267 #endif
54268 
54269 #ifdef __LITTLE_ENDIAN__
54270 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54271  uint32x4_t __ret;
54272  __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
54273  return __ret;
54274 }
54275 #else
54276 __ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54277  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54278  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54279  uint32x4_t __ret;
54281  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54282  return __ret;
54283 }
54284 #endif
54285 
54286 #ifdef __LITTLE_ENDIAN__
54287 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54288  int16x8_t __ret;
54289  __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
54290  return __ret;
54291 }
54292 #else
54293 __ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54294  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54295  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54296  int16x8_t __ret;
54298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54299  return __ret;
54300 }
54301 #endif
54302 
54303 #ifdef __LITTLE_ENDIAN__
54304 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54305  int64x2_t __ret;
54306  __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
54307  return __ret;
54308 }
54309 #else
54310 __ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54311  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54312  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54313  int64x2_t __ret;
54315  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54316  return __ret;
54317 }
54318 #endif
54319 
54320 #ifdef __LITTLE_ENDIAN__
54321 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54322  int32x4_t __ret;
54323  __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
54324  return __ret;
54325 }
54326 #else
54327 __ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54328  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54329  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54330  int32x4_t __ret;
54332  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54333  return __ret;
54334 }
54335 #endif
54336 
54337 #ifdef __LITTLE_ENDIAN__
54338 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54339  poly128_t __ret;
54340  __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
54341  return __ret;
54342 }
54343 #else
54344 __ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54345  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54346  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54347  poly128_t __ret;
54348  __ret = __noswap_vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
54349  return __ret;
54350 }
54351 #endif
54352 
54353 #ifdef __LITTLE_ENDIAN__
54354 #define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54355  uint32x4_t __s0 = __p0; \
54356  uint32x2_t __s1 = __p1; \
54357  uint64x2_t __ret; \
54358  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54359  __ret; \
54360 })
54361 #else
54362 #define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54363  uint32x4_t __s0 = __p0; \
54364  uint32x2_t __s1 = __p1; \
54365  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54366  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54367  uint64x2_t __ret; \
54368  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54369  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54370  __ret; \
54371 })
54372 #endif
54373 
54374 #ifdef __LITTLE_ENDIAN__
54375 #define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54376  uint16x8_t __s0 = __p0; \
54377  uint16x4_t __s1 = __p1; \
54378  uint32x4_t __ret; \
54379  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54380  __ret; \
54381 })
54382 #else
54383 #define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54384  uint16x8_t __s0 = __p0; \
54385  uint16x4_t __s1 = __p1; \
54386  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54387  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54388  uint32x4_t __ret; \
54389  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54391  __ret; \
54392 })
54393 #endif
54394 
54395 #ifdef __LITTLE_ENDIAN__
54396 #define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54397  int32x4_t __s0 = __p0; \
54398  int32x2_t __s1 = __p1; \
54399  int64x2_t __ret; \
54400  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54401  __ret; \
54402 })
54403 #else
54404 #define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54405  int32x4_t __s0 = __p0; \
54406  int32x2_t __s1 = __p1; \
54407  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54408  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54409  int64x2_t __ret; \
54410  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54411  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54412  __ret; \
54413 })
54414 #endif
54415 
54416 #ifdef __LITTLE_ENDIAN__
54417 #define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54418  int16x8_t __s0 = __p0; \
54419  int16x4_t __s1 = __p1; \
54420  int32x4_t __ret; \
54421  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54422  __ret; \
54423 })
54424 #else
54425 #define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54426  int16x8_t __s0 = __p0; \
54427  int16x4_t __s1 = __p1; \
54428  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54429  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54430  int32x4_t __ret; \
54431  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54432  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54433  __ret; \
54434 })
54435 #endif
54436 
54437 #ifdef __LITTLE_ENDIAN__
54438 #define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54439  uint32x4_t __s0 = __p0; \
54440  uint32x4_t __s1 = __p1; \
54441  uint64x2_t __ret; \
54442  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54443  __ret; \
54444 })
54445 #else
54446 #define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54447  uint32x4_t __s0 = __p0; \
54448  uint32x4_t __s1 = __p1; \
54449  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54450  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54451  uint64x2_t __ret; \
54452  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54454  __ret; \
54455 })
54456 #endif
54457 
54458 #ifdef __LITTLE_ENDIAN__
54459 #define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54460  uint16x8_t __s0 = __p0; \
54461  uint16x8_t __s1 = __p1; \
54462  uint32x4_t __ret; \
54463  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54464  __ret; \
54465 })
54466 #else
54467 #define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54468  uint16x8_t __s0 = __p0; \
54469  uint16x8_t __s1 = __p1; \
54470  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54471  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54472  uint32x4_t __ret; \
54473  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54474  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54475  __ret; \
54476 })
54477 #endif
54478 
54479 #ifdef __LITTLE_ENDIAN__
54480 #define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54481  int32x4_t __s0 = __p0; \
54482  int32x4_t __s1 = __p1; \
54483  int64x2_t __ret; \
54484  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54485  __ret; \
54486 })
54487 #else
54488 #define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54489  int32x4_t __s0 = __p0; \
54490  int32x4_t __s1 = __p1; \
54491  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54492  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54493  int64x2_t __ret; \
54494  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54495  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54496  __ret; \
54497 })
54498 #endif
54499 
54500 #ifdef __LITTLE_ENDIAN__
54501 #define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54502  int16x8_t __s0 = __p0; \
54503  int16x8_t __s1 = __p1; \
54504  int32x4_t __ret; \
54505  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54506  __ret; \
54507 })
54508 #else
54509 #define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54510  int16x8_t __s0 = __p0; \
54511  int16x8_t __s1 = __p1; \
54512  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54513  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54514  int32x4_t __ret; \
54515  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54517  __ret; \
54518 })
54519 #endif
54520 
54521 #ifdef __LITTLE_ENDIAN__
54522 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54523  uint64x2_t __ret;
54524  __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
54525  return __ret;
54526 }
54527 #else
54528 __ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54529  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54530  uint64x2_t __ret;
54531  __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
54532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54533  return __ret;
54534 }
54535 #endif
54536 
54537 #ifdef __LITTLE_ENDIAN__
54538 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54539  uint32x4_t __ret;
54540  __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
54541  return __ret;
54542 }
54543 #else
54544 __ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54545  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54546  uint32x4_t __ret;
54547  __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
54548  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54549  return __ret;
54550 }
54551 #endif
54552 
54553 #ifdef __LITTLE_ENDIAN__
54554 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54555  int64x2_t __ret;
54556  __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
54557  return __ret;
54558 }
54559 #else
54560 __ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54561  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54562  int64x2_t __ret;
54563  __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
54564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54565  return __ret;
54566 }
54567 #endif
54568 
54569 #ifdef __LITTLE_ENDIAN__
54570 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54571  int32x4_t __ret;
54572  __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
54573  return __ret;
54574 }
54575 #else
54576 __ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54577  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54578  int32x4_t __ret;
54579  __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
54580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54581  return __ret;
54582 }
54583 #endif
54584 
54585 #ifdef __LITTLE_ENDIAN__
54586 #define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54587  uint32x2_t __s0 = __p0; \
54588  uint32x4_t __s1 = __p1; \
54589  uint64x2_t __ret; \
54590  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54591  __ret; \
54592 })
54593 #else
54594 #define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54595  uint32x2_t __s0 = __p0; \
54596  uint32x4_t __s1 = __p1; \
54597  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54598  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54599  uint64x2_t __ret; \
54600  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54601  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54602  __ret; \
54603 })
54604 #endif
54605 
54606 #ifdef __LITTLE_ENDIAN__
54607 #define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54608  uint16x4_t __s0 = __p0; \
54609  uint16x8_t __s1 = __p1; \
54610  uint32x4_t __ret; \
54611  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54612  __ret; \
54613 })
54614 #else
54615 #define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54616  uint16x4_t __s0 = __p0; \
54617  uint16x8_t __s1 = __p1; \
54618  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54619  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54620  uint32x4_t __ret; \
54621  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54623  __ret; \
54624 })
54625 #endif
54626 
54627 #ifdef __LITTLE_ENDIAN__
54628 #define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54629  int32x2_t __s0 = __p0; \
54630  int32x4_t __s1 = __p1; \
54631  int64x2_t __ret; \
54632  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54633  __ret; \
54634 })
54635 #else
54636 #define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54637  int32x2_t __s0 = __p0; \
54638  int32x4_t __s1 = __p1; \
54639  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54640  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54641  int64x2_t __ret; \
54642  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54643  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54644  __ret; \
54645 })
54646 #endif
54647 
54648 #ifdef __LITTLE_ENDIAN__
54649 #define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54650  int16x4_t __s0 = __p0; \
54651  int16x8_t __s1 = __p1; \
54652  int32x4_t __ret; \
54653  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54654  __ret; \
54655 })
54656 #else
54657 #define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54658  int16x4_t __s0 = __p0; \
54659  int16x8_t __s1 = __p1; \
54660  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54661  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54662  int32x4_t __ret; \
54663  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54665  __ret; \
54666 })
54667 #endif
54668 
54669 #ifdef __LITTLE_ENDIAN__
54670 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54671  float64x2_t __ret;
54672  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54673  return __ret;
54674 }
54675 #else
54676 __ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54677  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54678  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54679  float64x2_t __ret;
54680  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54682  return __ret;
54683 }
54684 __ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54685  float64x2_t __ret;
54686  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54687  return __ret;
54688 }
54689 #endif
54690 
54691 #ifdef __LITTLE_ENDIAN__
54692 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54693  float32x4_t __ret;
54694  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54695  return __ret;
54696 }
54697 #else
54698 __ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54699  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54700  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54701  float32x4_t __ret;
54702  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54703  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54704  return __ret;
54705 }
54706 __ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54707  float32x4_t __ret;
54708  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54709  return __ret;
54710 }
54711 #endif
54712 
54713 #ifdef __LITTLE_ENDIAN__
54714 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54715  float64x1_t __ret;
54716  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54717  return __ret;
54718 }
54719 #else
54720 __ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54721  float64x1_t __ret;
54722  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54723  return __ret;
54724 }
54725 #endif
54726 
54727 #ifdef __LITTLE_ENDIAN__
54728 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54729  float32x2_t __ret;
54730  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54731  return __ret;
54732 }
54733 #else
54734 __ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54735  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54736  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54737  float32x2_t __ret;
54738  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54739  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54740  return __ret;
54741 }
54742 __ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54743  float32x2_t __ret;
54744  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54745  return __ret;
54746 }
54747 #endif
54748 
54749 #ifdef __LITTLE_ENDIAN__
54750 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54751  float64_t __ret;
54752  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54753  return __ret;
54754 }
54755 #else
54756 __ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54757  float64_t __ret;
54758  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54759  return __ret;
54760 }
54761 __ai float64_t __noswap_vmulxd_f64(float64_t __p0, float64_t __p1) {
54762  float64_t __ret;
54763  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54764  return __ret;
54765 }
54766 #endif
54767 
54768 #ifdef __LITTLE_ENDIAN__
54769 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54770  float32_t __ret;
54771  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54772  return __ret;
54773 }
54774 #else
54775 __ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54776  float32_t __ret;
54777  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54778  return __ret;
54779 }
54780 __ai float32_t __noswap_vmulxs_f32(float32_t __p0, float32_t __p1) {
54781  float32_t __ret;
54782  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54783  return __ret;
54784 }
54785 #endif
54786 
54787 #ifdef __LITTLE_ENDIAN__
54788 #define vmulxd_lane_f64(__p0_138, __p1_138, __p2_138) __extension__ ({ \
54789  float64_t __s0_138 = __p0_138; \
54790  float64x1_t __s1_138 = __p1_138; \
54791  float64_t __ret_138; \
54792  __ret_138 = vmulxd_f64(__s0_138, vget_lane_f64(__s1_138, __p2_138)); \
54793  __ret_138; \
54794 })
54795 #else
54796 #define vmulxd_lane_f64(__p0_139, __p1_139, __p2_139) __extension__ ({ \
54797  float64_t __s0_139 = __p0_139; \
54798  float64x1_t __s1_139 = __p1_139; \
54799  float64_t __ret_139; \
54800  __ret_139 = __noswap_vmulxd_f64(__s0_139, __noswap_vget_lane_f64(__s1_139, __p2_139)); \
54801  __ret_139; \
54802 })
54803 #endif
54804 
54805 #ifdef __LITTLE_ENDIAN__
54806 #define vmulxs_lane_f32(__p0_140, __p1_140, __p2_140) __extension__ ({ \
54807  float32_t __s0_140 = __p0_140; \
54808  float32x2_t __s1_140 = __p1_140; \
54809  float32_t __ret_140; \
54810  __ret_140 = vmulxs_f32(__s0_140, vget_lane_f32(__s1_140, __p2_140)); \
54811  __ret_140; \
54812 })
54813 #else
54814 #define vmulxs_lane_f32(__p0_141, __p1_141, __p2_141) __extension__ ({ \
54815  float32_t __s0_141 = __p0_141; \
54816  float32x2_t __s1_141 = __p1_141; \
54817  float32x2_t __rev1_141; __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 1, 0); \
54818  float32_t __ret_141; \
54819  __ret_141 = __noswap_vmulxs_f32(__s0_141, __noswap_vget_lane_f32(__rev1_141, __p2_141)); \
54820  __ret_141; \
54821 })
54822 #endif
54823 
54824 #ifdef __LITTLE_ENDIAN__
54825 #define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54826  float64x2_t __s0 = __p0; \
54827  float64x1_t __s1 = __p1; \
54828  float64x2_t __ret; \
54829  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54830  __ret; \
54831 })
54832 #else
54833 #define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54834  float64x2_t __s0 = __p0; \
54835  float64x1_t __s1 = __p1; \
54836  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54837  float64x2_t __ret; \
54838  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54839  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54840  __ret; \
54841 })
54842 #endif
54843 
54844 #ifdef __LITTLE_ENDIAN__
54845 #define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54846  float32x4_t __s0 = __p0; \
54847  float32x2_t __s1 = __p1; \
54848  float32x4_t __ret; \
54849  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54850  __ret; \
54851 })
54852 #else
54853 #define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54854  float32x4_t __s0 = __p0; \
54855  float32x2_t __s1 = __p1; \
54856  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54857  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54858  float32x4_t __ret; \
54859  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54860  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54861  __ret; \
54862 })
54863 #endif
54864 
54865 #ifdef __LITTLE_ENDIAN__
54866 #define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54867  float32x2_t __s0 = __p0; \
54868  float32x2_t __s1 = __p1; \
54869  float32x2_t __ret; \
54870  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54871  __ret; \
54872 })
54873 #else
54874 #define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54875  float32x2_t __s0 = __p0; \
54876  float32x2_t __s1 = __p1; \
54877  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54878  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54879  float32x2_t __ret; \
54880  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54881  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54882  __ret; \
54883 })
54884 #endif
54885 
54886 #ifdef __LITTLE_ENDIAN__
54887 #define vmulxd_laneq_f64(__p0_142, __p1_142, __p2_142) __extension__ ({ \
54888  float64_t __s0_142 = __p0_142; \
54889  float64x2_t __s1_142 = __p1_142; \
54890  float64_t __ret_142; \
54891  __ret_142 = vmulxd_f64(__s0_142, vgetq_lane_f64(__s1_142, __p2_142)); \
54892  __ret_142; \
54893 })
54894 #else
54895 #define vmulxd_laneq_f64(__p0_143, __p1_143, __p2_143) __extension__ ({ \
54896  float64_t __s0_143 = __p0_143; \
54897  float64x2_t __s1_143 = __p1_143; \
54898  float64x2_t __rev1_143; __rev1_143 = __builtin_shufflevector(__s1_143, __s1_143, 1, 0); \
54899  float64_t __ret_143; \
54900  __ret_143 = __noswap_vmulxd_f64(__s0_143, __noswap_vgetq_lane_f64(__rev1_143, __p2_143)); \
54901  __ret_143; \
54902 })
54903 #endif
54904 
54905 #ifdef __LITTLE_ENDIAN__
54906 #define vmulxs_laneq_f32(__p0_144, __p1_144, __p2_144) __extension__ ({ \
54907  float32_t __s0_144 = __p0_144; \
54908  float32x4_t __s1_144 = __p1_144; \
54909  float32_t __ret_144; \
54910  __ret_144 = vmulxs_f32(__s0_144, vgetq_lane_f32(__s1_144, __p2_144)); \
54911  __ret_144; \
54912 })
54913 #else
54914 #define vmulxs_laneq_f32(__p0_145, __p1_145, __p2_145) __extension__ ({ \
54915  float32_t __s0_145 = __p0_145; \
54916  float32x4_t __s1_145 = __p1_145; \
54917  float32x4_t __rev1_145; __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 3, 2, 1, 0); \
54918  float32_t __ret_145; \
54919  __ret_145 = __noswap_vmulxs_f32(__s0_145, __noswap_vgetq_lane_f32(__rev1_145, __p2_145)); \
54920  __ret_145; \
54921 })
54922 #endif
54923 
54924 #ifdef __LITTLE_ENDIAN__
54925 #define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54926  float64x2_t __s0 = __p0; \
54927  float64x2_t __s1 = __p1; \
54928  float64x2_t __ret; \
54929  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54930  __ret; \
54931 })
54932 #else
54933 #define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54934  float64x2_t __s0 = __p0; \
54935  float64x2_t __s1 = __p1; \
54936  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54937  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54938  float64x2_t __ret; \
54939  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54940  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54941  __ret; \
54942 })
54943 #endif
54944 
54945 #ifdef __LITTLE_ENDIAN__
54946 #define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54947  float32x4_t __s0 = __p0; \
54948  float32x4_t __s1 = __p1; \
54949  float32x4_t __ret; \
54950  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54951  __ret; \
54952 })
54953 #else
54954 #define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54955  float32x4_t __s0 = __p0; \
54956  float32x4_t __s1 = __p1; \
54957  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54958  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54959  float32x4_t __ret; \
54960  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54961  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54962  __ret; \
54963 })
54964 #endif
54965 
54966 #ifdef __LITTLE_ENDIAN__
54967 #define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54968  float32x2_t __s0 = __p0; \
54969  float32x4_t __s1 = __p1; \
54970  float32x2_t __ret; \
54971  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54972  __ret; \
54973 })
54974 #else
54975 #define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54976  float32x2_t __s0 = __p0; \
54977  float32x4_t __s1 = __p1; \
54978  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54979  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54980  float32x2_t __ret; \
54981  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54983  __ret; \
54984 })
54985 #endif
54986 
54987 #ifdef __LITTLE_ENDIAN__
54988 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
54989  float64x2_t __ret;
54990  __ret = -__p0;
54991  return __ret;
54992 }
54993 #else
54994 __ai float64x2_t vnegq_f64(float64x2_t __p0) {
54995  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54996  float64x2_t __ret;
54997  __ret = -__rev0;
54998  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54999  return __ret;
55000 }
55001 #endif
55002 
55003 #ifdef __LITTLE_ENDIAN__
55004 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
55005  int64x2_t __ret;
55006  __ret = -__p0;
55007  return __ret;
55008 }
55009 #else
55010 __ai int64x2_t vnegq_s64(int64x2_t __p0) {
55011  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55012  int64x2_t __ret;
55013  __ret = -__rev0;
55014  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55015  return __ret;
55016 }
55017 #endif
55018 
55019 #ifdef __LITTLE_ENDIAN__
55020 __ai float64x1_t vneg_f64(float64x1_t __p0) {
55021  float64x1_t __ret;
55022  __ret = -__p0;
55023  return __ret;
55024 }
55025 #else
55026 __ai float64x1_t vneg_f64(float64x1_t __p0) {
55027  float64x1_t __ret;
55028  __ret = -__p0;
55029  return __ret;
55030 }
55031 #endif
55032 
55033 #ifdef __LITTLE_ENDIAN__
55034 __ai int64x1_t vneg_s64(int64x1_t __p0) {
55035  int64x1_t __ret;
55036  __ret = -__p0;
55037  return __ret;
55038 }
55039 #else
55040 __ai int64x1_t vneg_s64(int64x1_t __p0) {
55041  int64x1_t __ret;
55042  __ret = -__p0;
55043  return __ret;
55044 }
55045 #endif
55046 
55047 #ifdef __LITTLE_ENDIAN__
55048 __ai int64_t vnegd_s64(int64_t __p0) {
55049  int64_t __ret;
55050  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55051  return __ret;
55052 }
55053 #else
55054 __ai int64_t vnegd_s64(int64_t __p0) {
55055  int64_t __ret;
55056  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55057  return __ret;
55058 }
55059 #endif
55060 
55061 #ifdef __LITTLE_ENDIAN__
55062 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55063  uint8x16_t __ret;
55064  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55065  return __ret;
55066 }
55067 #else
55068 __ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55069  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55070  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55071  uint8x16_t __ret;
55072  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55073  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55074  return __ret;
55075 }
55076 #endif
55077 
55078 #ifdef __LITTLE_ENDIAN__
55079 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55080  uint32x4_t __ret;
55081  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55082  return __ret;
55083 }
55084 #else
55085 __ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55086  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55087  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55088  uint32x4_t __ret;
55089  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55090  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55091  return __ret;
55092 }
55093 #endif
55094 
55095 #ifdef __LITTLE_ENDIAN__
55096 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55097  uint64x2_t __ret;
55098  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
55099  return __ret;
55100 }
55101 #else
55102 __ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55103  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55104  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55105  uint64x2_t __ret;
55106  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
55107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55108  return __ret;
55109 }
55110 #endif
55111 
55112 #ifdef __LITTLE_ENDIAN__
55113 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55114  uint16x8_t __ret;
55115  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55116  return __ret;
55117 }
55118 #else
55119 __ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55120  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55121  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55122  uint16x8_t __ret;
55123  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55124  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55125  return __ret;
55126 }
55127 #endif
55128 
55129 #ifdef __LITTLE_ENDIAN__
55130 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55131  int8x16_t __ret;
55132  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55133  return __ret;
55134 }
55135 #else
55136 __ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55137  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55138  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55139  int8x16_t __ret;
55140  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55141  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55142  return __ret;
55143 }
55144 #endif
55145 
55146 #ifdef __LITTLE_ENDIAN__
55147 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55148  float64x2_t __ret;
55149  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55150  return __ret;
55151 }
55152 #else
55153 __ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55154  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55155  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55156  float64x2_t __ret;
55157  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55158  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55159  return __ret;
55160 }
55161 #endif
55162 
55163 #ifdef __LITTLE_ENDIAN__
55164 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55165  float32x4_t __ret;
55166  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55167  return __ret;
55168 }
55169 #else
55170 __ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55171  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55172  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55173  float32x4_t __ret;
55174  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55175  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55176  return __ret;
55177 }
55178 #endif
55179 
55180 #ifdef __LITTLE_ENDIAN__
55181 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55182  int32x4_t __ret;
55183  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55184  return __ret;
55185 }
55186 #else
55187 __ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55188  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55189  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55190  int32x4_t __ret;
55191  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55192  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55193  return __ret;
55194 }
55195 #endif
55196 
55197 #ifdef __LITTLE_ENDIAN__
55198 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55199  int64x2_t __ret;
55200  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
55201  return __ret;
55202 }
55203 #else
55204 __ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55205  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55206  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55207  int64x2_t __ret;
55208  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
55209  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55210  return __ret;
55211 }
55212 #endif
55213 
55214 #ifdef __LITTLE_ENDIAN__
55215 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55216  int16x8_t __ret;
55217  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55218  return __ret;
55219 }
55220 #else
55221 __ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55222  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55223  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55224  int16x8_t __ret;
55225  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55226  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55227  return __ret;
55228 }
55229 #endif
55230 
55231 #ifdef __LITTLE_ENDIAN__
55232 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55233  uint64_t __ret;
55234  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__p0);
55235  return __ret;
55236 }
55237 #else
55238 __ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55239  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55240  uint64_t __ret;
55241  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__rev0);
55242  return __ret;
55243 }
55244 #endif
55245 
55246 #ifdef __LITTLE_ENDIAN__
55247 __ai float64_t vpaddd_f64(float64x2_t __p0) {
55248  float64_t __ret;
55249  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__p0);
55250  return __ret;
55251 }
55252 #else
55253 __ai float64_t vpaddd_f64(float64x2_t __p0) {
55254  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55255  float64_t __ret;
55256  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__rev0);
55257  return __ret;
55258 }
55259 #endif
55260 
55261 #ifdef __LITTLE_ENDIAN__
55262 __ai int64_t vpaddd_s64(int64x2_t __p0) {
55263  int64_t __ret;
55264  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__p0);
55265  return __ret;
55266 }
55267 #else
55268 __ai int64_t vpaddd_s64(int64x2_t __p0) {
55269  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55270  int64_t __ret;
55271  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__rev0);
55272  return __ret;
55273 }
55274 #endif
55275 
55276 #ifdef __LITTLE_ENDIAN__
55277 __ai float32_t vpadds_f32(float32x2_t __p0) {
55278  float32_t __ret;
55279  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__p0);
55280  return __ret;
55281 }
55282 #else
55283 __ai float32_t vpadds_f32(float32x2_t __p0) {
55284  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55285  float32_t __ret;
55286  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__rev0);
55287  return __ret;
55288 }
55289 #endif
55290 
55291 #ifdef __LITTLE_ENDIAN__
55292 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55293  uint8x16_t __ret;
55294  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55295  return __ret;
55296 }
55297 #else
55298 __ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55299  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55300  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55301  uint8x16_t __ret;
55302  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55303  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55304  return __ret;
55305 }
55306 #endif
55307 
55308 #ifdef __LITTLE_ENDIAN__
55309 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55310  uint32x4_t __ret;
55311  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55312  return __ret;
55313 }
55314 #else
55315 __ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55316  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55317  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55318  uint32x4_t __ret;
55319  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55320  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55321  return __ret;
55322 }
55323 #endif
55324 
55325 #ifdef __LITTLE_ENDIAN__
55326 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55327  uint16x8_t __ret;
55328  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55329  return __ret;
55330 }
55331 #else
55332 __ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55333  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55334  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55335  uint16x8_t __ret;
55336  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55337  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55338  return __ret;
55339 }
55340 #endif
55341 
55342 #ifdef __LITTLE_ENDIAN__
55343 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55344  int8x16_t __ret;
55345  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55346  return __ret;
55347 }
55348 #else
55349 __ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55350  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55351  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55352  int8x16_t __ret;
55353  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55354  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55355  return __ret;
55356 }
55357 #endif
55358 
55359 #ifdef __LITTLE_ENDIAN__
55360 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55361  float64x2_t __ret;
55362  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55363  return __ret;
55364 }
55365 #else
55366 __ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55367  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55368  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55369  float64x2_t __ret;
55370  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55371  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55372  return __ret;
55373 }
55374 #endif
55375 
55376 #ifdef __LITTLE_ENDIAN__
55377 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55378  float32x4_t __ret;
55379  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55380  return __ret;
55381 }
55382 #else
55383 __ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55384  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55385  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55386  float32x4_t __ret;
55387  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55388  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55389  return __ret;
55390 }
55391 #endif
55392 
55393 #ifdef __LITTLE_ENDIAN__
55394 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55395  int32x4_t __ret;
55396  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55397  return __ret;
55398 }
55399 #else
55400 __ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55401  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55402  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55403  int32x4_t __ret;
55404  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55405  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55406  return __ret;
55407 }
55408 #endif
55409 
55410 #ifdef __LITTLE_ENDIAN__
55411 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55412  int16x8_t __ret;
55413  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55414  return __ret;
55415 }
55416 #else
55417 __ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55418  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55419  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55420  int16x8_t __ret;
55421  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55422  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55423  return __ret;
55424 }
55425 #endif
55426 
55427 #ifdef __LITTLE_ENDIAN__
55428 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55429  float64_t __ret;
55430  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__p0);
55431  return __ret;
55432 }
55433 #else
55434 __ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55435  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55436  float64_t __ret;
55437  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__rev0);
55438  return __ret;
55439 }
55440 #endif
55441 
55442 #ifdef __LITTLE_ENDIAN__
55443 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
55444  float32_t __ret;
55445  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__p0);
55446  return __ret;
55447 }
55448 #else
55449 __ai float32_t vpmaxs_f32(float32x2_t __p0) {
55450  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55451  float32_t __ret;
55452  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__rev0);
55453  return __ret;
55454 }
55455 #endif
55456 
55457 #ifdef __LITTLE_ENDIAN__
55458 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55459  float64x2_t __ret;
55460  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55461  return __ret;
55462 }
55463 #else
55464 __ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55465  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55466  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55467  float64x2_t __ret;
55468  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55470  return __ret;
55471 }
55472 #endif
55473 
55474 #ifdef __LITTLE_ENDIAN__
55475 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55476  float32x4_t __ret;
55477  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55478  return __ret;
55479 }
55480 #else
55481 __ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55482  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55483  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55484  float32x4_t __ret;
55485  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55487  return __ret;
55488 }
55489 #endif
55490 
55491 #ifdef __LITTLE_ENDIAN__
55492 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55493  float32x2_t __ret;
55494  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55495  return __ret;
55496 }
55497 #else
55498 __ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55499  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55500  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55501  float32x2_t __ret;
55502  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55503  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55504  return __ret;
55505 }
55506 #endif
55507 
55508 #ifdef __LITTLE_ENDIAN__
55509 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55510  float64_t __ret;
55511  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__p0);
55512  return __ret;
55513 }
55514 #else
55515 __ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55516  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55517  float64_t __ret;
55518  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__rev0);
55519  return __ret;
55520 }
55521 #endif
55522 
55523 #ifdef __LITTLE_ENDIAN__
55524 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55525  float32_t __ret;
55526  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__p0);
55527  return __ret;
55528 }
55529 #else
55530 __ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55531  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55532  float32_t __ret;
55533  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__rev0);
55534  return __ret;
55535 }
55536 #endif
55537 
55538 #ifdef __LITTLE_ENDIAN__
55539 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55540  uint8x16_t __ret;
55541  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55542  return __ret;
55543 }
55544 #else
55545 __ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55546  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55547  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55548  uint8x16_t __ret;
55549  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55550  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55551  return __ret;
55552 }
55553 #endif
55554 
55555 #ifdef __LITTLE_ENDIAN__
55556 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55557  uint32x4_t __ret;
55558  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55559  return __ret;
55560 }
55561 #else
55562 __ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55563  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55564  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55565  uint32x4_t __ret;
55566  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55567  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55568  return __ret;
55569 }
55570 #endif
55571 
55572 #ifdef __LITTLE_ENDIAN__
55573 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55574  uint16x8_t __ret;
55575  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55576  return __ret;
55577 }
55578 #else
55579 __ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55580  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55581  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55582  uint16x8_t __ret;
55583  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55585  return __ret;
55586 }
55587 #endif
55588 
55589 #ifdef __LITTLE_ENDIAN__
55590 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55591  int8x16_t __ret;
55592  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55593  return __ret;
55594 }
55595 #else
55596 __ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55597  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55598  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55599  int8x16_t __ret;
55600  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55601  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55602  return __ret;
55603 }
55604 #endif
55605 
55606 #ifdef __LITTLE_ENDIAN__
55607 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55608  float64x2_t __ret;
55609  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55610  return __ret;
55611 }
55612 #else
55613 __ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55614  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55615  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55616  float64x2_t __ret;
55617  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55618  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55619  return __ret;
55620 }
55621 #endif
55622 
55623 #ifdef __LITTLE_ENDIAN__
55624 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55625  float32x4_t __ret;
55626  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55627  return __ret;
55628 }
55629 #else
55630 __ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55631  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55632  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55633  float32x4_t __ret;
55634  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55635  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55636  return __ret;
55637 }
55638 #endif
55639 
55640 #ifdef __LITTLE_ENDIAN__
55641 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55642  int32x4_t __ret;
55643  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55644  return __ret;
55645 }
55646 #else
55647 __ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55648  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55649  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55650  int32x4_t __ret;
55651  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55653  return __ret;
55654 }
55655 #endif
55656 
55657 #ifdef __LITTLE_ENDIAN__
55658 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55659  int16x8_t __ret;
55660  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55661  return __ret;
55662 }
55663 #else
55664 __ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55665  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55666  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55667  int16x8_t __ret;
55668  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55669  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55670  return __ret;
55671 }
55672 #endif
55673 
55674 #ifdef __LITTLE_ENDIAN__
55675 __ai float64_t vpminqd_f64(float64x2_t __p0) {
55676  float64_t __ret;
55677  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__p0);
55678  return __ret;
55679 }
55680 #else
55681 __ai float64_t vpminqd_f64(float64x2_t __p0) {
55682  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55683  float64_t __ret;
55684  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__rev0);
55685  return __ret;
55686 }
55687 #endif
55688 
55689 #ifdef __LITTLE_ENDIAN__
55690 __ai float32_t vpmins_f32(float32x2_t __p0) {
55691  float32_t __ret;
55692  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__p0);
55693  return __ret;
55694 }
55695 #else
55696 __ai float32_t vpmins_f32(float32x2_t __p0) {
55697  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55698  float32_t __ret;
55699  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__rev0);
55700  return __ret;
55701 }
55702 #endif
55703 
55704 #ifdef __LITTLE_ENDIAN__
55705 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55706  float64x2_t __ret;
55707  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55708  return __ret;
55709 }
55710 #else
55711 __ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55712  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55713  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55714  float64x2_t __ret;
55715  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55716  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55717  return __ret;
55718 }
55719 #endif
55720 
55721 #ifdef __LITTLE_ENDIAN__
55722 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55723  float32x4_t __ret;
55724  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55725  return __ret;
55726 }
55727 #else
55728 __ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55729  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55730  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55731  float32x4_t __ret;
55732  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55733  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55734  return __ret;
55735 }
55736 #endif
55737 
55738 #ifdef __LITTLE_ENDIAN__
55739 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55740  float32x2_t __ret;
55741  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55742  return __ret;
55743 }
55744 #else
55745 __ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55746  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55747  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55748  float32x2_t __ret;
55749  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55750  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55751  return __ret;
55752 }
55753 #endif
55754 
55755 #ifdef __LITTLE_ENDIAN__
55756 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55757  float64_t __ret;
55758  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__p0);
55759  return __ret;
55760 }
55761 #else
55762 __ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55763  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55764  float64_t __ret;
55765  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__rev0);
55766  return __ret;
55767 }
55768 #endif
55769 
55770 #ifdef __LITTLE_ENDIAN__
55771 __ai float32_t vpminnms_f32(float32x2_t __p0) {
55772  float32_t __ret;
55773  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__p0);
55774  return __ret;
55775 }
55776 #else
55777 __ai float32_t vpminnms_f32(float32x2_t __p0) {
55778  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55779  float32_t __ret;
55780  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__rev0);
55781  return __ret;
55782 }
55783 #endif
55784 
55785 #ifdef __LITTLE_ENDIAN__
55786 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55787  int64x2_t __ret;
55788  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
55789  return __ret;
55790 }
55791 #else
55792 __ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55793  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55794  int64x2_t __ret;
55795  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
55796  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55797  return __ret;
55798 }
55799 #endif
55800 
55801 #ifdef __LITTLE_ENDIAN__
55802 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
55803  int64x1_t __ret;
55804  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55805  return __ret;
55806 }
55807 #else
55808 __ai int64x1_t vqabs_s64(int64x1_t __p0) {
55809  int64x1_t __ret;
55810  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55811  return __ret;
55812 }
55813 #endif
55814 
55815 #ifdef __LITTLE_ENDIAN__
55816 __ai int8_t vqabsb_s8(int8_t __p0) {
55817  int8_t __ret;
55818  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55819  return __ret;
55820 }
55821 #else
55822 __ai int8_t vqabsb_s8(int8_t __p0) {
55823  int8_t __ret;
55824  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55825  return __ret;
55826 }
55827 #endif
55828 
55829 #ifdef __LITTLE_ENDIAN__
55830 __ai int32_t vqabss_s32(int32_t __p0) {
55831  int32_t __ret;
55832  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55833  return __ret;
55834 }
55835 #else
55836 __ai int32_t vqabss_s32(int32_t __p0) {
55837  int32_t __ret;
55838  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55839  return __ret;
55840 }
55841 #endif
55842 
55843 #ifdef __LITTLE_ENDIAN__
55844 __ai int64_t vqabsd_s64(int64_t __p0) {
55845  int64_t __ret;
55846  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55847  return __ret;
55848 }
55849 #else
55850 __ai int64_t vqabsd_s64(int64_t __p0) {
55851  int64_t __ret;
55852  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55853  return __ret;
55854 }
55855 #endif
55856 
55857 #ifdef __LITTLE_ENDIAN__
55858 __ai int16_t vqabsh_s16(int16_t __p0) {
55859  int16_t __ret;
55860  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55861  return __ret;
55862 }
55863 #else
55864 __ai int16_t vqabsh_s16(int16_t __p0) {
55865  int16_t __ret;
55866  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55867  return __ret;
55868 }
55869 #endif
55870 
55871 #ifdef __LITTLE_ENDIAN__
55872 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55873  uint8_t __ret;
55874  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55875  return __ret;
55876 }
55877 #else
55878 __ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55879  uint8_t __ret;
55880  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55881  return __ret;
55882 }
55883 #endif
55884 
55885 #ifdef __LITTLE_ENDIAN__
55886 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55887  uint32_t __ret;
55888  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55889  return __ret;
55890 }
55891 #else
55892 __ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55893  uint32_t __ret;
55894  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55895  return __ret;
55896 }
55897 #endif
55898 
55899 #ifdef __LITTLE_ENDIAN__
55900 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55901  uint64_t __ret;
55902  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55903  return __ret;
55904 }
55905 #else
55906 __ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55907  uint64_t __ret;
55908  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55909  return __ret;
55910 }
55911 #endif
55912 
55913 #ifdef __LITTLE_ENDIAN__
55914 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55915  uint16_t __ret;
55916  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55917  return __ret;
55918 }
55919 #else
55920 __ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55921  uint16_t __ret;
55922  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55923  return __ret;
55924 }
55925 #endif
55926 
55927 #ifdef __LITTLE_ENDIAN__
55928 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55929  int8_t __ret;
55930  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55931  return __ret;
55932 }
55933 #else
55934 __ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55935  int8_t __ret;
55936  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55937  return __ret;
55938 }
55939 #endif
55940 
55941 #ifdef __LITTLE_ENDIAN__
55942 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55943  int32_t __ret;
55944  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55945  return __ret;
55946 }
55947 #else
55948 __ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55949  int32_t __ret;
55950  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55951  return __ret;
55952 }
55953 __ai int32_t __noswap_vqadds_s32(int32_t __p0, int32_t __p1) {
55954  int32_t __ret;
55955  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55956  return __ret;
55957 }
55958 #endif
55959 
55960 #ifdef __LITTLE_ENDIAN__
55961 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55962  int64_t __ret;
55963  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55964  return __ret;
55965 }
55966 #else
55967 __ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55968  int64_t __ret;
55969  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55970  return __ret;
55971 }
55972 #endif
55973 
55974 #ifdef __LITTLE_ENDIAN__
55975 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55976  int16_t __ret;
55977  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55978  return __ret;
55979 }
55980 #else
55981 __ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55982  int16_t __ret;
55983  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55984  return __ret;
55985 }
55986 __ai int16_t __noswap_vqaddh_s16(int16_t __p0, int16_t __p1) {
55987  int16_t __ret;
55988  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55989  return __ret;
55990 }
55991 #endif
55992 
55993 #ifdef __LITTLE_ENDIAN__
55994 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55995  int64_t __ret;
55996  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
55997  return __ret;
55998 }
55999 #else
56000 __ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56001  int64_t __ret;
56002  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
56003  return __ret;
56004 }
56005 #endif
56006 
56007 #ifdef __LITTLE_ENDIAN__
56008 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56009  int32_t __ret;
56010  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56011  return __ret;
56012 }
56013 #else
56014 __ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56015  int32_t __ret;
56016  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56017  return __ret;
56018 }
56019 #endif
56020 
56021 #ifdef __LITTLE_ENDIAN__
56022 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56023  int64x2_t __ret;
56024  __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56025  return __ret;
56026 }
56027 #else
56028 __ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56029  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56030  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56031  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56032  int64x2_t __ret;
56033  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56035  return __ret;
56036 }
56037 #endif
56038 
56039 #ifdef __LITTLE_ENDIAN__
56040 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56041  int32x4_t __ret;
56042  __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56043  return __ret;
56044 }
56045 #else
56046 __ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56047  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56048  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56049  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56050  int32x4_t __ret;
56051  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56052  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56053  return __ret;
56054 }
56055 #endif
56056 
56057 #ifdef __LITTLE_ENDIAN__
56058 #define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56059  int64x2_t __s0 = __p0; \
56060  int32x4_t __s1 = __p1; \
56061  int32x2_t __s2 = __p2; \
56062  int64x2_t __ret; \
56063  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56064  __ret; \
56065 })
56066 #else
56067 #define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56068  int64x2_t __s0 = __p0; \
56069  int32x4_t __s1 = __p1; \
56070  int32x2_t __s2 = __p2; \
56071  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56072  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56073  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56074  int64x2_t __ret; \
56075  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56076  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56077  __ret; \
56078 })
56079 #endif
56080 
56081 #ifdef __LITTLE_ENDIAN__
56082 #define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56083  int32x4_t __s0 = __p0; \
56084  int16x8_t __s1 = __p1; \
56085  int16x4_t __s2 = __p2; \
56086  int32x4_t __ret; \
56087  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56088  __ret; \
56089 })
56090 #else
56091 #define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56092  int32x4_t __s0 = __p0; \
56093  int16x8_t __s1 = __p1; \
56094  int16x4_t __s2 = __p2; \
56095  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56096  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56097  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56098  int32x4_t __ret; \
56099  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56101  __ret; \
56102 })
56103 #endif
56104 
56105 #ifdef __LITTLE_ENDIAN__
56106 #define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56107  int64x2_t __s0 = __p0; \
56108  int32x4_t __s1 = __p1; \
56109  int32x4_t __s2 = __p2; \
56110  int64x2_t __ret; \
56111  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56112  __ret; \
56113 })
56114 #else
56115 #define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56116  int64x2_t __s0 = __p0; \
56117  int32x4_t __s1 = __p1; \
56118  int32x4_t __s2 = __p2; \
56119  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56120  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56121  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56122  int64x2_t __ret; \
56123  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56124  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56125  __ret; \
56126 })
56127 #endif
56128 
56129 #ifdef __LITTLE_ENDIAN__
56130 #define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56131  int32x4_t __s0 = __p0; \
56132  int16x8_t __s1 = __p1; \
56133  int16x8_t __s2 = __p2; \
56134  int32x4_t __ret; \
56135  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56136  __ret; \
56137 })
56138 #else
56139 #define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56140  int32x4_t __s0 = __p0; \
56141  int16x8_t __s1 = __p1; \
56142  int16x8_t __s2 = __p2; \
56143  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56144  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56145  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56146  int32x4_t __ret; \
56147  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56148  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56149  __ret; \
56150 })
56151 #endif
56152 
56153 #ifdef __LITTLE_ENDIAN__
56154 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56155  int64x2_t __ret;
56156  __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
56157  return __ret;
56158 }
56159 #else
56160 __ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56161  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56162  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56163  int64x2_t __ret;
56164  __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56166  return __ret;
56167 }
56168 #endif
56169 
56170 #ifdef __LITTLE_ENDIAN__
56171 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56172  int32x4_t __ret;
56173  __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
56174  return __ret;
56175 }
56176 #else
56177 __ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56178  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56179  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56180  int32x4_t __ret;
56181  __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56182  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56183  return __ret;
56184 }
56185 #endif
56186 
56187 #ifdef __LITTLE_ENDIAN__
56188 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56189  int64_t __s0 = __p0; \
56190  int32_t __s1 = __p1; \
56191  int32x2_t __s2 = __p2; \
56192  int64_t __ret; \
56193  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56194  __ret; \
56195 })
56196 #else
56197 #define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56198  int64_t __s0 = __p0; \
56199  int32_t __s1 = __p1; \
56200  int32x2_t __s2 = __p2; \
56201  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56202  int64_t __ret; \
56203  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56204  __ret; \
56205 })
56206 #endif
56207 
56208 #ifdef __LITTLE_ENDIAN__
56209 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56210  int32_t __s0 = __p0; \
56211  int16_t __s1 = __p1; \
56212  int16x4_t __s2 = __p2; \
56213  int32_t __ret; \
56214  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56215  __ret; \
56216 })
56217 #else
56218 #define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56219  int32_t __s0 = __p0; \
56220  int16_t __s1 = __p1; \
56221  int16x4_t __s2 = __p2; \
56222  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56223  int32_t __ret; \
56224  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56225  __ret; \
56226 })
56227 #endif
56228 
56229 #ifdef __LITTLE_ENDIAN__
56230 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56231  int64_t __s0 = __p0; \
56232  int32_t __s1 = __p1; \
56233  int32x4_t __s2 = __p2; \
56234  int64_t __ret; \
56235  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56236  __ret; \
56237 })
56238 #else
56239 #define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56240  int64_t __s0 = __p0; \
56241  int32_t __s1 = __p1; \
56242  int32x4_t __s2 = __p2; \
56243  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56244  int64_t __ret; \
56245  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56246  __ret; \
56247 })
56248 #endif
56249 
56250 #ifdef __LITTLE_ENDIAN__
56251 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56252  int32_t __s0 = __p0; \
56253  int16_t __s1 = __p1; \
56254  int16x8_t __s2 = __p2; \
56255  int32_t __ret; \
56256  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56257  __ret; \
56258 })
56259 #else
56260 #define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56261  int32_t __s0 = __p0; \
56262  int16_t __s1 = __p1; \
56263  int16x8_t __s2 = __p2; \
56264  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56265  int32_t __ret; \
56266  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56267  __ret; \
56268 })
56269 #endif
56270 
56271 #ifdef __LITTLE_ENDIAN__
56272 #define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56273  int64x2_t __s0 = __p0; \
56274  int32x2_t __s1 = __p1; \
56275  int32x4_t __s2 = __p2; \
56276  int64x2_t __ret; \
56277  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56278  __ret; \
56279 })
56280 #else
56281 #define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56282  int64x2_t __s0 = __p0; \
56283  int32x2_t __s1 = __p1; \
56284  int32x4_t __s2 = __p2; \
56285  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56286  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56287  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56288  int64x2_t __ret; \
56289  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56291  __ret; \
56292 })
56293 #endif
56294 
56295 #ifdef __LITTLE_ENDIAN__
56296 #define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56297  int32x4_t __s0 = __p0; \
56298  int16x4_t __s1 = __p1; \
56299  int16x8_t __s2 = __p2; \
56300  int32x4_t __ret; \
56301  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56302  __ret; \
56303 })
56304 #else
56305 #define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56306  int32x4_t __s0 = __p0; \
56307  int16x4_t __s1 = __p1; \
56308  int16x8_t __s2 = __p2; \
56309  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56310  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56311  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56312  int32x4_t __ret; \
56313  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56315  __ret; \
56316 })
56317 #endif
56318 
56319 #ifdef __LITTLE_ENDIAN__
56320 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56321  int64_t __ret;
56322  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56323  return __ret;
56324 }
56325 #else
56326 __ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56327  int64_t __ret;
56328  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56329  return __ret;
56330 }
56331 #endif
56332 
56333 #ifdef __LITTLE_ENDIAN__
56334 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56335  int32_t __ret;
56336  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56337  return __ret;
56338 }
56339 #else
56340 __ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56341  int32_t __ret;
56342  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56343  return __ret;
56344 }
56345 #endif
56346 
56347 #ifdef __LITTLE_ENDIAN__
56348 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56349  int64x2_t __ret;
56350  __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56351  return __ret;
56352 }
56353 #else
56354 __ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56355  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56356  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56357  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56358  int64x2_t __ret;
56359  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56360  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56361  return __ret;
56362 }
56363 #endif
56364 
56365 #ifdef __LITTLE_ENDIAN__
56366 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56367  int32x4_t __ret;
56368  __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56369  return __ret;
56370 }
56371 #else
56372 __ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56373  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56374  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56375  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56376  int32x4_t __ret;
56377  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56378  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56379  return __ret;
56380 }
56381 #endif
56382 
56383 #ifdef __LITTLE_ENDIAN__
56384 #define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56385  int64x2_t __s0 = __p0; \
56386  int32x4_t __s1 = __p1; \
56387  int32x2_t __s2 = __p2; \
56388  int64x2_t __ret; \
56389  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56390  __ret; \
56391 })
56392 #else
56393 #define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56394  int64x2_t __s0 = __p0; \
56395  int32x4_t __s1 = __p1; \
56396  int32x2_t __s2 = __p2; \
56397  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56398  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56399  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56400  int64x2_t __ret; \
56401  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56402  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56403  __ret; \
56404 })
56405 #endif
56406 
56407 #ifdef __LITTLE_ENDIAN__
56408 #define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56409  int32x4_t __s0 = __p0; \
56410  int16x8_t __s1 = __p1; \
56411  int16x4_t __s2 = __p2; \
56412  int32x4_t __ret; \
56413  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56414  __ret; \
56415 })
56416 #else
56417 #define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56418  int32x4_t __s0 = __p0; \
56419  int16x8_t __s1 = __p1; \
56420  int16x4_t __s2 = __p2; \
56421  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56422  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56423  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56424  int32x4_t __ret; \
56425  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56426  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56427  __ret; \
56428 })
56429 #endif
56430 
56431 #ifdef __LITTLE_ENDIAN__
56432 #define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56433  int64x2_t __s0 = __p0; \
56434  int32x4_t __s1 = __p1; \
56435  int32x4_t __s2 = __p2; \
56436  int64x2_t __ret; \
56437  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56438  __ret; \
56439 })
56440 #else
56441 #define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56442  int64x2_t __s0 = __p0; \
56443  int32x4_t __s1 = __p1; \
56444  int32x4_t __s2 = __p2; \
56445  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56446  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56447  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56448  int64x2_t __ret; \
56449  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56451  __ret; \
56452 })
56453 #endif
56454 
56455 #ifdef __LITTLE_ENDIAN__
56456 #define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56457  int32x4_t __s0 = __p0; \
56458  int16x8_t __s1 = __p1; \
56459  int16x8_t __s2 = __p2; \
56460  int32x4_t __ret; \
56461  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56462  __ret; \
56463 })
56464 #else
56465 #define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56466  int32x4_t __s0 = __p0; \
56467  int16x8_t __s1 = __p1; \
56468  int16x8_t __s2 = __p2; \
56469  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56470  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56471  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56472  int32x4_t __ret; \
56473  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56474  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56475  __ret; \
56476 })
56477 #endif
56478 
56479 #ifdef __LITTLE_ENDIAN__
56480 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56481  int64x2_t __ret;
56482  __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
56483  return __ret;
56484 }
56485 #else
56486 __ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56487  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56488  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56489  int64x2_t __ret;
56490  __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56491  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56492  return __ret;
56493 }
56494 #endif
56495 
56496 #ifdef __LITTLE_ENDIAN__
56497 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56498  int32x4_t __ret;
56499  __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
56500  return __ret;
56501 }
56502 #else
56503 __ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56504  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56505  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56506  int32x4_t __ret;
56507  __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56509  return __ret;
56510 }
56511 #endif
56512 
56513 #ifdef __LITTLE_ENDIAN__
56514 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56515  int64_t __s0 = __p0; \
56516  int32_t __s1 = __p1; \
56517  int32x2_t __s2 = __p2; \
56518  int64_t __ret; \
56519  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56520  __ret; \
56521 })
56522 #else
56523 #define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56524  int64_t __s0 = __p0; \
56525  int32_t __s1 = __p1; \
56526  int32x2_t __s2 = __p2; \
56527  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56528  int64_t __ret; \
56529  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56530  __ret; \
56531 })
56532 #endif
56533 
56534 #ifdef __LITTLE_ENDIAN__
56535 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56536  int32_t __s0 = __p0; \
56537  int16_t __s1 = __p1; \
56538  int16x4_t __s2 = __p2; \
56539  int32_t __ret; \
56540  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56541  __ret; \
56542 })
56543 #else
56544 #define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56545  int32_t __s0 = __p0; \
56546  int16_t __s1 = __p1; \
56547  int16x4_t __s2 = __p2; \
56548  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56549  int32_t __ret; \
56550  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56551  __ret; \
56552 })
56553 #endif
56554 
56555 #ifdef __LITTLE_ENDIAN__
56556 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56557  int64_t __s0 = __p0; \
56558  int32_t __s1 = __p1; \
56559  int32x4_t __s2 = __p2; \
56560  int64_t __ret; \
56561  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56562  __ret; \
56563 })
56564 #else
56565 #define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56566  int64_t __s0 = __p0; \
56567  int32_t __s1 = __p1; \
56568  int32x4_t __s2 = __p2; \
56569  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56570  int64_t __ret; \
56571  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56572  __ret; \
56573 })
56574 #endif
56575 
56576 #ifdef __LITTLE_ENDIAN__
56577 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56578  int32_t __s0 = __p0; \
56579  int16_t __s1 = __p1; \
56580  int16x8_t __s2 = __p2; \
56581  int32_t __ret; \
56582  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56583  __ret; \
56584 })
56585 #else
56586 #define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56587  int32_t __s0 = __p0; \
56588  int16_t __s1 = __p1; \
56589  int16x8_t __s2 = __p2; \
56590  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56591  int32_t __ret; \
56592  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56593  __ret; \
56594 })
56595 #endif
56596 
56597 #ifdef __LITTLE_ENDIAN__
56598 #define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56599  int64x2_t __s0 = __p0; \
56600  int32x2_t __s1 = __p1; \
56601  int32x4_t __s2 = __p2; \
56602  int64x2_t __ret; \
56603  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56604  __ret; \
56605 })
56606 #else
56607 #define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56608  int64x2_t __s0 = __p0; \
56609  int32x2_t __s1 = __p1; \
56610  int32x4_t __s2 = __p2; \
56611  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56612  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56613  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56614  int64x2_t __ret; \
56615  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56616  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56617  __ret; \
56618 })
56619 #endif
56620 
56621 #ifdef __LITTLE_ENDIAN__
56622 #define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56623  int32x4_t __s0 = __p0; \
56624  int16x4_t __s1 = __p1; \
56625  int16x8_t __s2 = __p2; \
56626  int32x4_t __ret; \
56627  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56628  __ret; \
56629 })
56630 #else
56631 #define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56632  int32x4_t __s0 = __p0; \
56633  int16x4_t __s1 = __p1; \
56634  int16x8_t __s2 = __p2; \
56635  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56636  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56637  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56638  int32x4_t __ret; \
56639  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56641  __ret; \
56642 })
56643 #endif
56644 
56645 #ifdef __LITTLE_ENDIAN__
56646 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56647  int32_t __ret;
56648  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56649  return __ret;
56650 }
56651 #else
56652 __ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56653  int32_t __ret;
56654  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56655  return __ret;
56656 }
56657 __ai int32_t __noswap_vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56658  int32_t __ret;
56659  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56660  return __ret;
56661 }
56662 #endif
56663 
56664 #ifdef __LITTLE_ENDIAN__
56665 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56666  int16_t __ret;
56667  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56668  return __ret;
56669 }
56670 #else
56671 __ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56672  int16_t __ret;
56673  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56674  return __ret;
56675 }
56676 __ai int16_t __noswap_vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56677  int16_t __ret;
56678  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56679  return __ret;
56680 }
56681 #endif
56682 
56683 #ifdef __LITTLE_ENDIAN__
56684 #define vqdmulhs_lane_s32(__p0_146, __p1_146, __p2_146) __extension__ ({ \
56685  int32_t __s0_146 = __p0_146; \
56686  int32x2_t __s1_146 = __p1_146; \
56687  int32_t __ret_146; \
56688  __ret_146 = vqdmulhs_s32(__s0_146, vget_lane_s32(__s1_146, __p2_146)); \
56689  __ret_146; \
56690 })
56691 #else
56692 #define vqdmulhs_lane_s32(__p0_147, __p1_147, __p2_147) __extension__ ({ \
56693  int32_t __s0_147 = __p0_147; \
56694  int32x2_t __s1_147 = __p1_147; \
56695  int32x2_t __rev1_147; __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 1, 0); \
56696  int32_t __ret_147; \
56697  __ret_147 = __noswap_vqdmulhs_s32(__s0_147, __noswap_vget_lane_s32(__rev1_147, __p2_147)); \
56698  __ret_147; \
56699 })
56700 #endif
56701 
56702 #ifdef __LITTLE_ENDIAN__
56703 #define vqdmulhh_lane_s16(__p0_148, __p1_148, __p2_148) __extension__ ({ \
56704  int16_t __s0_148 = __p0_148; \
56705  int16x4_t __s1_148 = __p1_148; \
56706  int16_t __ret_148; \
56707  __ret_148 = vqdmulhh_s16(__s0_148, vget_lane_s16(__s1_148, __p2_148)); \
56708  __ret_148; \
56709 })
56710 #else
56711 #define vqdmulhh_lane_s16(__p0_149, __p1_149, __p2_149) __extension__ ({ \
56712  int16_t __s0_149 = __p0_149; \
56713  int16x4_t __s1_149 = __p1_149; \
56714  int16x4_t __rev1_149; __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
56715  int16_t __ret_149; \
56716  __ret_149 = __noswap_vqdmulhh_s16(__s0_149, __noswap_vget_lane_s16(__rev1_149, __p2_149)); \
56717  __ret_149; \
56718 })
56719 #endif
56720 
56721 #ifdef __LITTLE_ENDIAN__
56722 #define vqdmulhs_laneq_s32(__p0_150, __p1_150, __p2_150) __extension__ ({ \
56723  int32_t __s0_150 = __p0_150; \
56724  int32x4_t __s1_150 = __p1_150; \
56725  int32_t __ret_150; \
56726  __ret_150 = vqdmulhs_s32(__s0_150, vgetq_lane_s32(__s1_150, __p2_150)); \
56727  __ret_150; \
56728 })
56729 #else
56730 #define vqdmulhs_laneq_s32(__p0_151, __p1_151, __p2_151) __extension__ ({ \
56731  int32_t __s0_151 = __p0_151; \
56732  int32x4_t __s1_151 = __p1_151; \
56733  int32x4_t __rev1_151; __rev1_151 = __builtin_shufflevector(__s1_151, __s1_151, 3, 2, 1, 0); \
56734  int32_t __ret_151; \
56735  __ret_151 = __noswap_vqdmulhs_s32(__s0_151, __noswap_vgetq_lane_s32(__rev1_151, __p2_151)); \
56736  __ret_151; \
56737 })
56738 #endif
56739 
56740 #ifdef __LITTLE_ENDIAN__
56741 #define vqdmulhh_laneq_s16(__p0_152, __p1_152, __p2_152) __extension__ ({ \
56742  int16_t __s0_152 = __p0_152; \
56743  int16x8_t __s1_152 = __p1_152; \
56744  int16_t __ret_152; \
56745  __ret_152 = vqdmulhh_s16(__s0_152, vgetq_lane_s16(__s1_152, __p2_152)); \
56746  __ret_152; \
56747 })
56748 #else
56749 #define vqdmulhh_laneq_s16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
56750  int16_t __s0_153 = __p0_153; \
56751  int16x8_t __s1_153 = __p1_153; \
56752  int16x8_t __rev1_153; __rev1_153 = __builtin_shufflevector(__s1_153, __s1_153, 7, 6, 5, 4, 3, 2, 1, 0); \
56753  int16_t __ret_153; \
56754  __ret_153 = __noswap_vqdmulhh_s16(__s0_153, __noswap_vgetq_lane_s16(__rev1_153, __p2_153)); \
56755  __ret_153; \
56756 })
56757 #endif
56758 
56759 #ifdef __LITTLE_ENDIAN__
56760 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56761  int32x4_t __s0 = __p0; \
56762  int32x4_t __s1 = __p1; \
56763  int32x4_t __ret; \
56764  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56765  __ret; \
56766 })
56767 #else
56768 #define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56769  int32x4_t __s0 = __p0; \
56770  int32x4_t __s1 = __p1; \
56771  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56772  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56773  int32x4_t __ret; \
56774  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56775  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56776  __ret; \
56777 })
56778 #endif
56779 
56780 #ifdef __LITTLE_ENDIAN__
56781 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56782  int16x8_t __s0 = __p0; \
56783  int16x8_t __s1 = __p1; \
56784  int16x8_t __ret; \
56785  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56786  __ret; \
56787 })
56788 #else
56789 #define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56790  int16x8_t __s0 = __p0; \
56791  int16x8_t __s1 = __p1; \
56792  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56793  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56794  int16x8_t __ret; \
56795  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56796  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56797  __ret; \
56798 })
56799 #endif
56800 
56801 #ifdef __LITTLE_ENDIAN__
56802 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56803  int32x2_t __s0 = __p0; \
56804  int32x4_t __s1 = __p1; \
56805  int32x2_t __ret; \
56806  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56807  __ret; \
56808 })
56809 #else
56810 #define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56811  int32x2_t __s0 = __p0; \
56812  int32x4_t __s1 = __p1; \
56813  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56814  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56815  int32x2_t __ret; \
56816  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56818  __ret; \
56819 })
56820 #endif
56821 
56822 #ifdef __LITTLE_ENDIAN__
56823 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56824  int16x4_t __s0 = __p0; \
56825  int16x8_t __s1 = __p1; \
56826  int16x4_t __ret; \
56827  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56828  __ret; \
56829 })
56830 #else
56831 #define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56832  int16x4_t __s0 = __p0; \
56833  int16x8_t __s1 = __p1; \
56834  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56835  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56836  int16x4_t __ret; \
56837  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56839  __ret; \
56840 })
56841 #endif
56842 
56843 #ifdef __LITTLE_ENDIAN__
56844 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56845  int64_t __ret;
56846  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56847  return __ret;
56848 }
56849 #else
56850 __ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56851  int64_t __ret;
56852  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56853  return __ret;
56854 }
56855 __ai int64_t __noswap_vqdmulls_s32(int32_t __p0, int32_t __p1) {
56856  int64_t __ret;
56857  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56858  return __ret;
56859 }
56860 #endif
56861 
56862 #ifdef __LITTLE_ENDIAN__
56863 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56864  int32_t __ret;
56865  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56866  return __ret;
56867 }
56868 #else
56869 __ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56870  int32_t __ret;
56871  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56872  return __ret;
56873 }
56874 __ai int32_t __noswap_vqdmullh_s16(int16_t __p0, int16_t __p1) {
56875  int32_t __ret;
56876  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56877  return __ret;
56878 }
56879 #endif
56880 
56881 #ifdef __LITTLE_ENDIAN__
56882 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56883  int64x2_t __ret;
56884  __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
56885  return __ret;
56886 }
56887 #else
56888 __ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56889  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56890  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56891  int64x2_t __ret;
56893  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56894  return __ret;
56895 }
56896 #endif
56897 
56898 #ifdef __LITTLE_ENDIAN__
56899 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56900  int32x4_t __ret;
56901  __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
56902  return __ret;
56903 }
56904 #else
56905 __ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56906  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56907  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56908  int32x4_t __ret;
56910  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56911  return __ret;
56912 }
56913 #endif
56914 
56915 #ifdef __LITTLE_ENDIAN__
56916 #define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56917  int32x4_t __s0 = __p0; \
56918  int32x2_t __s1 = __p1; \
56919  int64x2_t __ret; \
56920  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56921  __ret; \
56922 })
56923 #else
56924 #define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56925  int32x4_t __s0 = __p0; \
56926  int32x2_t __s1 = __p1; \
56927  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56928  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56929  int64x2_t __ret; \
56930  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56931  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56932  __ret; \
56933 })
56934 #endif
56935 
56936 #ifdef __LITTLE_ENDIAN__
56937 #define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56938  int16x8_t __s0 = __p0; \
56939  int16x4_t __s1 = __p1; \
56940  int32x4_t __ret; \
56941  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56942  __ret; \
56943 })
56944 #else
56945 #define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56946  int16x8_t __s0 = __p0; \
56947  int16x4_t __s1 = __p1; \
56948  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56949  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56950  int32x4_t __ret; \
56951  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56952  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56953  __ret; \
56954 })
56955 #endif
56956 
56957 #ifdef __LITTLE_ENDIAN__
56958 #define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56959  int32x4_t __s0 = __p0; \
56960  int32x4_t __s1 = __p1; \
56961  int64x2_t __ret; \
56962  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56963  __ret; \
56964 })
56965 #else
56966 #define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56967  int32x4_t __s0 = __p0; \
56968  int32x4_t __s1 = __p1; \
56969  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56970  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56971  int64x2_t __ret; \
56972  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56974  __ret; \
56975 })
56976 #endif
56977 
56978 #ifdef __LITTLE_ENDIAN__
56979 #define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56980  int16x8_t __s0 = __p0; \
56981  int16x8_t __s1 = __p1; \
56982  int32x4_t __ret; \
56983  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56984  __ret; \
56985 })
56986 #else
56987 #define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56988  int16x8_t __s0 = __p0; \
56989  int16x8_t __s1 = __p1; \
56990  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56991  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56992  int32x4_t __ret; \
56993  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56995  __ret; \
56996 })
56997 #endif
56998 
56999 #ifdef __LITTLE_ENDIAN__
57000 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
57001  int64x2_t __ret;
57002  __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
57003  return __ret;
57004 }
57005 #else
57006 __ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
57007  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57008  int64x2_t __ret;
57009  __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
57010  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57011  return __ret;
57012 }
57013 #endif
57014 
57015 #ifdef __LITTLE_ENDIAN__
57016 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57017  int32x4_t __ret;
57018  __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
57019  return __ret;
57020 }
57021 #else
57022 __ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57023  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57024  int32x4_t __ret;
57025  __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
57026  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57027  return __ret;
57028 }
57029 #endif
57030 
57031 #ifdef __LITTLE_ENDIAN__
57032 #define vqdmulls_lane_s32(__p0_154, __p1_154, __p2_154) __extension__ ({ \
57033  int32_t __s0_154 = __p0_154; \
57034  int32x2_t __s1_154 = __p1_154; \
57035  int64_t __ret_154; \
57036  __ret_154 = vqdmulls_s32(__s0_154, vget_lane_s32(__s1_154, __p2_154)); \
57037  __ret_154; \
57038 })
57039 #else
57040 #define vqdmulls_lane_s32(__p0_155, __p1_155, __p2_155) __extension__ ({ \
57041  int32_t __s0_155 = __p0_155; \
57042  int32x2_t __s1_155 = __p1_155; \
57043  int32x2_t __rev1_155; __rev1_155 = __builtin_shufflevector(__s1_155, __s1_155, 1, 0); \
57044  int64_t __ret_155; \
57045  __ret_155 = __noswap_vqdmulls_s32(__s0_155, __noswap_vget_lane_s32(__rev1_155, __p2_155)); \
57046  __ret_155; \
57047 })
57048 #endif
57049 
57050 #ifdef __LITTLE_ENDIAN__
57051 #define vqdmullh_lane_s16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
57052  int16_t __s0_156 = __p0_156; \
57053  int16x4_t __s1_156 = __p1_156; \
57054  int32_t __ret_156; \
57055  __ret_156 = vqdmullh_s16(__s0_156, vget_lane_s16(__s1_156, __p2_156)); \
57056  __ret_156; \
57057 })
57058 #else
57059 #define vqdmullh_lane_s16(__p0_157, __p1_157, __p2_157) __extension__ ({ \
57060  int16_t __s0_157 = __p0_157; \
57061  int16x4_t __s1_157 = __p1_157; \
57062  int16x4_t __rev1_157; __rev1_157 = __builtin_shufflevector(__s1_157, __s1_157, 3, 2, 1, 0); \
57063  int32_t __ret_157; \
57064  __ret_157 = __noswap_vqdmullh_s16(__s0_157, __noswap_vget_lane_s16(__rev1_157, __p2_157)); \
57065  __ret_157; \
57066 })
57067 #endif
57068 
57069 #ifdef __LITTLE_ENDIAN__
57070 #define vqdmulls_laneq_s32(__p0_158, __p1_158, __p2_158) __extension__ ({ \
57071  int32_t __s0_158 = __p0_158; \
57072  int32x4_t __s1_158 = __p1_158; \
57073  int64_t __ret_158; \
57074  __ret_158 = vqdmulls_s32(__s0_158, vgetq_lane_s32(__s1_158, __p2_158)); \
57075  __ret_158; \
57076 })
57077 #else
57078 #define vqdmulls_laneq_s32(__p0_159, __p1_159, __p2_159) __extension__ ({ \
57079  int32_t __s0_159 = __p0_159; \
57080  int32x4_t __s1_159 = __p1_159; \
57081  int32x4_t __rev1_159; __rev1_159 = __builtin_shufflevector(__s1_159, __s1_159, 3, 2, 1, 0); \
57082  int64_t __ret_159; \
57083  __ret_159 = __noswap_vqdmulls_s32(__s0_159, __noswap_vgetq_lane_s32(__rev1_159, __p2_159)); \
57084  __ret_159; \
57085 })
57086 #endif
57087 
57088 #ifdef __LITTLE_ENDIAN__
57089 #define vqdmullh_laneq_s16(__p0_160, __p1_160, __p2_160) __extension__ ({ \
57090  int16_t __s0_160 = __p0_160; \
57091  int16x8_t __s1_160 = __p1_160; \
57092  int32_t __ret_160; \
57093  __ret_160 = vqdmullh_s16(__s0_160, vgetq_lane_s16(__s1_160, __p2_160)); \
57094  __ret_160; \
57095 })
57096 #else
57097 #define vqdmullh_laneq_s16(__p0_161, __p1_161, __p2_161) __extension__ ({ \
57098  int16_t __s0_161 = __p0_161; \
57099  int16x8_t __s1_161 = __p1_161; \
57100  int16x8_t __rev1_161; __rev1_161 = __builtin_shufflevector(__s1_161, __s1_161, 7, 6, 5, 4, 3, 2, 1, 0); \
57101  int32_t __ret_161; \
57102  __ret_161 = __noswap_vqdmullh_s16(__s0_161, __noswap_vgetq_lane_s16(__rev1_161, __p2_161)); \
57103  __ret_161; \
57104 })
57105 #endif
57106 
57107 #ifdef __LITTLE_ENDIAN__
57108 #define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57109  int32x2_t __s0 = __p0; \
57110  int32x4_t __s1 = __p1; \
57111  int64x2_t __ret; \
57112  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57113  __ret; \
57114 })
57115 #else
57116 #define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57117  int32x2_t __s0 = __p0; \
57118  int32x4_t __s1 = __p1; \
57119  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57120  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57121  int64x2_t __ret; \
57122  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57123  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57124  __ret; \
57125 })
57126 #endif
57127 
57128 #ifdef __LITTLE_ENDIAN__
57129 #define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57130  int16x4_t __s0 = __p0; \
57131  int16x8_t __s1 = __p1; \
57132  int32x4_t __ret; \
57133  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57134  __ret; \
57135 })
57136 #else
57137 #define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57138  int16x4_t __s0 = __p0; \
57139  int16x8_t __s1 = __p1; \
57140  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57141  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57142  int32x4_t __ret; \
57143  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57144  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57145  __ret; \
57146 })
57147 #endif
57148 
57149 #ifdef __LITTLE_ENDIAN__
57150 __ai int16_t vqmovns_s32(int32_t __p0) {
57151  int16_t __ret;
57152  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57153  return __ret;
57154 }
57155 #else
57156 __ai int16_t vqmovns_s32(int32_t __p0) {
57157  int16_t __ret;
57158  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57159  return __ret;
57160 }
57161 #endif
57162 
57163 #ifdef __LITTLE_ENDIAN__
57164 __ai int32_t vqmovnd_s64(int64_t __p0) {
57165  int32_t __ret;
57166  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57167  return __ret;
57168 }
57169 #else
57170 __ai int32_t vqmovnd_s64(int64_t __p0) {
57171  int32_t __ret;
57172  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57173  return __ret;
57174 }
57175 #endif
57176 
57177 #ifdef __LITTLE_ENDIAN__
57178 __ai int8_t vqmovnh_s16(int16_t __p0) {
57179  int8_t __ret;
57180  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57181  return __ret;
57182 }
57183 #else
57184 __ai int8_t vqmovnh_s16(int16_t __p0) {
57185  int8_t __ret;
57186  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57187  return __ret;
57188 }
57189 #endif
57190 
57191 #ifdef __LITTLE_ENDIAN__
57192 __ai uint16_t vqmovns_u32(uint32_t __p0) {
57193  uint16_t __ret;
57194  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57195  return __ret;
57196 }
57197 #else
57198 __ai uint16_t vqmovns_u32(uint32_t __p0) {
57199  uint16_t __ret;
57200  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57201  return __ret;
57202 }
57203 #endif
57204 
57205 #ifdef __LITTLE_ENDIAN__
57206 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
57207  uint32_t __ret;
57208  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57209  return __ret;
57210 }
57211 #else
57212 __ai uint32_t vqmovnd_u64(uint64_t __p0) {
57213  uint32_t __ret;
57214  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57215  return __ret;
57216 }
57217 #endif
57218 
57219 #ifdef __LITTLE_ENDIAN__
57220 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
57221  uint8_t __ret;
57222  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57223  return __ret;
57224 }
57225 #else
57226 __ai uint8_t vqmovnh_u16(uint16_t __p0) {
57227  uint8_t __ret;
57228  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57229  return __ret;
57230 }
57231 #endif
57232 
57233 #ifdef __LITTLE_ENDIAN__
57234 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57235  uint16x8_t __ret;
57236  __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
57237  return __ret;
57238 }
57239 #else
57240 __ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57241  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57242  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57243  uint16x8_t __ret;
57244  __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
57245  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57246  return __ret;
57247 }
57248 #endif
57249 
57250 #ifdef __LITTLE_ENDIAN__
57251 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57252  uint32x4_t __ret;
57253  __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
57254  return __ret;
57255 }
57256 #else
57257 __ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57258  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57259  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57260  uint32x4_t __ret;
57261  __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
57262  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57263  return __ret;
57264 }
57265 #endif
57266 
57267 #ifdef __LITTLE_ENDIAN__
57268 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57269  uint8x16_t __ret;
57270  __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
57271  return __ret;
57272 }
57273 #else
57274 __ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57275  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57276  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57277  uint8x16_t __ret;
57278  __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
57279  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57280  return __ret;
57281 }
57282 #endif
57283 
57284 #ifdef __LITTLE_ENDIAN__
57285 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57286  int16x8_t __ret;
57287  __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
57288  return __ret;
57289 }
57290 #else
57291 __ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57292  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57293  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57294  int16x8_t __ret;
57295  __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
57296  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57297  return __ret;
57298 }
57299 #endif
57300 
57301 #ifdef __LITTLE_ENDIAN__
57302 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57303  int32x4_t __ret;
57304  __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
57305  return __ret;
57306 }
57307 #else
57308 __ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57309  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57310  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57311  int32x4_t __ret;
57312  __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
57313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57314  return __ret;
57315 }
57316 #endif
57317 
57318 #ifdef __LITTLE_ENDIAN__
57319 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57320  int8x16_t __ret;
57321  __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
57322  return __ret;
57323 }
57324 #else
57325 __ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57326  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57327  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57328  int8x16_t __ret;
57329  __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
57330  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57331  return __ret;
57332 }
57333 #endif
57334 
57335 #ifdef __LITTLE_ENDIAN__
57336 __ai int16_t vqmovuns_s32(int32_t __p0) {
57337  int16_t __ret;
57338  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57339  return __ret;
57340 }
57341 #else
57342 __ai int16_t vqmovuns_s32(int32_t __p0) {
57343  int16_t __ret;
57344  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57345  return __ret;
57346 }
57347 #endif
57348 
57349 #ifdef __LITTLE_ENDIAN__
57350 __ai int32_t vqmovund_s64(int64_t __p0) {
57351  int32_t __ret;
57352  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57353  return __ret;
57354 }
57355 #else
57356 __ai int32_t vqmovund_s64(int64_t __p0) {
57357  int32_t __ret;
57358  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57359  return __ret;
57360 }
57361 #endif
57362 
57363 #ifdef __LITTLE_ENDIAN__
57364 __ai int8_t vqmovunh_s16(int16_t __p0) {
57365  int8_t __ret;
57366  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57367  return __ret;
57368 }
57369 #else
57370 __ai int8_t vqmovunh_s16(int16_t __p0) {
57371  int8_t __ret;
57372  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57373  return __ret;
57374 }
57375 #endif
57376 
57377 #ifdef __LITTLE_ENDIAN__
57378 __ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57379  int16x8_t __ret;
57380  __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
57381  return __ret;
57382 }
57383 #else
57384 __ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57385  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57386  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57387  int16x8_t __ret;
57388  __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
57389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57390  return __ret;
57391 }
57392 #endif
57393 
57394 #ifdef __LITTLE_ENDIAN__
57395 __ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57396  int32x4_t __ret;
57397  __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
57398  return __ret;
57399 }
57400 #else
57401 __ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57402  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57403  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57404  int32x4_t __ret;
57405  __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
57406  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57407  return __ret;
57408 }
57409 #endif
57410 
57411 #ifdef __LITTLE_ENDIAN__
57412 __ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57413  int8x16_t __ret;
57414  __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
57415  return __ret;
57416 }
57417 #else
57418 __ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57419  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57420  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57421  int8x16_t __ret;
57422  __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
57423  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57424  return __ret;
57425 }
57426 #endif
57427 
57428 #ifdef __LITTLE_ENDIAN__
57429 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57430  int64x2_t __ret;
57431  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
57432  return __ret;
57433 }
57434 #else
57435 __ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57436  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57437  int64x2_t __ret;
57438  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
57439  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57440  return __ret;
57441 }
57442 #endif
57443 
57444 #ifdef __LITTLE_ENDIAN__
57445 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
57446  int64x1_t __ret;
57447  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57448  return __ret;
57449 }
57450 #else
57451 __ai int64x1_t vqneg_s64(int64x1_t __p0) {
57452  int64x1_t __ret;
57453  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57454  return __ret;
57455 }
57456 #endif
57457 
57458 #ifdef __LITTLE_ENDIAN__
57459 __ai int8_t vqnegb_s8(int8_t __p0) {
57460  int8_t __ret;
57461  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57462  return __ret;
57463 }
57464 #else
57465 __ai int8_t vqnegb_s8(int8_t __p0) {
57466  int8_t __ret;
57467  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57468  return __ret;
57469 }
57470 #endif
57471 
57472 #ifdef __LITTLE_ENDIAN__
57473 __ai int32_t vqnegs_s32(int32_t __p0) {
57474  int32_t __ret;
57475  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57476  return __ret;
57477 }
57478 #else
57479 __ai int32_t vqnegs_s32(int32_t __p0) {
57480  int32_t __ret;
57481  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57482  return __ret;
57483 }
57484 #endif
57485 
57486 #ifdef __LITTLE_ENDIAN__
57487 __ai int64_t vqnegd_s64(int64_t __p0) {
57488  int64_t __ret;
57489  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57490  return __ret;
57491 }
57492 #else
57493 __ai int64_t vqnegd_s64(int64_t __p0) {
57494  int64_t __ret;
57495  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57496  return __ret;
57497 }
57498 #endif
57499 
57500 #ifdef __LITTLE_ENDIAN__
57501 __ai int16_t vqnegh_s16(int16_t __p0) {
57502  int16_t __ret;
57503  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57504  return __ret;
57505 }
57506 #else
57507 __ai int16_t vqnegh_s16(int16_t __p0) {
57508  int16_t __ret;
57509  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57510  return __ret;
57511 }
57512 #endif
57513 
57514 #ifdef __LITTLE_ENDIAN__
57515 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57516  int32_t __ret;
57517  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57518  return __ret;
57519 }
57520 #else
57521 __ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57522  int32_t __ret;
57523  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57524  return __ret;
57525 }
57526 __ai int32_t __noswap_vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57527  int32_t __ret;
57528  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57529  return __ret;
57530 }
57531 #endif
57532 
57533 #ifdef __LITTLE_ENDIAN__
57534 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57535  int16_t __ret;
57536  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57537  return __ret;
57538 }
57539 #else
57540 __ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57541  int16_t __ret;
57542  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57543  return __ret;
57544 }
57545 __ai int16_t __noswap_vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57546  int16_t __ret;
57547  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57548  return __ret;
57549 }
57550 #endif
57551 
57552 #ifdef __LITTLE_ENDIAN__
57553 #define vqrdmulhs_lane_s32(__p0_162, __p1_162, __p2_162) __extension__ ({ \
57554  int32_t __s0_162 = __p0_162; \
57555  int32x2_t __s1_162 = __p1_162; \
57556  int32_t __ret_162; \
57557  __ret_162 = vqrdmulhs_s32(__s0_162, vget_lane_s32(__s1_162, __p2_162)); \
57558  __ret_162; \
57559 })
57560 #else
57561 #define vqrdmulhs_lane_s32(__p0_163, __p1_163, __p2_163) __extension__ ({ \
57562  int32_t __s0_163 = __p0_163; \
57563  int32x2_t __s1_163 = __p1_163; \
57564  int32x2_t __rev1_163; __rev1_163 = __builtin_shufflevector(__s1_163, __s1_163, 1, 0); \
57565  int32_t __ret_163; \
57566  __ret_163 = __noswap_vqrdmulhs_s32(__s0_163, __noswap_vget_lane_s32(__rev1_163, __p2_163)); \
57567  __ret_163; \
57568 })
57569 #endif
57570 
57571 #ifdef __LITTLE_ENDIAN__
57572 #define vqrdmulhh_lane_s16(__p0_164, __p1_164, __p2_164) __extension__ ({ \
57573  int16_t __s0_164 = __p0_164; \
57574  int16x4_t __s1_164 = __p1_164; \
57575  int16_t __ret_164; \
57576  __ret_164 = vqrdmulhh_s16(__s0_164, vget_lane_s16(__s1_164, __p2_164)); \
57577  __ret_164; \
57578 })
57579 #else
57580 #define vqrdmulhh_lane_s16(__p0_165, __p1_165, __p2_165) __extension__ ({ \
57581  int16_t __s0_165 = __p0_165; \
57582  int16x4_t __s1_165 = __p1_165; \
57583  int16x4_t __rev1_165; __rev1_165 = __builtin_shufflevector(__s1_165, __s1_165, 3, 2, 1, 0); \
57584  int16_t __ret_165; \
57585  __ret_165 = __noswap_vqrdmulhh_s16(__s0_165, __noswap_vget_lane_s16(__rev1_165, __p2_165)); \
57586  __ret_165; \
57587 })
57588 #endif
57589 
57590 #ifdef __LITTLE_ENDIAN__
57591 #define vqrdmulhs_laneq_s32(__p0_166, __p1_166, __p2_166) __extension__ ({ \
57592  int32_t __s0_166 = __p0_166; \
57593  int32x4_t __s1_166 = __p1_166; \
57594  int32_t __ret_166; \
57595  __ret_166 = vqrdmulhs_s32(__s0_166, vgetq_lane_s32(__s1_166, __p2_166)); \
57596  __ret_166; \
57597 })
57598 #else
57599 #define vqrdmulhs_laneq_s32(__p0_167, __p1_167, __p2_167) __extension__ ({ \
57600  int32_t __s0_167 = __p0_167; \
57601  int32x4_t __s1_167 = __p1_167; \
57602  int32x4_t __rev1_167; __rev1_167 = __builtin_shufflevector(__s1_167, __s1_167, 3, 2, 1, 0); \
57603  int32_t __ret_167; \
57604  __ret_167 = __noswap_vqrdmulhs_s32(__s0_167, __noswap_vgetq_lane_s32(__rev1_167, __p2_167)); \
57605  __ret_167; \
57606 })
57607 #endif
57608 
57609 #ifdef __LITTLE_ENDIAN__
57610 #define vqrdmulhh_laneq_s16(__p0_168, __p1_168, __p2_168) __extension__ ({ \
57611  int16_t __s0_168 = __p0_168; \
57612  int16x8_t __s1_168 = __p1_168; \
57613  int16_t __ret_168; \
57614  __ret_168 = vqrdmulhh_s16(__s0_168, vgetq_lane_s16(__s1_168, __p2_168)); \
57615  __ret_168; \
57616 })
57617 #else
57618 #define vqrdmulhh_laneq_s16(__p0_169, __p1_169, __p2_169) __extension__ ({ \
57619  int16_t __s0_169 = __p0_169; \
57620  int16x8_t __s1_169 = __p1_169; \
57621  int16x8_t __rev1_169; __rev1_169 = __builtin_shufflevector(__s1_169, __s1_169, 7, 6, 5, 4, 3, 2, 1, 0); \
57622  int16_t __ret_169; \
57623  __ret_169 = __noswap_vqrdmulhh_s16(__s0_169, __noswap_vgetq_lane_s16(__rev1_169, __p2_169)); \
57624  __ret_169; \
57625 })
57626 #endif
57627 
57628 #ifdef __LITTLE_ENDIAN__
57629 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57630  int32x4_t __s0 = __p0; \
57631  int32x4_t __s1 = __p1; \
57632  int32x4_t __ret; \
57633  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57634  __ret; \
57635 })
57636 #else
57637 #define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57638  int32x4_t __s0 = __p0; \
57639  int32x4_t __s1 = __p1; \
57640  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57641  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57642  int32x4_t __ret; \
57643  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57645  __ret; \
57646 })
57647 #endif
57648 
57649 #ifdef __LITTLE_ENDIAN__
57650 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57651  int16x8_t __s0 = __p0; \
57652  int16x8_t __s1 = __p1; \
57653  int16x8_t __ret; \
57654  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57655  __ret; \
57656 })
57657 #else
57658 #define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57659  int16x8_t __s0 = __p0; \
57660  int16x8_t __s1 = __p1; \
57661  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
57662  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57663  int16x8_t __ret; \
57664  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57665  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
57666  __ret; \
57667 })
57668 #endif
57669 
57670 #ifdef __LITTLE_ENDIAN__
57671 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57672  int32x2_t __s0 = __p0; \
57673  int32x4_t __s1 = __p1; \
57674  int32x2_t __ret; \
57675  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57676  __ret; \
57677 })
57678 #else
57679 #define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57680  int32x2_t __s0 = __p0; \
57681  int32x4_t __s1 = __p1; \
57682  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57683  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57684  int32x2_t __ret; \
57685  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57686  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57687  __ret; \
57688 })
57689 #endif
57690 
57691 #ifdef __LITTLE_ENDIAN__
57692 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57693  int16x4_t __s0 = __p0; \
57694  int16x8_t __s1 = __p1; \
57695  int16x4_t __ret; \
57696  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57697  __ret; \
57698 })
57699 #else
57700 #define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57701  int16x4_t __s0 = __p0; \
57702  int16x8_t __s1 = __p1; \
57703  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57704  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57705  int16x4_t __ret; \
57706  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57707  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57708  __ret; \
57709 })
57710 #endif
57711 
57712 #ifdef __LITTLE_ENDIAN__
57713 __ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57714  uint8_t __ret;
57715  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57716  return __ret;
57717 }
57718 #else
57719 __ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57720  uint8_t __ret;
57721  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57722  return __ret;
57723 }
57724 #endif
57725 
57726 #ifdef __LITTLE_ENDIAN__
57727 __ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57728  uint32_t __ret;
57729  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57730  return __ret;
57731 }
57732 #else
57733 __ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57734  uint32_t __ret;
57735  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57736  return __ret;
57737 }
57738 #endif
57739 
57740 #ifdef __LITTLE_ENDIAN__
57741 __ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57742  uint64_t __ret;
57743  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57744  return __ret;
57745 }
57746 #else
57747 __ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57748  uint64_t __ret;
57749  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57750  return __ret;
57751 }
57752 #endif
57753 
57754 #ifdef __LITTLE_ENDIAN__
57755 __ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57756  uint16_t __ret;
57757  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57758  return __ret;
57759 }
57760 #else
57761 __ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57762  uint16_t __ret;
57763  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57764  return __ret;
57765 }
57766 #endif
57767 
57768 #ifdef __LITTLE_ENDIAN__
57769 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57770  int8_t __ret;
57771  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57772  return __ret;
57773 }
57774 #else
57775 __ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57776  int8_t __ret;
57777  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57778  return __ret;
57779 }
57780 #endif
57781 
57782 #ifdef __LITTLE_ENDIAN__
57783 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57784  int32_t __ret;
57785  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57786  return __ret;
57787 }
57788 #else
57789 __ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57790  int32_t __ret;
57791  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57792  return __ret;
57793 }
57794 #endif
57795 
57796 #ifdef __LITTLE_ENDIAN__
57797 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57798  int64_t __ret;
57799  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57800  return __ret;
57801 }
57802 #else
57803 __ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57804  int64_t __ret;
57805  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57806  return __ret;
57807 }
57808 #endif
57809 
57810 #ifdef __LITTLE_ENDIAN__
57811 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57812  int16_t __ret;
57813  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57814  return __ret;
57815 }
57816 #else
57817 __ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57818  int16_t __ret;
57819  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57820  return __ret;
57821 }
57822 #endif
57823 
57824 #ifdef __LITTLE_ENDIAN__
57825 #define vqrshrn_high_n_u32(__p0_170, __p1_170, __p2_170) __extension__ ({ \
57826  uint16x4_t __s0_170 = __p0_170; \
57827  uint32x4_t __s1_170 = __p1_170; \
57828  uint16x8_t __ret_170; \
57829  __ret_170 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_170), (uint16x4_t)(vqrshrn_n_u32(__s1_170, __p2_170)))); \
57830  __ret_170; \
57831 })
57832 #else
57833 #define vqrshrn_high_n_u32(__p0_171, __p1_171, __p2_171) __extension__ ({ \
57834  uint16x4_t __s0_171 = __p0_171; \
57835  uint32x4_t __s1_171 = __p1_171; \
57836  uint16x4_t __rev0_171; __rev0_171 = __builtin_shufflevector(__s0_171, __s0_171, 3, 2, 1, 0); \
57837  uint32x4_t __rev1_171; __rev1_171 = __builtin_shufflevector(__s1_171, __s1_171, 3, 2, 1, 0); \
57838  uint16x8_t __ret_171; \
57839  __ret_171 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_171), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_171, __p2_171)))); \
57840  __ret_171 = __builtin_shufflevector(__ret_171, __ret_171, 7, 6, 5, 4, 3, 2, 1, 0); \
57841  __ret_171; \
57842 })
57843 #endif
57844 
57845 #ifdef __LITTLE_ENDIAN__
57846 #define vqrshrn_high_n_u64(__p0_172, __p1_172, __p2_172) __extension__ ({ \
57847  uint32x2_t __s0_172 = __p0_172; \
57848  uint64x2_t __s1_172 = __p1_172; \
57849  uint32x4_t __ret_172; \
57850  __ret_172 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_172), (uint32x2_t)(vqrshrn_n_u64(__s1_172, __p2_172)))); \
57851  __ret_172; \
57852 })
57853 #else
57854 #define vqrshrn_high_n_u64(__p0_173, __p1_173, __p2_173) __extension__ ({ \
57855  uint32x2_t __s0_173 = __p0_173; \
57856  uint64x2_t __s1_173 = __p1_173; \
57857  uint32x2_t __rev0_173; __rev0_173 = __builtin_shufflevector(__s0_173, __s0_173, 1, 0); \
57858  uint64x2_t __rev1_173; __rev1_173 = __builtin_shufflevector(__s1_173, __s1_173, 1, 0); \
57859  uint32x4_t __ret_173; \
57860  __ret_173 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_173), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_173, __p2_173)))); \
57861  __ret_173 = __builtin_shufflevector(__ret_173, __ret_173, 3, 2, 1, 0); \
57862  __ret_173; \
57863 })
57864 #endif
57865 
57866 #ifdef __LITTLE_ENDIAN__
57867 #define vqrshrn_high_n_u16(__p0_174, __p1_174, __p2_174) __extension__ ({ \
57868  uint8x8_t __s0_174 = __p0_174; \
57869  uint16x8_t __s1_174 = __p1_174; \
57870  uint8x16_t __ret_174; \
57871  __ret_174 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_174), (uint8x8_t)(vqrshrn_n_u16(__s1_174, __p2_174)))); \
57872  __ret_174; \
57873 })
57874 #else
57875 #define vqrshrn_high_n_u16(__p0_175, __p1_175, __p2_175) __extension__ ({ \
57876  uint8x8_t __s0_175 = __p0_175; \
57877  uint16x8_t __s1_175 = __p1_175; \
57878  uint8x8_t __rev0_175; __rev0_175 = __builtin_shufflevector(__s0_175, __s0_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57879  uint16x8_t __rev1_175; __rev1_175 = __builtin_shufflevector(__s1_175, __s1_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57880  uint8x16_t __ret_175; \
57881  __ret_175 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_175), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_175, __p2_175)))); \
57882  __ret_175 = __builtin_shufflevector(__ret_175, __ret_175, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57883  __ret_175; \
57884 })
57885 #endif
57886 
57887 #ifdef __LITTLE_ENDIAN__
57888 #define vqrshrn_high_n_s32(__p0_176, __p1_176, __p2_176) __extension__ ({ \
57889  int16x4_t __s0_176 = __p0_176; \
57890  int32x4_t __s1_176 = __p1_176; \
57891  int16x8_t __ret_176; \
57892  __ret_176 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_176), (int16x4_t)(vqrshrn_n_s32(__s1_176, __p2_176)))); \
57893  __ret_176; \
57894 })
57895 #else
57896 #define vqrshrn_high_n_s32(__p0_177, __p1_177, __p2_177) __extension__ ({ \
57897  int16x4_t __s0_177 = __p0_177; \
57898  int32x4_t __s1_177 = __p1_177; \
57899  int16x4_t __rev0_177; __rev0_177 = __builtin_shufflevector(__s0_177, __s0_177, 3, 2, 1, 0); \
57900  int32x4_t __rev1_177; __rev1_177 = __builtin_shufflevector(__s1_177, __s1_177, 3, 2, 1, 0); \
57901  int16x8_t __ret_177; \
57902  __ret_177 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_177), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_177, __p2_177)))); \
57903  __ret_177 = __builtin_shufflevector(__ret_177, __ret_177, 7, 6, 5, 4, 3, 2, 1, 0); \
57904  __ret_177; \
57905 })
57906 #endif
57907 
57908 #ifdef __LITTLE_ENDIAN__
57909 #define vqrshrn_high_n_s64(__p0_178, __p1_178, __p2_178) __extension__ ({ \
57910  int32x2_t __s0_178 = __p0_178; \
57911  int64x2_t __s1_178 = __p1_178; \
57912  int32x4_t __ret_178; \
57913  __ret_178 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_178), (int32x2_t)(vqrshrn_n_s64(__s1_178, __p2_178)))); \
57914  __ret_178; \
57915 })
57916 #else
57917 #define vqrshrn_high_n_s64(__p0_179, __p1_179, __p2_179) __extension__ ({ \
57918  int32x2_t __s0_179 = __p0_179; \
57919  int64x2_t __s1_179 = __p1_179; \
57920  int32x2_t __rev0_179; __rev0_179 = __builtin_shufflevector(__s0_179, __s0_179, 1, 0); \
57921  int64x2_t __rev1_179; __rev1_179 = __builtin_shufflevector(__s1_179, __s1_179, 1, 0); \
57922  int32x4_t __ret_179; \
57923  __ret_179 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_179), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_179, __p2_179)))); \
57924  __ret_179 = __builtin_shufflevector(__ret_179, __ret_179, 3, 2, 1, 0); \
57925  __ret_179; \
57926 })
57927 #endif
57928 
57929 #ifdef __LITTLE_ENDIAN__
57930 #define vqrshrn_high_n_s16(__p0_180, __p1_180, __p2_180) __extension__ ({ \
57931  int8x8_t __s0_180 = __p0_180; \
57932  int16x8_t __s1_180 = __p1_180; \
57933  int8x16_t __ret_180; \
57934  __ret_180 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_180), (int8x8_t)(vqrshrn_n_s16(__s1_180, __p2_180)))); \
57935  __ret_180; \
57936 })
57937 #else
57938 #define vqrshrn_high_n_s16(__p0_181, __p1_181, __p2_181) __extension__ ({ \
57939  int8x8_t __s0_181 = __p0_181; \
57940  int16x8_t __s1_181 = __p1_181; \
57941  int8x8_t __rev0_181; __rev0_181 = __builtin_shufflevector(__s0_181, __s0_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57942  int16x8_t __rev1_181; __rev1_181 = __builtin_shufflevector(__s1_181, __s1_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57943  int8x16_t __ret_181; \
57944  __ret_181 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_181), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_181, __p2_181)))); \
57945  __ret_181 = __builtin_shufflevector(__ret_181, __ret_181, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57946  __ret_181; \
57947 })
57948 #endif
57949 
57950 #ifdef __LITTLE_ENDIAN__
57951 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57952  uint32_t __s0 = __p0; \
57953  uint16_t __ret; \
57954  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57955  __ret; \
57956 })
57957 #else
57958 #define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57959  uint32_t __s0 = __p0; \
57960  uint16_t __ret; \
57961  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57962  __ret; \
57963 })
57964 #endif
57965 
57966 #ifdef __LITTLE_ENDIAN__
57967 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57968  uint64_t __s0 = __p0; \
57969  uint32_t __ret; \
57970  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57971  __ret; \
57972 })
57973 #else
57974 #define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57975  uint64_t __s0 = __p0; \
57976  uint32_t __ret; \
57977  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57978  __ret; \
57979 })
57980 #endif
57981 
57982 #ifdef __LITTLE_ENDIAN__
57983 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57984  uint16_t __s0 = __p0; \
57985  uint8_t __ret; \
57986  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57987  __ret; \
57988 })
57989 #else
57990 #define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57991  uint16_t __s0 = __p0; \
57992  uint8_t __ret; \
57993  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57994  __ret; \
57995 })
57996 #endif
57997 
57998 #ifdef __LITTLE_ENDIAN__
57999 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
58000  int32_t __s0 = __p0; \
58001  int16_t __ret; \
58002  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
58003  __ret; \
58004 })
58005 #else
58006 #define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
58007  int32_t __s0 = __p0; \
58008  int16_t __ret; \
58009  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
58010  __ret; \
58011 })
58012 #endif
58013 
58014 #ifdef __LITTLE_ENDIAN__
58015 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58016  int64_t __s0 = __p0; \
58017  int32_t __ret; \
58018  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58019  __ret; \
58020 })
58021 #else
58022 #define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58023  int64_t __s0 = __p0; \
58024  int32_t __ret; \
58025  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58026  __ret; \
58027 })
58028 #endif
58029 
58030 #ifdef __LITTLE_ENDIAN__
58031 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58032  int16_t __s0 = __p0; \
58033  int8_t __ret; \
58034  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58035  __ret; \
58036 })
58037 #else
58038 #define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58039  int16_t __s0 = __p0; \
58040  int8_t __ret; \
58041  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58042  __ret; \
58043 })
58044 #endif
58045 
58046 #ifdef __LITTLE_ENDIAN__
58047 #define vqrshrun_high_n_s32(__p0_182, __p1_182, __p2_182) __extension__ ({ \
58048  int16x4_t __s0_182 = __p0_182; \
58049  int32x4_t __s1_182 = __p1_182; \
58050  int16x8_t __ret_182; \
58051  __ret_182 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_182), (int16x4_t)(vqrshrun_n_s32(__s1_182, __p2_182)))); \
58052  __ret_182; \
58053 })
58054 #else
58055 #define vqrshrun_high_n_s32(__p0_183, __p1_183, __p2_183) __extension__ ({ \
58056  int16x4_t __s0_183 = __p0_183; \
58057  int32x4_t __s1_183 = __p1_183; \
58058  int16x4_t __rev0_183; __rev0_183 = __builtin_shufflevector(__s0_183, __s0_183, 3, 2, 1, 0); \
58059  int32x4_t __rev1_183; __rev1_183 = __builtin_shufflevector(__s1_183, __s1_183, 3, 2, 1, 0); \
58060  int16x8_t __ret_183; \
58061  __ret_183 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_183), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_183, __p2_183)))); \
58062  __ret_183 = __builtin_shufflevector(__ret_183, __ret_183, 7, 6, 5, 4, 3, 2, 1, 0); \
58063  __ret_183; \
58064 })
58065 #endif
58066 
58067 #ifdef __LITTLE_ENDIAN__
58068 #define vqrshrun_high_n_s64(__p0_184, __p1_184, __p2_184) __extension__ ({ \
58069  int32x2_t __s0_184 = __p0_184; \
58070  int64x2_t __s1_184 = __p1_184; \
58071  int32x4_t __ret_184; \
58072  __ret_184 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_184), (int32x2_t)(vqrshrun_n_s64(__s1_184, __p2_184)))); \
58073  __ret_184; \
58074 })
58075 #else
58076 #define vqrshrun_high_n_s64(__p0_185, __p1_185, __p2_185) __extension__ ({ \
58077  int32x2_t __s0_185 = __p0_185; \
58078  int64x2_t __s1_185 = __p1_185; \
58079  int32x2_t __rev0_185; __rev0_185 = __builtin_shufflevector(__s0_185, __s0_185, 1, 0); \
58080  int64x2_t __rev1_185; __rev1_185 = __builtin_shufflevector(__s1_185, __s1_185, 1, 0); \
58081  int32x4_t __ret_185; \
58082  __ret_185 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_185), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_185, __p2_185)))); \
58083  __ret_185 = __builtin_shufflevector(__ret_185, __ret_185, 3, 2, 1, 0); \
58084  __ret_185; \
58085 })
58086 #endif
58087 
58088 #ifdef __LITTLE_ENDIAN__
58089 #define vqrshrun_high_n_s16(__p0_186, __p1_186, __p2_186) __extension__ ({ \
58090  int8x8_t __s0_186 = __p0_186; \
58091  int16x8_t __s1_186 = __p1_186; \
58092  int8x16_t __ret_186; \
58093  __ret_186 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_186), (int8x8_t)(vqrshrun_n_s16(__s1_186, __p2_186)))); \
58094  __ret_186; \
58095 })
58096 #else
58097 #define vqrshrun_high_n_s16(__p0_187, __p1_187, __p2_187) __extension__ ({ \
58098  int8x8_t __s0_187 = __p0_187; \
58099  int16x8_t __s1_187 = __p1_187; \
58100  int8x8_t __rev0_187; __rev0_187 = __builtin_shufflevector(__s0_187, __s0_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58101  int16x8_t __rev1_187; __rev1_187 = __builtin_shufflevector(__s1_187, __s1_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58102  int8x16_t __ret_187; \
58103  __ret_187 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_187), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_187, __p2_187)))); \
58104  __ret_187 = __builtin_shufflevector(__ret_187, __ret_187, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58105  __ret_187; \
58106 })
58107 #endif
58108 
58109 #ifdef __LITTLE_ENDIAN__
58110 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58111  int32_t __s0 = __p0; \
58112  int16_t __ret; \
58113  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58114  __ret; \
58115 })
58116 #else
58117 #define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58118  int32_t __s0 = __p0; \
58119  int16_t __ret; \
58120  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58121  __ret; \
58122 })
58123 #endif
58124 
58125 #ifdef __LITTLE_ENDIAN__
58126 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58127  int64_t __s0 = __p0; \
58128  int32_t __ret; \
58129  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58130  __ret; \
58131 })
58132 #else
58133 #define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58134  int64_t __s0 = __p0; \
58135  int32_t __ret; \
58136  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58137  __ret; \
58138 })
58139 #endif
58140 
58141 #ifdef __LITTLE_ENDIAN__
58142 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58143  int16_t __s0 = __p0; \
58144  int8_t __ret; \
58145  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58146  __ret; \
58147 })
58148 #else
58149 #define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58150  int16_t __s0 = __p0; \
58151  int8_t __ret; \
58152  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58153  __ret; \
58154 })
58155 #endif
58156 
58157 #ifdef __LITTLE_ENDIAN__
58158 __ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58159  uint8_t __ret;
58160  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58161  return __ret;
58162 }
58163 #else
58164 __ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58165  uint8_t __ret;
58166  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58167  return __ret;
58168 }
58169 #endif
58170 
58171 #ifdef __LITTLE_ENDIAN__
58172 __ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58173  uint32_t __ret;
58174  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58175  return __ret;
58176 }
58177 #else
58178 __ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58179  uint32_t __ret;
58180  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58181  return __ret;
58182 }
58183 #endif
58184 
58185 #ifdef __LITTLE_ENDIAN__
58186 __ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58187  uint64_t __ret;
58188  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58189  return __ret;
58190 }
58191 #else
58192 __ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58193  uint64_t __ret;
58194  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58195  return __ret;
58196 }
58197 #endif
58198 
58199 #ifdef __LITTLE_ENDIAN__
58200 __ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58201  uint16_t __ret;
58202  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58203  return __ret;
58204 }
58205 #else
58206 __ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58207  uint16_t __ret;
58208  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58209  return __ret;
58210 }
58211 #endif
58212 
58213 #ifdef __LITTLE_ENDIAN__
58214 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58215  int8_t __ret;
58216  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58217  return __ret;
58218 }
58219 #else
58220 __ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58221  int8_t __ret;
58222  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58223  return __ret;
58224 }
58225 #endif
58226 
58227 #ifdef __LITTLE_ENDIAN__
58228 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58229  int32_t __ret;
58230  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58231  return __ret;
58232 }
58233 #else
58234 __ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58235  int32_t __ret;
58236  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58237  return __ret;
58238 }
58239 #endif
58240 
58241 #ifdef __LITTLE_ENDIAN__
58242 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58243  int64_t __ret;
58244  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58245  return __ret;
58246 }
58247 #else
58248 __ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58249  int64_t __ret;
58250  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58251  return __ret;
58252 }
58253 #endif
58254 
58255 #ifdef __LITTLE_ENDIAN__
58256 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58257  int16_t __ret;
58258  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58259  return __ret;
58260 }
58261 #else
58262 __ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58263  int16_t __ret;
58264  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58265  return __ret;
58266 }
58267 #endif
58268 
58269 #ifdef __LITTLE_ENDIAN__
58270 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58271  uint8_t __s0 = __p0; \
58272  uint8_t __ret; \
58273  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58274  __ret; \
58275 })
58276 #else
58277 #define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58278  uint8_t __s0 = __p0; \
58279  uint8_t __ret; \
58280  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58281  __ret; \
58282 })
58283 #endif
58284 
58285 #ifdef __LITTLE_ENDIAN__
58286 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58287  uint32_t __s0 = __p0; \
58288  uint32_t __ret; \
58289  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58290  __ret; \
58291 })
58292 #else
58293 #define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58294  uint32_t __s0 = __p0; \
58295  uint32_t __ret; \
58296  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58297  __ret; \
58298 })
58299 #endif
58300 
58301 #ifdef __LITTLE_ENDIAN__
58302 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58303  uint64_t __s0 = __p0; \
58304  uint64_t __ret; \
58305  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58306  __ret; \
58307 })
58308 #else
58309 #define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58310  uint64_t __s0 = __p0; \
58311  uint64_t __ret; \
58312  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58313  __ret; \
58314 })
58315 #endif
58316 
58317 #ifdef __LITTLE_ENDIAN__
58318 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58319  uint16_t __s0 = __p0; \
58320  uint16_t __ret; \
58321  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58322  __ret; \
58323 })
58324 #else
58325 #define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58326  uint16_t __s0 = __p0; \
58327  uint16_t __ret; \
58328  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58329  __ret; \
58330 })
58331 #endif
58332 
58333 #ifdef __LITTLE_ENDIAN__
58334 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58335  int8_t __s0 = __p0; \
58336  int8_t __ret; \
58337  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58338  __ret; \
58339 })
58340 #else
58341 #define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58342  int8_t __s0 = __p0; \
58343  int8_t __ret; \
58344  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58345  __ret; \
58346 })
58347 #endif
58348 
58349 #ifdef __LITTLE_ENDIAN__
58350 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58351  int32_t __s0 = __p0; \
58352  int32_t __ret; \
58353  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58354  __ret; \
58355 })
58356 #else
58357 #define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58358  int32_t __s0 = __p0; \
58359  int32_t __ret; \
58360  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58361  __ret; \
58362 })
58363 #endif
58364 
58365 #ifdef __LITTLE_ENDIAN__
58366 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58367  int64_t __s0 = __p0; \
58368  int64_t __ret; \
58369  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58370  __ret; \
58371 })
58372 #else
58373 #define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58374  int64_t __s0 = __p0; \
58375  int64_t __ret; \
58376  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58377  __ret; \
58378 })
58379 #endif
58380 
58381 #ifdef __LITTLE_ENDIAN__
58382 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58383  int16_t __s0 = __p0; \
58384  int16_t __ret; \
58385  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58386  __ret; \
58387 })
58388 #else
58389 #define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58390  int16_t __s0 = __p0; \
58391  int16_t __ret; \
58392  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58393  __ret; \
58394 })
58395 #endif
58396 
58397 #ifdef __LITTLE_ENDIAN__
58398 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58399  int8_t __s0 = __p0; \
58400  int8_t __ret; \
58401  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58402  __ret; \
58403 })
58404 #else
58405 #define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58406  int8_t __s0 = __p0; \
58407  int8_t __ret; \
58408  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58409  __ret; \
58410 })
58411 #endif
58412 
58413 #ifdef __LITTLE_ENDIAN__
58414 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58415  int32_t __s0 = __p0; \
58416  int32_t __ret; \
58417  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58418  __ret; \
58419 })
58420 #else
58421 #define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58422  int32_t __s0 = __p0; \
58423  int32_t __ret; \
58424  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58425  __ret; \
58426 })
58427 #endif
58428 
58429 #ifdef __LITTLE_ENDIAN__
58430 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58431  int64_t __s0 = __p0; \
58432  int64_t __ret; \
58433  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58434  __ret; \
58435 })
58436 #else
58437 #define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58438  int64_t __s0 = __p0; \
58439  int64_t __ret; \
58440  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58441  __ret; \
58442 })
58443 #endif
58444 
58445 #ifdef __LITTLE_ENDIAN__
58446 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58447  int16_t __s0 = __p0; \
58448  int16_t __ret; \
58449  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58450  __ret; \
58451 })
58452 #else
58453 #define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58454  int16_t __s0 = __p0; \
58455  int16_t __ret; \
58456  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58457  __ret; \
58458 })
58459 #endif
58460 
58461 #ifdef __LITTLE_ENDIAN__
58462 #define vqshrn_high_n_u32(__p0_188, __p1_188, __p2_188) __extension__ ({ \
58463  uint16x4_t __s0_188 = __p0_188; \
58464  uint32x4_t __s1_188 = __p1_188; \
58465  uint16x8_t __ret_188; \
58466  __ret_188 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_188), (uint16x4_t)(vqshrn_n_u32(__s1_188, __p2_188)))); \
58467  __ret_188; \
58468 })
58469 #else
58470 #define vqshrn_high_n_u32(__p0_189, __p1_189, __p2_189) __extension__ ({ \
58471  uint16x4_t __s0_189 = __p0_189; \
58472  uint32x4_t __s1_189 = __p1_189; \
58473  uint16x4_t __rev0_189; __rev0_189 = __builtin_shufflevector(__s0_189, __s0_189, 3, 2, 1, 0); \
58474  uint32x4_t __rev1_189; __rev1_189 = __builtin_shufflevector(__s1_189, __s1_189, 3, 2, 1, 0); \
58475  uint16x8_t __ret_189; \
58476  __ret_189 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_189), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_189, __p2_189)))); \
58477  __ret_189 = __builtin_shufflevector(__ret_189, __ret_189, 7, 6, 5, 4, 3, 2, 1, 0); \
58478  __ret_189; \
58479 })
58480 #endif
58481 
58482 #ifdef __LITTLE_ENDIAN__
58483 #define vqshrn_high_n_u64(__p0_190, __p1_190, __p2_190) __extension__ ({ \
58484  uint32x2_t __s0_190 = __p0_190; \
58485  uint64x2_t __s1_190 = __p1_190; \
58486  uint32x4_t __ret_190; \
58487  __ret_190 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_190), (uint32x2_t)(vqshrn_n_u64(__s1_190, __p2_190)))); \
58488  __ret_190; \
58489 })
58490 #else
58491 #define vqshrn_high_n_u64(__p0_191, __p1_191, __p2_191) __extension__ ({ \
58492  uint32x2_t __s0_191 = __p0_191; \
58493  uint64x2_t __s1_191 = __p1_191; \
58494  uint32x2_t __rev0_191; __rev0_191 = __builtin_shufflevector(__s0_191, __s0_191, 1, 0); \
58495  uint64x2_t __rev1_191; __rev1_191 = __builtin_shufflevector(__s1_191, __s1_191, 1, 0); \
58496  uint32x4_t __ret_191; \
58497  __ret_191 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_191), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_191, __p2_191)))); \
58498  __ret_191 = __builtin_shufflevector(__ret_191, __ret_191, 3, 2, 1, 0); \
58499  __ret_191; \
58500 })
58501 #endif
58502 
58503 #ifdef __LITTLE_ENDIAN__
58504 #define vqshrn_high_n_u16(__p0_192, __p1_192, __p2_192) __extension__ ({ \
58505  uint8x8_t __s0_192 = __p0_192; \
58506  uint16x8_t __s1_192 = __p1_192; \
58507  uint8x16_t __ret_192; \
58508  __ret_192 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_192), (uint8x8_t)(vqshrn_n_u16(__s1_192, __p2_192)))); \
58509  __ret_192; \
58510 })
58511 #else
58512 #define vqshrn_high_n_u16(__p0_193, __p1_193, __p2_193) __extension__ ({ \
58513  uint8x8_t __s0_193 = __p0_193; \
58514  uint16x8_t __s1_193 = __p1_193; \
58515  uint8x8_t __rev0_193; __rev0_193 = __builtin_shufflevector(__s0_193, __s0_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58516  uint16x8_t __rev1_193; __rev1_193 = __builtin_shufflevector(__s1_193, __s1_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58517  uint8x16_t __ret_193; \
58518  __ret_193 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_193), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_193, __p2_193)))); \
58519  __ret_193 = __builtin_shufflevector(__ret_193, __ret_193, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58520  __ret_193; \
58521 })
58522 #endif
58523 
58524 #ifdef __LITTLE_ENDIAN__
58525 #define vqshrn_high_n_s32(__p0_194, __p1_194, __p2_194) __extension__ ({ \
58526  int16x4_t __s0_194 = __p0_194; \
58527  int32x4_t __s1_194 = __p1_194; \
58528  int16x8_t __ret_194; \
58529  __ret_194 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_194), (int16x4_t)(vqshrn_n_s32(__s1_194, __p2_194)))); \
58530  __ret_194; \
58531 })
58532 #else
58533 #define vqshrn_high_n_s32(__p0_195, __p1_195, __p2_195) __extension__ ({ \
58534  int16x4_t __s0_195 = __p0_195; \
58535  int32x4_t __s1_195 = __p1_195; \
58536  int16x4_t __rev0_195; __rev0_195 = __builtin_shufflevector(__s0_195, __s0_195, 3, 2, 1, 0); \
58537  int32x4_t __rev1_195; __rev1_195 = __builtin_shufflevector(__s1_195, __s1_195, 3, 2, 1, 0); \
58538  int16x8_t __ret_195; \
58539  __ret_195 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_195), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_195, __p2_195)))); \
58540  __ret_195 = __builtin_shufflevector(__ret_195, __ret_195, 7, 6, 5, 4, 3, 2, 1, 0); \
58541  __ret_195; \
58542 })
58543 #endif
58544 
58545 #ifdef __LITTLE_ENDIAN__
58546 #define vqshrn_high_n_s64(__p0_196, __p1_196, __p2_196) __extension__ ({ \
58547  int32x2_t __s0_196 = __p0_196; \
58548  int64x2_t __s1_196 = __p1_196; \
58549  int32x4_t __ret_196; \
58550  __ret_196 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_196), (int32x2_t)(vqshrn_n_s64(__s1_196, __p2_196)))); \
58551  __ret_196; \
58552 })
58553 #else
58554 #define vqshrn_high_n_s64(__p0_197, __p1_197, __p2_197) __extension__ ({ \
58555  int32x2_t __s0_197 = __p0_197; \
58556  int64x2_t __s1_197 = __p1_197; \
58557  int32x2_t __rev0_197; __rev0_197 = __builtin_shufflevector(__s0_197, __s0_197, 1, 0); \
58558  int64x2_t __rev1_197; __rev1_197 = __builtin_shufflevector(__s1_197, __s1_197, 1, 0); \
58559  int32x4_t __ret_197; \
58560  __ret_197 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_197), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_197, __p2_197)))); \
58561  __ret_197 = __builtin_shufflevector(__ret_197, __ret_197, 3, 2, 1, 0); \
58562  __ret_197; \
58563 })
58564 #endif
58565 
58566 #ifdef __LITTLE_ENDIAN__
58567 #define vqshrn_high_n_s16(__p0_198, __p1_198, __p2_198) __extension__ ({ \
58568  int8x8_t __s0_198 = __p0_198; \
58569  int16x8_t __s1_198 = __p1_198; \
58570  int8x16_t __ret_198; \
58571  __ret_198 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_198), (int8x8_t)(vqshrn_n_s16(__s1_198, __p2_198)))); \
58572  __ret_198; \
58573 })
58574 #else
58575 #define vqshrn_high_n_s16(__p0_199, __p1_199, __p2_199) __extension__ ({ \
58576  int8x8_t __s0_199 = __p0_199; \
58577  int16x8_t __s1_199 = __p1_199; \
58578  int8x8_t __rev0_199; __rev0_199 = __builtin_shufflevector(__s0_199, __s0_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58579  int16x8_t __rev1_199; __rev1_199 = __builtin_shufflevector(__s1_199, __s1_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58580  int8x16_t __ret_199; \
58581  __ret_199 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_199), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_199, __p2_199)))); \
58582  __ret_199 = __builtin_shufflevector(__ret_199, __ret_199, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58583  __ret_199; \
58584 })
58585 #endif
58586 
58587 #ifdef __LITTLE_ENDIAN__
58588 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58589  uint32_t __s0 = __p0; \
58590  uint16_t __ret; \
58591  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58592  __ret; \
58593 })
58594 #else
58595 #define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58596  uint32_t __s0 = __p0; \
58597  uint16_t __ret; \
58598  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58599  __ret; \
58600 })
58601 #endif
58602 
58603 #ifdef __LITTLE_ENDIAN__
58604 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58605  uint64_t __s0 = __p0; \
58606  uint32_t __ret; \
58607  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58608  __ret; \
58609 })
58610 #else
58611 #define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58612  uint64_t __s0 = __p0; \
58613  uint32_t __ret; \
58614  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58615  __ret; \
58616 })
58617 #endif
58618 
58619 #ifdef __LITTLE_ENDIAN__
58620 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58621  uint16_t __s0 = __p0; \
58622  uint8_t __ret; \
58623  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58624  __ret; \
58625 })
58626 #else
58627 #define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58628  uint16_t __s0 = __p0; \
58629  uint8_t __ret; \
58630  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58631  __ret; \
58632 })
58633 #endif
58634 
58635 #ifdef __LITTLE_ENDIAN__
58636 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58637  int32_t __s0 = __p0; \
58638  int16_t __ret; \
58639  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58640  __ret; \
58641 })
58642 #else
58643 #define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58644  int32_t __s0 = __p0; \
58645  int16_t __ret; \
58646  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58647  __ret; \
58648 })
58649 #endif
58650 
58651 #ifdef __LITTLE_ENDIAN__
58652 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58653  int64_t __s0 = __p0; \
58654  int32_t __ret; \
58655  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58656  __ret; \
58657 })
58658 #else
58659 #define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58660  int64_t __s0 = __p0; \
58661  int32_t __ret; \
58662  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58663  __ret; \
58664 })
58665 #endif
58666 
58667 #ifdef __LITTLE_ENDIAN__
58668 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58669  int16_t __s0 = __p0; \
58670  int8_t __ret; \
58671  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58672  __ret; \
58673 })
58674 #else
58675 #define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58676  int16_t __s0 = __p0; \
58677  int8_t __ret; \
58678  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58679  __ret; \
58680 })
58681 #endif
58682 
58683 #ifdef __LITTLE_ENDIAN__
58684 #define vqshrun_high_n_s32(__p0_200, __p1_200, __p2_200) __extension__ ({ \
58685  int16x4_t __s0_200 = __p0_200; \
58686  int32x4_t __s1_200 = __p1_200; \
58687  int16x8_t __ret_200; \
58688  __ret_200 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_200), (int16x4_t)(vqshrun_n_s32(__s1_200, __p2_200)))); \
58689  __ret_200; \
58690 })
58691 #else
58692 #define vqshrun_high_n_s32(__p0_201, __p1_201, __p2_201) __extension__ ({ \
58693  int16x4_t __s0_201 = __p0_201; \
58694  int32x4_t __s1_201 = __p1_201; \
58695  int16x4_t __rev0_201; __rev0_201 = __builtin_shufflevector(__s0_201, __s0_201, 3, 2, 1, 0); \
58696  int32x4_t __rev1_201; __rev1_201 = __builtin_shufflevector(__s1_201, __s1_201, 3, 2, 1, 0); \
58697  int16x8_t __ret_201; \
58698  __ret_201 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_201), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_201, __p2_201)))); \
58699  __ret_201 = __builtin_shufflevector(__ret_201, __ret_201, 7, 6, 5, 4, 3, 2, 1, 0); \
58700  __ret_201; \
58701 })
58702 #endif
58703 
58704 #ifdef __LITTLE_ENDIAN__
58705 #define vqshrun_high_n_s64(__p0_202, __p1_202, __p2_202) __extension__ ({ \
58706  int32x2_t __s0_202 = __p0_202; \
58707  int64x2_t __s1_202 = __p1_202; \
58708  int32x4_t __ret_202; \
58709  __ret_202 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_202), (int32x2_t)(vqshrun_n_s64(__s1_202, __p2_202)))); \
58710  __ret_202; \
58711 })
58712 #else
58713 #define vqshrun_high_n_s64(__p0_203, __p1_203, __p2_203) __extension__ ({ \
58714  int32x2_t __s0_203 = __p0_203; \
58715  int64x2_t __s1_203 = __p1_203; \
58716  int32x2_t __rev0_203; __rev0_203 = __builtin_shufflevector(__s0_203, __s0_203, 1, 0); \
58717  int64x2_t __rev1_203; __rev1_203 = __builtin_shufflevector(__s1_203, __s1_203, 1, 0); \
58718  int32x4_t __ret_203; \
58719  __ret_203 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_203), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_203, __p2_203)))); \
58720  __ret_203 = __builtin_shufflevector(__ret_203, __ret_203, 3, 2, 1, 0); \
58721  __ret_203; \
58722 })
58723 #endif
58724 
58725 #ifdef __LITTLE_ENDIAN__
58726 #define vqshrun_high_n_s16(__p0_204, __p1_204, __p2_204) __extension__ ({ \
58727  int8x8_t __s0_204 = __p0_204; \
58728  int16x8_t __s1_204 = __p1_204; \
58729  int8x16_t __ret_204; \
58730  __ret_204 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_204), (int8x8_t)(vqshrun_n_s16(__s1_204, __p2_204)))); \
58731  __ret_204; \
58732 })
58733 #else
58734 #define vqshrun_high_n_s16(__p0_205, __p1_205, __p2_205) __extension__ ({ \
58735  int8x8_t __s0_205 = __p0_205; \
58736  int16x8_t __s1_205 = __p1_205; \
58737  int8x8_t __rev0_205; __rev0_205 = __builtin_shufflevector(__s0_205, __s0_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58738  int16x8_t __rev1_205; __rev1_205 = __builtin_shufflevector(__s1_205, __s1_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58739  int8x16_t __ret_205; \
58740  __ret_205 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_205), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_205, __p2_205)))); \
58741  __ret_205 = __builtin_shufflevector(__ret_205, __ret_205, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58742  __ret_205; \
58743 })
58744 #endif
58745 
58746 #ifdef __LITTLE_ENDIAN__
58747 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58748  int32_t __s0 = __p0; \
58749  int16_t __ret; \
58750  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58751  __ret; \
58752 })
58753 #else
58754 #define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58755  int32_t __s0 = __p0; \
58756  int16_t __ret; \
58757  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58758  __ret; \
58759 })
58760 #endif
58761 
58762 #ifdef __LITTLE_ENDIAN__
58763 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58764  int64_t __s0 = __p0; \
58765  int32_t __ret; \
58766  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58767  __ret; \
58768 })
58769 #else
58770 #define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58771  int64_t __s0 = __p0; \
58772  int32_t __ret; \
58773  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58774  __ret; \
58775 })
58776 #endif
58777 
58778 #ifdef __LITTLE_ENDIAN__
58779 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58780  int16_t __s0 = __p0; \
58781  int8_t __ret; \
58782  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58783  __ret; \
58784 })
58785 #else
58786 #define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58787  int16_t __s0 = __p0; \
58788  int8_t __ret; \
58789  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58790  __ret; \
58791 })
58792 #endif
58793 
58794 #ifdef __LITTLE_ENDIAN__
58795 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58796  uint8_t __ret;
58797  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58798  return __ret;
58799 }
58800 #else
58801 __ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58802  uint8_t __ret;
58803  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58804  return __ret;
58805 }
58806 #endif
58807 
58808 #ifdef __LITTLE_ENDIAN__
58809 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58810  uint32_t __ret;
58811  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58812  return __ret;
58813 }
58814 #else
58815 __ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58816  uint32_t __ret;
58817  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58818  return __ret;
58819 }
58820 #endif
58821 
58822 #ifdef __LITTLE_ENDIAN__
58823 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58824  uint64_t __ret;
58825  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58826  return __ret;
58827 }
58828 #else
58829 __ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58830  uint64_t __ret;
58831  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58832  return __ret;
58833 }
58834 #endif
58835 
58836 #ifdef __LITTLE_ENDIAN__
58837 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58838  uint16_t __ret;
58839  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58840  return __ret;
58841 }
58842 #else
58843 __ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58844  uint16_t __ret;
58845  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58846  return __ret;
58847 }
58848 #endif
58849 
58850 #ifdef __LITTLE_ENDIAN__
58851 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58852  int8_t __ret;
58853  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58854  return __ret;
58855 }
58856 #else
58857 __ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58858  int8_t __ret;
58859  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58860  return __ret;
58861 }
58862 #endif
58863 
58864 #ifdef __LITTLE_ENDIAN__
58865 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58866  int32_t __ret;
58867  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58868  return __ret;
58869 }
58870 #else
58871 __ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58872  int32_t __ret;
58873  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58874  return __ret;
58875 }
58876 __ai int32_t __noswap_vqsubs_s32(int32_t __p0, int32_t __p1) {
58877  int32_t __ret;
58878  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58879  return __ret;
58880 }
58881 #endif
58882 
58883 #ifdef __LITTLE_ENDIAN__
58884 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58885  int64_t __ret;
58886  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58887  return __ret;
58888 }
58889 #else
58890 __ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58891  int64_t __ret;
58892  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58893  return __ret;
58894 }
58895 #endif
58896 
58897 #ifdef __LITTLE_ENDIAN__
58898 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58899  int16_t __ret;
58900  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58901  return __ret;
58902 }
58903 #else
58904 __ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58905  int16_t __ret;
58906  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58907  return __ret;
58908 }
58909 __ai int16_t __noswap_vqsubh_s16(int16_t __p0, int16_t __p1) {
58910  int16_t __ret;
58911  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58912  return __ret;
58913 }
58914 #endif
58915 
58916 #ifdef __LITTLE_ENDIAN__
58917 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58918  poly8x8_t __ret;
58919  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
58920  return __ret;
58921 }
58922 #else
58923 __ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58924  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58925  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58926  poly8x8_t __ret;
58927  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
58928  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58929  return __ret;
58930 }
58931 #endif
58932 
58933 #ifdef __LITTLE_ENDIAN__
58934 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58935  poly8x16_t __ret;
58936  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
58937  return __ret;
58938 }
58939 #else
58940 __ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58941  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58942  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58943  poly8x16_t __ret;
58944  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
58945  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58946  return __ret;
58947 }
58948 #endif
58949 
58950 #ifdef __LITTLE_ENDIAN__
58951 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58952  uint8x16_t __ret;
58953  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
58954  return __ret;
58955 }
58956 #else
58957 __ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58958  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58959  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58960  uint8x16_t __ret;
58961  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
58962  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58963  return __ret;
58964 }
58965 #endif
58966 
58967 #ifdef __LITTLE_ENDIAN__
58968 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58969  int8x16_t __ret;
58970  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
58971  return __ret;
58972 }
58973 #else
58974 __ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58975  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58976  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58977  int8x16_t __ret;
58978  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
58979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58980  return __ret;
58981 }
58982 #endif
58983 
58984 #ifdef __LITTLE_ENDIAN__
58985 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58986  uint8x8_t __ret;
58987  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
58988  return __ret;
58989 }
58990 #else
58991 __ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58992  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58993  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58994  uint8x8_t __ret;
58995  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
58996  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58997  return __ret;
58998 }
58999 #endif
59000 
59001 #ifdef __LITTLE_ENDIAN__
59002 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
59003  int8x8_t __ret;
59004  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
59005  return __ret;
59006 }
59007 #else
59008 __ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
59009  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59010  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59011  int8x8_t __ret;
59012  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
59013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59014  return __ret;
59015 }
59016 #endif
59017 
59018 #ifdef __LITTLE_ENDIAN__
59019 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59020  poly8x8_t __ret;
59021  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
59022  return __ret;
59023 }
59024 #else
59025 __ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59026  poly8x16x2_t __rev0;
59027  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59028  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59029  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59030  poly8x8_t __ret;
59031  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
59032  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59033  return __ret;
59034 }
59035 #endif
59036 
59037 #ifdef __LITTLE_ENDIAN__
59038 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59039  poly8x16_t __ret;
59040  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
59041  return __ret;
59042 }
59043 #else
59044 __ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59045  poly8x16x2_t __rev0;
59046  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59047  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59048  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59049  poly8x16_t __ret;
59050  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
59051  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59052  return __ret;
59053 }
59054 #endif
59055 
59056 #ifdef __LITTLE_ENDIAN__
59057 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59058  uint8x16_t __ret;
59059  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
59060  return __ret;
59061 }
59062 #else
59063 __ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59064  uint8x16x2_t __rev0;
59065  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59066  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59067  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59068  uint8x16_t __ret;
59069  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
59070  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59071  return __ret;
59072 }
59073 #endif
59074 
59075 #ifdef __LITTLE_ENDIAN__
59076 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59077  int8x16_t __ret;
59078  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
59079  return __ret;
59080 }
59081 #else
59082 __ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59083  int8x16x2_t __rev0;
59084  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59085  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59086  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59087  int8x16_t __ret;
59088  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
59089  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59090  return __ret;
59091 }
59092 #endif
59093 
59094 #ifdef __LITTLE_ENDIAN__
59095 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59096  uint8x8_t __ret;
59097  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
59098  return __ret;
59099 }
59100 #else
59101 __ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59102  uint8x16x2_t __rev0;
59103  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59104  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59105  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59106  uint8x8_t __ret;
59107  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
59108  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59109  return __ret;
59110 }
59111 #endif
59112 
59113 #ifdef __LITTLE_ENDIAN__
59114 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59115  int8x8_t __ret;
59116  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
59117  return __ret;
59118 }
59119 #else
59120 __ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59121  int8x16x2_t __rev0;
59122  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59123  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59124  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59125  int8x8_t __ret;
59126  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
59127  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59128  return __ret;
59129 }
59130 #endif
59131 
59132 #ifdef __LITTLE_ENDIAN__
59133 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59134  poly8x8_t __ret;
59135  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
59136  return __ret;
59137 }
59138 #else
59139 __ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59140  poly8x16x3_t __rev0;
59141  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59142  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59143  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59144  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59145  poly8x8_t __ret;
59146  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
59147  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59148  return __ret;
59149 }
59150 #endif
59151 
59152 #ifdef __LITTLE_ENDIAN__
59153 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59154  poly8x16_t __ret;
59155  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
59156  return __ret;
59157 }
59158 #else
59159 __ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59160  poly8x16x3_t __rev0;
59161  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59162  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59163  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59164  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59165  poly8x16_t __ret;
59166  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
59167  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59168  return __ret;
59169 }
59170 #endif
59171 
59172 #ifdef __LITTLE_ENDIAN__
59173 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59174  uint8x16_t __ret;
59175  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
59176  return __ret;
59177 }
59178 #else
59179 __ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59180  uint8x16x3_t __rev0;
59181  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59182  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59183  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59184  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59185  uint8x16_t __ret;
59186  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
59187  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59188  return __ret;
59189 }
59190 #endif
59191 
59192 #ifdef __LITTLE_ENDIAN__
59193 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59194  int8x16_t __ret;
59195  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
59196  return __ret;
59197 }
59198 #else
59199 __ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59200  int8x16x3_t __rev0;
59201  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59202  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59203  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59204  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59205  int8x16_t __ret;
59206  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
59207  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59208  return __ret;
59209 }
59210 #endif
59211 
59212 #ifdef __LITTLE_ENDIAN__
59213 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59214  uint8x8_t __ret;
59215  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
59216  return __ret;
59217 }
59218 #else
59219 __ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59220  uint8x16x3_t __rev0;
59221  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59222  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59223  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59224  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59225  uint8x8_t __ret;
59226  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
59227  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59228  return __ret;
59229 }
59230 #endif
59231 
59232 #ifdef __LITTLE_ENDIAN__
59233 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59234  int8x8_t __ret;
59235  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
59236  return __ret;
59237 }
59238 #else
59239 __ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59240  int8x16x3_t __rev0;
59241  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59242  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59243  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59244  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59245  int8x8_t __ret;
59246  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
59247  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59248  return __ret;
59249 }
59250 #endif
59251 
59252 #ifdef __LITTLE_ENDIAN__
59253 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59254  poly8x8_t __ret;
59255  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
59256  return __ret;
59257 }
59258 #else
59259 __ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59260  poly8x16x4_t __rev0;
59261  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59262  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59263  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59264  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59265  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59266  poly8x8_t __ret;
59267  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
59268  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59269  return __ret;
59270 }
59271 #endif
59272 
59273 #ifdef __LITTLE_ENDIAN__
59274 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59275  poly8x16_t __ret;
59276  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
59277  return __ret;
59278 }
59279 #else
59280 __ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59281  poly8x16x4_t __rev0;
59282  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59283  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59284  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59285  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59286  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59287  poly8x16_t __ret;
59288  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
59289  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59290  return __ret;
59291 }
59292 #endif
59293 
59294 #ifdef __LITTLE_ENDIAN__
59295 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59296  uint8x16_t __ret;
59297  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
59298  return __ret;
59299 }
59300 #else
59301 __ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59302  uint8x16x4_t __rev0;
59303  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59304  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59305  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59306  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59307  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59308  uint8x16_t __ret;
59309  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
59310  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59311  return __ret;
59312 }
59313 #endif
59314 
59315 #ifdef __LITTLE_ENDIAN__
59316 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59317  int8x16_t __ret;
59318  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
59319  return __ret;
59320 }
59321 #else
59322 __ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59323  int8x16x4_t __rev0;
59324  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59325  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59326  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59327  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59328  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59329  int8x16_t __ret;
59330  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
59331  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59332  return __ret;
59333 }
59334 #endif
59335 
59336 #ifdef __LITTLE_ENDIAN__
59337 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59338  uint8x8_t __ret;
59339  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
59340  return __ret;
59341 }
59342 #else
59343 __ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59344  uint8x16x4_t __rev0;
59345  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59346  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59347  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59348  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59349  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59350  uint8x8_t __ret;
59351  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
59352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59353  return __ret;
59354 }
59355 #endif
59356 
59357 #ifdef __LITTLE_ENDIAN__
59358 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59359  int8x8_t __ret;
59360  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
59361  return __ret;
59362 }
59363 #else
59364 __ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59365  int8x16x4_t __rev0;
59366  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59367  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59368  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59369  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59370  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59371  int8x8_t __ret;
59372  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
59373  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59374  return __ret;
59375 }
59376 #endif
59377 
59378 #ifdef __LITTLE_ENDIAN__
59379 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59380  poly8x8_t __ret;
59381  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
59382  return __ret;
59383 }
59384 #else
59385 __ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59386  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59387  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59388  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59389  poly8x8_t __ret;
59390  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
59391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59392  return __ret;
59393 }
59394 #endif
59395 
59396 #ifdef __LITTLE_ENDIAN__
59397 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59398  poly8x16_t __ret;
59399  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
59400  return __ret;
59401 }
59402 #else
59403 __ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59404  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59405  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59406  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59407  poly8x16_t __ret;
59408  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
59409  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59410  return __ret;
59411 }
59412 #endif
59413 
59414 #ifdef __LITTLE_ENDIAN__
59415 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59416  uint8x16_t __ret;
59417  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
59418  return __ret;
59419 }
59420 #else
59421 __ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59422  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59423  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59424  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59425  uint8x16_t __ret;
59426  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
59427  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59428  return __ret;
59429 }
59430 #endif
59431 
59432 #ifdef __LITTLE_ENDIAN__
59433 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59434  int8x16_t __ret;
59435  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
59436  return __ret;
59437 }
59438 #else
59439 __ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59440  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59441  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59442  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59443  int8x16_t __ret;
59444  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
59445  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59446  return __ret;
59447 }
59448 #endif
59449 
59450 #ifdef __LITTLE_ENDIAN__
59451 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59452  uint8x8_t __ret;
59453  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
59454  return __ret;
59455 }
59456 #else
59457 __ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59458  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59459  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59460  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59461  uint8x8_t __ret;
59462  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
59463  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59464  return __ret;
59465 }
59466 #endif
59467 
59468 #ifdef __LITTLE_ENDIAN__
59469 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59470  int8x8_t __ret;
59471  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
59472  return __ret;
59473 }
59474 #else
59475 __ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59476  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59477  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59478  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59479  int8x8_t __ret;
59480  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
59481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59482  return __ret;
59483 }
59484 #endif
59485 
59486 #ifdef __LITTLE_ENDIAN__
59487 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59488  poly8x8_t __ret;
59489  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
59490  return __ret;
59491 }
59492 #else
59493 __ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59494  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59495  poly8x16x2_t __rev1;
59496  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59497  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59498  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59499  poly8x8_t __ret;
59500  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
59501  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59502  return __ret;
59503 }
59504 #endif
59505 
59506 #ifdef __LITTLE_ENDIAN__
59507 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59508  poly8x16_t __ret;
59509  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
59510  return __ret;
59511 }
59512 #else
59513 __ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59514  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59515  poly8x16x2_t __rev1;
59516  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59517  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59518  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59519  poly8x16_t __ret;
59520  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
59521  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59522  return __ret;
59523 }
59524 #endif
59525 
59526 #ifdef __LITTLE_ENDIAN__
59527 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59528  uint8x16_t __ret;
59529  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
59530  return __ret;
59531 }
59532 #else
59533 __ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59534  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59535  uint8x16x2_t __rev1;
59536  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59537  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59538  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59539  uint8x16_t __ret;
59540  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
59541  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59542  return __ret;
59543 }
59544 #endif
59545 
59546 #ifdef __LITTLE_ENDIAN__
59547 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59548  int8x16_t __ret;
59549  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
59550  return __ret;
59551 }
59552 #else
59553 __ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59554  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59555  int8x16x2_t __rev1;
59556  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59557  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59558  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59559  int8x16_t __ret;
59560  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
59561  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59562  return __ret;
59563 }
59564 #endif
59565 
59566 #ifdef __LITTLE_ENDIAN__
59567 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59568  uint8x8_t __ret;
59569  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
59570  return __ret;
59571 }
59572 #else
59573 __ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59574  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59575  uint8x16x2_t __rev1;
59576  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59577  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59578  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59579  uint8x8_t __ret;
59580  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
59581  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59582  return __ret;
59583 }
59584 #endif
59585 
59586 #ifdef __LITTLE_ENDIAN__
59587 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59588  int8x8_t __ret;
59589  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
59590  return __ret;
59591 }
59592 #else
59593 __ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59594  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59595  int8x16x2_t __rev1;
59596  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59597  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59598  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59599  int8x8_t __ret;
59600  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
59601  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59602  return __ret;
59603 }
59604 #endif
59605 
59606 #ifdef __LITTLE_ENDIAN__
59607 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59608  poly8x8_t __ret;
59609  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
59610  return __ret;
59611 }
59612 #else
59613 __ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59614  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59615  poly8x16x3_t __rev1;
59616  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59617  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59618  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59619  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59620  poly8x8_t __ret;
59621  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
59622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59623  return __ret;
59624 }
59625 #endif
59626 
59627 #ifdef __LITTLE_ENDIAN__
59628 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59629  poly8x16_t __ret;
59630  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
59631  return __ret;
59632 }
59633 #else
59634 __ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59635  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59636  poly8x16x3_t __rev1;
59637  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59638  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59639  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59640  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59641  poly8x16_t __ret;
59642  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
59643  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59644  return __ret;
59645 }
59646 #endif
59647 
59648 #ifdef __LITTLE_ENDIAN__
59649 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59650  uint8x16_t __ret;
59651  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
59652  return __ret;
59653 }
59654 #else
59655 __ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59656  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59657  uint8x16x3_t __rev1;
59658  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59659  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59660  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59661  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59662  uint8x16_t __ret;
59663  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
59664  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59665  return __ret;
59666 }
59667 #endif
59668 
59669 #ifdef __LITTLE_ENDIAN__
59670 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59671  int8x16_t __ret;
59672  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
59673  return __ret;
59674 }
59675 #else
59676 __ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59677  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59678  int8x16x3_t __rev1;
59679  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59680  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59681  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59682  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59683  int8x16_t __ret;
59684  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
59685  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59686  return __ret;
59687 }
59688 #endif
59689 
59690 #ifdef __LITTLE_ENDIAN__
59691 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59692  uint8x8_t __ret;
59693  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
59694  return __ret;
59695 }
59696 #else
59697 __ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59698  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59699  uint8x16x3_t __rev1;
59700  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59701  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59702  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59703  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59704  uint8x8_t __ret;
59705  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
59706  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59707  return __ret;
59708 }
59709 #endif
59710 
59711 #ifdef __LITTLE_ENDIAN__
59712 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59713  int8x8_t __ret;
59714  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
59715  return __ret;
59716 }
59717 #else
59718 __ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59719  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59720  int8x16x3_t __rev1;
59721  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59722  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59723  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59724  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59725  int8x8_t __ret;
59726  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
59727  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59728  return __ret;
59729 }
59730 #endif
59731 
59732 #ifdef __LITTLE_ENDIAN__
59733 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59734  poly8x8_t __ret;
59735  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
59736  return __ret;
59737 }
59738 #else
59739 __ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59740  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59741  poly8x16x4_t __rev1;
59742  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59743  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59744  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59745  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59746  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59747  poly8x8_t __ret;
59748  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
59749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59750  return __ret;
59751 }
59752 #endif
59753 
59754 #ifdef __LITTLE_ENDIAN__
59755 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59756  poly8x16_t __ret;
59757  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
59758  return __ret;
59759 }
59760 #else
59761 __ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59762  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59763  poly8x16x4_t __rev1;
59764  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59765  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59766  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59767  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59768  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59769  poly8x16_t __ret;
59770  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
59771  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59772  return __ret;
59773 }
59774 #endif
59775 
59776 #ifdef __LITTLE_ENDIAN__
59777 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59778  uint8x16_t __ret;
59779  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
59780  return __ret;
59781 }
59782 #else
59783 __ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59784  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59785  uint8x16x4_t __rev1;
59786  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59787  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59788  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59789  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59790  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59791  uint8x16_t __ret;
59792  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
59793  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59794  return __ret;
59795 }
59796 #endif
59797 
59798 #ifdef __LITTLE_ENDIAN__
59799 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59800  int8x16_t __ret;
59801  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
59802  return __ret;
59803 }
59804 #else
59805 __ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59806  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59807  int8x16x4_t __rev1;
59808  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59809  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59810  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59811  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59812  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59813  int8x16_t __ret;
59814  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
59815  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59816  return __ret;
59817 }
59818 #endif
59819 
59820 #ifdef __LITTLE_ENDIAN__
59821 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59822  uint8x8_t __ret;
59823  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
59824  return __ret;
59825 }
59826 #else
59827 __ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59828  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59829  uint8x16x4_t __rev1;
59830  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59831  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59832  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59833  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59834  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59835  uint8x8_t __ret;
59836  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
59837  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59838  return __ret;
59839 }
59840 #endif
59841 
59842 #ifdef __LITTLE_ENDIAN__
59843 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59844  int8x8_t __ret;
59845  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
59846  return __ret;
59847 }
59848 #else
59849 __ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59850  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59851  int8x16x4_t __rev1;
59852  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59853  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59854  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59855  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59856  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59857  int8x8_t __ret;
59858  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
59859  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59860  return __ret;
59861 }
59862 #endif
59863 
59864 #ifdef __LITTLE_ENDIAN__
59865 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59866  uint16x8_t __ret;
59867  __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
59868  return __ret;
59869 }
59870 #else
59871 __ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59872  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59873  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59874  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59875  uint16x8_t __ret;
59876  __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
59877  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59878  return __ret;
59879 }
59880 #endif
59881 
59882 #ifdef __LITTLE_ENDIAN__
59883 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59884  uint32x4_t __ret;
59885  __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
59886  return __ret;
59887 }
59888 #else
59889 __ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59890  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59891  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59892  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59893  uint32x4_t __ret;
59894  __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
59895  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59896  return __ret;
59897 }
59898 #endif
59899 
59900 #ifdef __LITTLE_ENDIAN__
59901 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59902  uint8x16_t __ret;
59903  __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
59904  return __ret;
59905 }
59906 #else
59907 __ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59908  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59909  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59910  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59911  uint8x16_t __ret;
59912  __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
59913  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59914  return __ret;
59915 }
59916 #endif
59917 
59918 #ifdef __LITTLE_ENDIAN__
59919 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59920  int16x8_t __ret;
59921  __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
59922  return __ret;
59923 }
59924 #else
59925 __ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59926  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59927  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59928  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59929  int16x8_t __ret;
59930  __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
59931  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59932  return __ret;
59933 }
59934 #endif
59935 
59936 #ifdef __LITTLE_ENDIAN__
59937 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59938  int32x4_t __ret;
59939  __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
59940  return __ret;
59941 }
59942 #else
59943 __ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59944  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59945  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59946  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59947  int32x4_t __ret;
59948  __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
59949  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59950  return __ret;
59951 }
59952 #endif
59953 
59954 #ifdef __LITTLE_ENDIAN__
59955 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59956  int8x16_t __ret;
59957  __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
59958  return __ret;
59959 }
59960 #else
59961 __ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59962  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59963  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59964  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59965  int8x16_t __ret;
59966  __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
59967  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59968  return __ret;
59969 }
59970 #endif
59971 
59972 #ifdef __LITTLE_ENDIAN__
59973 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59974  poly8x8_t __ret;
59975  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
59976  return __ret;
59977 }
59978 #else
59979 __ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59980  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59981  poly8x8_t __ret;
59982  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
59983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59984  return __ret;
59985 }
59986 #endif
59987 
59988 #ifdef __LITTLE_ENDIAN__
59989 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59990  poly8x16_t __ret;
59991  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
59992  return __ret;
59993 }
59994 #else
59995 __ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59996  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59997  poly8x16_t __ret;
59998  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
59999  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60000  return __ret;
60001 }
60002 #endif
60003 
60004 #ifdef __LITTLE_ENDIAN__
60005 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60006  uint8x16_t __ret;
60007  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
60008  return __ret;
60009 }
60010 #else
60011 __ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60012  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60013  uint8x16_t __ret;
60014  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
60015  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60016  return __ret;
60017 }
60018 #endif
60019 
60020 #ifdef __LITTLE_ENDIAN__
60021 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60022  int8x16_t __ret;
60023  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
60024  return __ret;
60025 }
60026 #else
60027 __ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60028  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60029  int8x16_t __ret;
60030  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
60031  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60032  return __ret;
60033 }
60034 #endif
60035 
60036 #ifdef __LITTLE_ENDIAN__
60037 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60038  uint8x8_t __ret;
60039  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
60040  return __ret;
60041 }
60042 #else
60043 __ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60044  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60045  uint8x8_t __ret;
60046  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
60047  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60048  return __ret;
60049 }
60050 #endif
60051 
60052 #ifdef __LITTLE_ENDIAN__
60053 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
60054  int8x8_t __ret;
60055  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
60056  return __ret;
60057 }
60058 #else
60059 __ai int8x8_t vrbit_s8(int8x8_t __p0) {
60060  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60061  int8x8_t __ret;
60062  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
60063  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60064  return __ret;
60065 }
60066 #endif
60067 
60068 #ifdef __LITTLE_ENDIAN__
60069 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60070  float64x2_t __ret;
60071  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
60072  return __ret;
60073 }
60074 #else
60075 __ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60076  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60077  float64x2_t __ret;
60078  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
60079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60080  return __ret;
60081 }
60082 #endif
60083 
60084 #ifdef __LITTLE_ENDIAN__
60085 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60086  float64x1_t __ret;
60087  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60088  return __ret;
60089 }
60090 #else
60091 __ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60092  float64x1_t __ret;
60093  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60094  return __ret;
60095 }
60096 #endif
60097 
60098 #ifdef __LITTLE_ENDIAN__
60099 __ai float64_t vrecped_f64(float64_t __p0) {
60100  float64_t __ret;
60101  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60102  return __ret;
60103 }
60104 #else
60105 __ai float64_t vrecped_f64(float64_t __p0) {
60106  float64_t __ret;
60107  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60108  return __ret;
60109 }
60110 #endif
60111 
60112 #ifdef __LITTLE_ENDIAN__
60113 __ai float32_t vrecpes_f32(float32_t __p0) {
60114  float32_t __ret;
60115  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60116  return __ret;
60117 }
60118 #else
60119 __ai float32_t vrecpes_f32(float32_t __p0) {
60120  float32_t __ret;
60121  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60122  return __ret;
60123 }
60124 #endif
60125 
60126 #ifdef __LITTLE_ENDIAN__
60127 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60128  float64x2_t __ret;
60129  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60130  return __ret;
60131 }
60132 #else
60133 __ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60134  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60135  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60136  float64x2_t __ret;
60137  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60138  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60139  return __ret;
60140 }
60141 #endif
60142 
60143 #ifdef __LITTLE_ENDIAN__
60144 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60145  float64x1_t __ret;
60146  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60147  return __ret;
60148 }
60149 #else
60150 __ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60151  float64x1_t __ret;
60152  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60153  return __ret;
60154 }
60155 #endif
60156 
60157 #ifdef __LITTLE_ENDIAN__
60158 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60159  float64_t __ret;
60160  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60161  return __ret;
60162 }
60163 #else
60164 __ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60165  float64_t __ret;
60166  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60167  return __ret;
60168 }
60169 #endif
60170 
60171 #ifdef __LITTLE_ENDIAN__
60172 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60173  float32_t __ret;
60174  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60175  return __ret;
60176 }
60177 #else
60178 __ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60179  float32_t __ret;
60180  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60181  return __ret;
60182 }
60183 #endif
60184 
60185 #ifdef __LITTLE_ENDIAN__
60186 __ai float64_t vrecpxd_f64(float64_t __p0) {
60187  float64_t __ret;
60188  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60189  return __ret;
60190 }
60191 #else
60192 __ai float64_t vrecpxd_f64(float64_t __p0) {
60193  float64_t __ret;
60194  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60195  return __ret;
60196 }
60197 #endif
60198 
60199 #ifdef __LITTLE_ENDIAN__
60200 __ai float32_t vrecpxs_f32(float32_t __p0) {
60201  float32_t __ret;
60202  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60203  return __ret;
60204 }
60205 #else
60206 __ai float32_t vrecpxs_f32(float32_t __p0) {
60207  float32_t __ret;
60208  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60209  return __ret;
60210 }
60211 #endif
60212 
60213 #ifdef __LITTLE_ENDIAN__
60214 __ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60215  uint64_t __ret;
60216  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60217  return __ret;
60218 }
60219 #else
60220 __ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60221  uint64_t __ret;
60222  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60223  return __ret;
60224 }
60225 #endif
60226 
60227 #ifdef __LITTLE_ENDIAN__
60228 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60229  int64_t __ret;
60230  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60231  return __ret;
60232 }
60233 #else
60234 __ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60235  int64_t __ret;
60236  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60237  return __ret;
60238 }
60239 #endif
60240 
60241 #ifdef __LITTLE_ENDIAN__
60242 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60243  uint64_t __s0 = __p0; \
60244  uint64_t __ret; \
60245  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60246  __ret; \
60247 })
60248 #else
60249 #define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60250  uint64_t __s0 = __p0; \
60251  uint64_t __ret; \
60252  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60253  __ret; \
60254 })
60255 #endif
60256 
60257 #ifdef __LITTLE_ENDIAN__
60258 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60259  int64_t __s0 = __p0; \
60260  int64_t __ret; \
60261  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60262  __ret; \
60263 })
60264 #else
60265 #define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60266  int64_t __s0 = __p0; \
60267  int64_t __ret; \
60268  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60269  __ret; \
60270 })
60271 #endif
60272 
60273 #ifdef __LITTLE_ENDIAN__
60274 #define vrshrn_high_n_u32(__p0_206, __p1_206, __p2_206) __extension__ ({ \
60275  uint16x4_t __s0_206 = __p0_206; \
60276  uint32x4_t __s1_206 = __p1_206; \
60277  uint16x8_t __ret_206; \
60278  __ret_206 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_206), (uint16x4_t)(vrshrn_n_u32(__s1_206, __p2_206)))); \
60279  __ret_206; \
60280 })
60281 #else
60282 #define vrshrn_high_n_u32(__p0_207, __p1_207, __p2_207) __extension__ ({ \
60283  uint16x4_t __s0_207 = __p0_207; \
60284  uint32x4_t __s1_207 = __p1_207; \
60285  uint16x4_t __rev0_207; __rev0_207 = __builtin_shufflevector(__s0_207, __s0_207, 3, 2, 1, 0); \
60286  uint32x4_t __rev1_207; __rev1_207 = __builtin_shufflevector(__s1_207, __s1_207, 3, 2, 1, 0); \
60287  uint16x8_t __ret_207; \
60288  __ret_207 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_207), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_207, __p2_207)))); \
60289  __ret_207 = __builtin_shufflevector(__ret_207, __ret_207, 7, 6, 5, 4, 3, 2, 1, 0); \
60290  __ret_207; \
60291 })
60292 #endif
60293 
60294 #ifdef __LITTLE_ENDIAN__
60295 #define vrshrn_high_n_u64(__p0_208, __p1_208, __p2_208) __extension__ ({ \
60296  uint32x2_t __s0_208 = __p0_208; \
60297  uint64x2_t __s1_208 = __p1_208; \
60298  uint32x4_t __ret_208; \
60299  __ret_208 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_208), (uint32x2_t)(vrshrn_n_u64(__s1_208, __p2_208)))); \
60300  __ret_208; \
60301 })
60302 #else
60303 #define vrshrn_high_n_u64(__p0_209, __p1_209, __p2_209) __extension__ ({ \
60304  uint32x2_t __s0_209 = __p0_209; \
60305  uint64x2_t __s1_209 = __p1_209; \
60306  uint32x2_t __rev0_209; __rev0_209 = __builtin_shufflevector(__s0_209, __s0_209, 1, 0); \
60307  uint64x2_t __rev1_209; __rev1_209 = __builtin_shufflevector(__s1_209, __s1_209, 1, 0); \
60308  uint32x4_t __ret_209; \
60309  __ret_209 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_209), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_209, __p2_209)))); \
60310  __ret_209 = __builtin_shufflevector(__ret_209, __ret_209, 3, 2, 1, 0); \
60311  __ret_209; \
60312 })
60313 #endif
60314 
60315 #ifdef __LITTLE_ENDIAN__
60316 #define vrshrn_high_n_u16(__p0_210, __p1_210, __p2_210) __extension__ ({ \
60317  uint8x8_t __s0_210 = __p0_210; \
60318  uint16x8_t __s1_210 = __p1_210; \
60319  uint8x16_t __ret_210; \
60320  __ret_210 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_210), (uint8x8_t)(vrshrn_n_u16(__s1_210, __p2_210)))); \
60321  __ret_210; \
60322 })
60323 #else
60324 #define vrshrn_high_n_u16(__p0_211, __p1_211, __p2_211) __extension__ ({ \
60325  uint8x8_t __s0_211 = __p0_211; \
60326  uint16x8_t __s1_211 = __p1_211; \
60327  uint8x8_t __rev0_211; __rev0_211 = __builtin_shufflevector(__s0_211, __s0_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60328  uint16x8_t __rev1_211; __rev1_211 = __builtin_shufflevector(__s1_211, __s1_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60329  uint8x16_t __ret_211; \
60330  __ret_211 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_211), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_211, __p2_211)))); \
60331  __ret_211 = __builtin_shufflevector(__ret_211, __ret_211, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60332  __ret_211; \
60333 })
60334 #endif
60335 
60336 #ifdef __LITTLE_ENDIAN__
60337 #define vrshrn_high_n_s32(__p0_212, __p1_212, __p2_212) __extension__ ({ \
60338  int16x4_t __s0_212 = __p0_212; \
60339  int32x4_t __s1_212 = __p1_212; \
60340  int16x8_t __ret_212; \
60341  __ret_212 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_212), (int16x4_t)(vrshrn_n_s32(__s1_212, __p2_212)))); \
60342  __ret_212; \
60343 })
60344 #else
60345 #define vrshrn_high_n_s32(__p0_213, __p1_213, __p2_213) __extension__ ({ \
60346  int16x4_t __s0_213 = __p0_213; \
60347  int32x4_t __s1_213 = __p1_213; \
60348  int16x4_t __rev0_213; __rev0_213 = __builtin_shufflevector(__s0_213, __s0_213, 3, 2, 1, 0); \
60349  int32x4_t __rev1_213; __rev1_213 = __builtin_shufflevector(__s1_213, __s1_213, 3, 2, 1, 0); \
60350  int16x8_t __ret_213; \
60351  __ret_213 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_213), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_213, __p2_213)))); \
60352  __ret_213 = __builtin_shufflevector(__ret_213, __ret_213, 7, 6, 5, 4, 3, 2, 1, 0); \
60353  __ret_213; \
60354 })
60355 #endif
60356 
60357 #ifdef __LITTLE_ENDIAN__
60358 #define vrshrn_high_n_s64(__p0_214, __p1_214, __p2_214) __extension__ ({ \
60359  int32x2_t __s0_214 = __p0_214; \
60360  int64x2_t __s1_214 = __p1_214; \
60361  int32x4_t __ret_214; \
60362  __ret_214 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_214), (int32x2_t)(vrshrn_n_s64(__s1_214, __p2_214)))); \
60363  __ret_214; \
60364 })
60365 #else
60366 #define vrshrn_high_n_s64(__p0_215, __p1_215, __p2_215) __extension__ ({ \
60367  int32x2_t __s0_215 = __p0_215; \
60368  int64x2_t __s1_215 = __p1_215; \
60369  int32x2_t __rev0_215; __rev0_215 = __builtin_shufflevector(__s0_215, __s0_215, 1, 0); \
60370  int64x2_t __rev1_215; __rev1_215 = __builtin_shufflevector(__s1_215, __s1_215, 1, 0); \
60371  int32x4_t __ret_215; \
60372  __ret_215 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_215), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_215, __p2_215)))); \
60373  __ret_215 = __builtin_shufflevector(__ret_215, __ret_215, 3, 2, 1, 0); \
60374  __ret_215; \
60375 })
60376 #endif
60377 
60378 #ifdef __LITTLE_ENDIAN__
60379 #define vrshrn_high_n_s16(__p0_216, __p1_216, __p2_216) __extension__ ({ \
60380  int8x8_t __s0_216 = __p0_216; \
60381  int16x8_t __s1_216 = __p1_216; \
60382  int8x16_t __ret_216; \
60383  __ret_216 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_216), (int8x8_t)(vrshrn_n_s16(__s1_216, __p2_216)))); \
60384  __ret_216; \
60385 })
60386 #else
60387 #define vrshrn_high_n_s16(__p0_217, __p1_217, __p2_217) __extension__ ({ \
60388  int8x8_t __s0_217 = __p0_217; \
60389  int16x8_t __s1_217 = __p1_217; \
60390  int8x8_t __rev0_217; __rev0_217 = __builtin_shufflevector(__s0_217, __s0_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60391  int16x8_t __rev1_217; __rev1_217 = __builtin_shufflevector(__s1_217, __s1_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60392  int8x16_t __ret_217; \
60393  __ret_217 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_217), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_217, __p2_217)))); \
60394  __ret_217 = __builtin_shufflevector(__ret_217, __ret_217, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60395  __ret_217; \
60396 })
60397 #endif
60398 
60399 #ifdef __LITTLE_ENDIAN__
60400 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60401  float64x2_t __ret;
60402  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
60403  return __ret;
60404 }
60405 #else
60406 __ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60407  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60408  float64x2_t __ret;
60409  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
60410  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60411  return __ret;
60412 }
60413 #endif
60414 
60415 #ifdef __LITTLE_ENDIAN__
60416 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60417  float64x1_t __ret;
60418  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60419  return __ret;
60420 }
60421 #else
60422 __ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60423  float64x1_t __ret;
60424  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60425  return __ret;
60426 }
60427 #endif
60428 
60429 #ifdef __LITTLE_ENDIAN__
60430 __ai float64_t vrsqrted_f64(float64_t __p0) {
60431  float64_t __ret;
60432  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60433  return __ret;
60434 }
60435 #else
60436 __ai float64_t vrsqrted_f64(float64_t __p0) {
60437  float64_t __ret;
60438  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60439  return __ret;
60440 }
60441 #endif
60442 
60443 #ifdef __LITTLE_ENDIAN__
60444 __ai float32_t vrsqrtes_f32(float32_t __p0) {
60445  float32_t __ret;
60446  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60447  return __ret;
60448 }
60449 #else
60450 __ai float32_t vrsqrtes_f32(float32_t __p0) {
60451  float32_t __ret;
60452  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60453  return __ret;
60454 }
60455 #endif
60456 
60457 #ifdef __LITTLE_ENDIAN__
60458 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60459  float64x2_t __ret;
60460  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60461  return __ret;
60462 }
60463 #else
60464 __ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60465  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60466  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60467  float64x2_t __ret;
60468  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60470  return __ret;
60471 }
60472 #endif
60473 
60474 #ifdef __LITTLE_ENDIAN__
60475 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60476  float64x1_t __ret;
60477  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60478  return __ret;
60479 }
60480 #else
60481 __ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60482  float64x1_t __ret;
60483  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60484  return __ret;
60485 }
60486 #endif
60487 
60488 #ifdef __LITTLE_ENDIAN__
60489 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60490  float64_t __ret;
60491  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60492  return __ret;
60493 }
60494 #else
60495 __ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60496  float64_t __ret;
60497  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60498  return __ret;
60499 }
60500 #endif
60501 
60502 #ifdef __LITTLE_ENDIAN__
60503 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60504  float32_t __ret;
60505  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60506  return __ret;
60507 }
60508 #else
60509 __ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60510  float32_t __ret;
60511  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60512  return __ret;
60513 }
60514 #endif
60515 
60516 #ifdef __LITTLE_ENDIAN__
60517 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60518  uint64_t __s0 = __p0; \
60519  uint64_t __s1 = __p1; \
60520  uint64_t __ret; \
60521  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60522  __ret; \
60523 })
60524 #else
60525 #define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60526  uint64_t __s0 = __p0; \
60527  uint64_t __s1 = __p1; \
60528  uint64_t __ret; \
60529  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60530  __ret; \
60531 })
60532 #endif
60533 
60534 #ifdef __LITTLE_ENDIAN__
60535 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60536  int64_t __s0 = __p0; \
60537  int64_t __s1 = __p1; \
60538  int64_t __ret; \
60539  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60540  __ret; \
60541 })
60542 #else
60543 #define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60544  int64_t __s0 = __p0; \
60545  int64_t __s1 = __p1; \
60546  int64_t __ret; \
60547  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60548  __ret; \
60549 })
60550 #endif
60551 
60552 #ifdef __LITTLE_ENDIAN__
60553 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60554  uint16x8_t __ret;
60555  __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
60556  return __ret;
60557 }
60558 #else
60559 __ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60560  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60561  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60562  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60563  uint16x8_t __ret;
60564  __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
60565  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60566  return __ret;
60567 }
60568 #endif
60569 
60570 #ifdef __LITTLE_ENDIAN__
60571 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60572  uint32x4_t __ret;
60573  __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
60574  return __ret;
60575 }
60576 #else
60577 __ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60578  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60579  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60580  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60581  uint32x4_t __ret;
60582  __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
60583  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60584  return __ret;
60585 }
60586 #endif
60587 
60588 #ifdef __LITTLE_ENDIAN__
60589 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60590  uint8x16_t __ret;
60591  __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
60592  return __ret;
60593 }
60594 #else
60595 __ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60596  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60597  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60598  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60599  uint8x16_t __ret;
60600  __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
60601  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60602  return __ret;
60603 }
60604 #endif
60605 
60606 #ifdef __LITTLE_ENDIAN__
60607 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60608  int16x8_t __ret;
60609  __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
60610  return __ret;
60611 }
60612 #else
60613 __ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60614  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60615  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60616  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60617  int16x8_t __ret;
60618  __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
60619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60620  return __ret;
60621 }
60622 #endif
60623 
60624 #ifdef __LITTLE_ENDIAN__
60625 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60626  int32x4_t __ret;
60627  __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
60628  return __ret;
60629 }
60630 #else
60631 __ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60632  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60633  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60634  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60635  int32x4_t __ret;
60636  __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
60637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60638  return __ret;
60639 }
60640 #endif
60641 
60642 #ifdef __LITTLE_ENDIAN__
60643 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60644  int8x16_t __ret;
60645  __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
60646  return __ret;
60647 }
60648 #else
60649 __ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60650  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60651  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60652  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60653  int8x16_t __ret;
60654  __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
60655  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60656  return __ret;
60657 }
60658 #endif
60659 
60660 #ifdef __LITTLE_ENDIAN__
60661 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60662  poly64_t __s0 = __p0; \
60663  poly64x1_t __s1 = __p1; \
60664  poly64x1_t __ret; \
60665  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60666  __ret; \
60667 })
60668 #else
60669 #define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60670  poly64_t __s0 = __p0; \
60671  poly64x1_t __s1 = __p1; \
60672  poly64x1_t __ret; \
60673  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60674  __ret; \
60675 })
60676 #define __noswap_vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60677  poly64_t __s0 = __p0; \
60678  poly64x1_t __s1 = __p1; \
60679  poly64x1_t __ret; \
60680  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60681  __ret; \
60682 })
60683 #endif
60684 
60685 #ifdef __LITTLE_ENDIAN__
60686 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60687  poly64_t __s0 = __p0; \
60688  poly64x2_t __s1 = __p1; \
60689  poly64x2_t __ret; \
60690  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60691  __ret; \
60692 })
60693 #else
60694 #define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60695  poly64_t __s0 = __p0; \
60696  poly64x2_t __s1 = __p1; \
60697  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60698  poly64x2_t __ret; \
60699  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
60700  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60701  __ret; \
60702 })
60703 #define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60704  poly64_t __s0 = __p0; \
60705  poly64x2_t __s1 = __p1; \
60706  poly64x2_t __ret; \
60707  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60708  __ret; \
60709 })
60710 #endif
60711 
60712 #ifdef __LITTLE_ENDIAN__
60713 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60714  float64_t __s0 = __p0; \
60715  float64x2_t __s1 = __p1; \
60716  float64x2_t __ret; \
60717  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60718  __ret; \
60719 })
60720 #else
60721 #define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60722  float64_t __s0 = __p0; \
60723  float64x2_t __s1 = __p1; \
60724  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60725  float64x2_t __ret; \
60726  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__rev1, __p2); \
60727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60728  __ret; \
60729 })
60730 #define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60731  float64_t __s0 = __p0; \
60732  float64x2_t __s1 = __p1; \
60733  float64x2_t __ret; \
60734  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60735  __ret; \
60736 })
60737 #endif
60738 
60739 #ifdef __LITTLE_ENDIAN__
60740 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60741  float64_t __s0 = __p0; \
60742  float64x1_t __s1 = __p1; \
60743  float64x1_t __ret; \
60744  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60745  __ret; \
60746 })
60747 #else
60748 #define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60749  float64_t __s0 = __p0; \
60750  float64x1_t __s1 = __p1; \
60751  float64x1_t __ret; \
60752  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60753  __ret; \
60754 })
60755 #define __noswap_vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60756  float64_t __s0 = __p0; \
60757  float64x1_t __s1 = __p1; \
60758  float64x1_t __ret; \
60759  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60760  __ret; \
60761 })
60762 #endif
60763 
60764 #ifdef __LITTLE_ENDIAN__
60765 __ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60766  uint64_t __ret;
60767  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60768  return __ret;
60769 }
60770 #else
60771 __ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60772  uint64_t __ret;
60773  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60774  return __ret;
60775 }
60776 #endif
60777 
60778 #ifdef __LITTLE_ENDIAN__
60779 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60780  int64_t __ret;
60781  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60782  return __ret;
60783 }
60784 #else
60785 __ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60786  int64_t __ret;
60787  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60788  return __ret;
60789 }
60790 #endif
60791 
60792 #ifdef __LITTLE_ENDIAN__
60793 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
60794  uint64_t __s0 = __p0; \
60795  uint64_t __ret; \
60796  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60797  __ret; \
60798 })
60799 #else
60800 #define vshld_n_u64(__p0, __p1) __extension__ ({ \
60801  uint64_t __s0 = __p0; \
60802  uint64_t __ret; \
60803  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60804  __ret; \
60805 })
60806 #endif
60807 
60808 #ifdef __LITTLE_ENDIAN__
60809 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
60810  int64_t __s0 = __p0; \
60811  int64_t __ret; \
60812  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60813  __ret; \
60814 })
60815 #else
60816 #define vshld_n_s64(__p0, __p1) __extension__ ({ \
60817  int64_t __s0 = __p0; \
60818  int64_t __ret; \
60819  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60820  __ret; \
60821 })
60822 #endif
60823 
60824 #ifdef __LITTLE_ENDIAN__
60825 #define vshll_high_n_u8(__p0_218, __p1_218) __extension__ ({ \
60826  uint8x16_t __s0_218 = __p0_218; \
60827  uint16x8_t __ret_218; \
60828  __ret_218 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_218), __p1_218)); \
60829  __ret_218; \
60830 })
60831 #else
60832 #define vshll_high_n_u8(__p0_219, __p1_219) __extension__ ({ \
60833  uint8x16_t __s0_219 = __p0_219; \
60834  uint8x16_t __rev0_219; __rev0_219 = __builtin_shufflevector(__s0_219, __s0_219, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60835  uint16x8_t __ret_219; \
60836  __ret_219 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_219), __p1_219)); \
60837  __ret_219 = __builtin_shufflevector(__ret_219, __ret_219, 7, 6, 5, 4, 3, 2, 1, 0); \
60838  __ret_219; \
60839 })
60840 #endif
60841 
60842 #ifdef __LITTLE_ENDIAN__
60843 #define vshll_high_n_u32(__p0_220, __p1_220) __extension__ ({ \
60844  uint32x4_t __s0_220 = __p0_220; \
60845  uint64x2_t __ret_220; \
60846  __ret_220 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_220), __p1_220)); \
60847  __ret_220; \
60848 })
60849 #else
60850 #define vshll_high_n_u32(__p0_221, __p1_221) __extension__ ({ \
60851  uint32x4_t __s0_221 = __p0_221; \
60852  uint32x4_t __rev0_221; __rev0_221 = __builtin_shufflevector(__s0_221, __s0_221, 3, 2, 1, 0); \
60853  uint64x2_t __ret_221; \
60854  __ret_221 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_221), __p1_221)); \
60855  __ret_221 = __builtin_shufflevector(__ret_221, __ret_221, 1, 0); \
60856  __ret_221; \
60857 })
60858 #endif
60859 
60860 #ifdef __LITTLE_ENDIAN__
60861 #define vshll_high_n_u16(__p0_222, __p1_222) __extension__ ({ \
60862  uint16x8_t __s0_222 = __p0_222; \
60863  uint32x4_t __ret_222; \
60864  __ret_222 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_222), __p1_222)); \
60865  __ret_222; \
60866 })
60867 #else
60868 #define vshll_high_n_u16(__p0_223, __p1_223) __extension__ ({ \
60869  uint16x8_t __s0_223 = __p0_223; \
60870  uint16x8_t __rev0_223; __rev0_223 = __builtin_shufflevector(__s0_223, __s0_223, 7, 6, 5, 4, 3, 2, 1, 0); \
60871  uint32x4_t __ret_223; \
60872  __ret_223 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_223), __p1_223)); \
60873  __ret_223 = __builtin_shufflevector(__ret_223, __ret_223, 3, 2, 1, 0); \
60874  __ret_223; \
60875 })
60876 #endif
60877 
60878 #ifdef __LITTLE_ENDIAN__
60879 #define vshll_high_n_s8(__p0_224, __p1_224) __extension__ ({ \
60880  int8x16_t __s0_224 = __p0_224; \
60881  int16x8_t __ret_224; \
60882  __ret_224 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_224), __p1_224)); \
60883  __ret_224; \
60884 })
60885 #else
60886 #define vshll_high_n_s8(__p0_225, __p1_225) __extension__ ({ \
60887  int8x16_t __s0_225 = __p0_225; \
60888  int8x16_t __rev0_225; __rev0_225 = __builtin_shufflevector(__s0_225, __s0_225, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60889  int16x8_t __ret_225; \
60890  __ret_225 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_225), __p1_225)); \
60891  __ret_225 = __builtin_shufflevector(__ret_225, __ret_225, 7, 6, 5, 4, 3, 2, 1, 0); \
60892  __ret_225; \
60893 })
60894 #endif
60895 
60896 #ifdef __LITTLE_ENDIAN__
60897 #define vshll_high_n_s32(__p0_226, __p1_226) __extension__ ({ \
60898  int32x4_t __s0_226 = __p0_226; \
60899  int64x2_t __ret_226; \
60900  __ret_226 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_226), __p1_226)); \
60901  __ret_226; \
60902 })
60903 #else
60904 #define vshll_high_n_s32(__p0_227, __p1_227) __extension__ ({ \
60905  int32x4_t __s0_227 = __p0_227; \
60906  int32x4_t __rev0_227; __rev0_227 = __builtin_shufflevector(__s0_227, __s0_227, 3, 2, 1, 0); \
60907  int64x2_t __ret_227; \
60908  __ret_227 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_227), __p1_227)); \
60909  __ret_227 = __builtin_shufflevector(__ret_227, __ret_227, 1, 0); \
60910  __ret_227; \
60911 })
60912 #endif
60913 
60914 #ifdef __LITTLE_ENDIAN__
60915 #define vshll_high_n_s16(__p0_228, __p1_228) __extension__ ({ \
60916  int16x8_t __s0_228 = __p0_228; \
60917  int32x4_t __ret_228; \
60918  __ret_228 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_228), __p1_228)); \
60919  __ret_228; \
60920 })
60921 #else
60922 #define vshll_high_n_s16(__p0_229, __p1_229) __extension__ ({ \
60923  int16x8_t __s0_229 = __p0_229; \
60924  int16x8_t __rev0_229; __rev0_229 = __builtin_shufflevector(__s0_229, __s0_229, 7, 6, 5, 4, 3, 2, 1, 0); \
60925  int32x4_t __ret_229; \
60926  __ret_229 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_229), __p1_229)); \
60927  __ret_229 = __builtin_shufflevector(__ret_229, __ret_229, 3, 2, 1, 0); \
60928  __ret_229; \
60929 })
60930 #endif
60931 
60932 #ifdef __LITTLE_ENDIAN__
60933 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60934  uint64_t __s0 = __p0; \
60935  uint64_t __ret; \
60936  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60937  __ret; \
60938 })
60939 #else
60940 #define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60941  uint64_t __s0 = __p0; \
60942  uint64_t __ret; \
60943  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60944  __ret; \
60945 })
60946 #endif
60947 
60948 #ifdef __LITTLE_ENDIAN__
60949 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60950  int64_t __s0 = __p0; \
60951  int64_t __ret; \
60952  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60953  __ret; \
60954 })
60955 #else
60956 #define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60957  int64_t __s0 = __p0; \
60958  int64_t __ret; \
60959  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60960  __ret; \
60961 })
60962 #endif
60963 
60964 #ifdef __LITTLE_ENDIAN__
60965 #define vshrn_high_n_u32(__p0_230, __p1_230, __p2_230) __extension__ ({ \
60966  uint16x4_t __s0_230 = __p0_230; \
60967  uint32x4_t __s1_230 = __p1_230; \
60968  uint16x8_t __ret_230; \
60969  __ret_230 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_230), (uint16x4_t)(vshrn_n_u32(__s1_230, __p2_230)))); \
60970  __ret_230; \
60971 })
60972 #else
60973 #define vshrn_high_n_u32(__p0_231, __p1_231, __p2_231) __extension__ ({ \
60974  uint16x4_t __s0_231 = __p0_231; \
60975  uint32x4_t __s1_231 = __p1_231; \
60976  uint16x4_t __rev0_231; __rev0_231 = __builtin_shufflevector(__s0_231, __s0_231, 3, 2, 1, 0); \
60977  uint32x4_t __rev1_231; __rev1_231 = __builtin_shufflevector(__s1_231, __s1_231, 3, 2, 1, 0); \
60978  uint16x8_t __ret_231; \
60979  __ret_231 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_231), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_231, __p2_231)))); \
60980  __ret_231 = __builtin_shufflevector(__ret_231, __ret_231, 7, 6, 5, 4, 3, 2, 1, 0); \
60981  __ret_231; \
60982 })
60983 #endif
60984 
60985 #ifdef __LITTLE_ENDIAN__
60986 #define vshrn_high_n_u64(__p0_232, __p1_232, __p2_232) __extension__ ({ \
60987  uint32x2_t __s0_232 = __p0_232; \
60988  uint64x2_t __s1_232 = __p1_232; \
60989  uint32x4_t __ret_232; \
60990  __ret_232 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_232), (uint32x2_t)(vshrn_n_u64(__s1_232, __p2_232)))); \
60991  __ret_232; \
60992 })
60993 #else
60994 #define vshrn_high_n_u64(__p0_233, __p1_233, __p2_233) __extension__ ({ \
60995  uint32x2_t __s0_233 = __p0_233; \
60996  uint64x2_t __s1_233 = __p1_233; \
60997  uint32x2_t __rev0_233; __rev0_233 = __builtin_shufflevector(__s0_233, __s0_233, 1, 0); \
60998  uint64x2_t __rev1_233; __rev1_233 = __builtin_shufflevector(__s1_233, __s1_233, 1, 0); \
60999  uint32x4_t __ret_233; \
61000  __ret_233 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_233), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_233, __p2_233)))); \
61001  __ret_233 = __builtin_shufflevector(__ret_233, __ret_233, 3, 2, 1, 0); \
61002  __ret_233; \
61003 })
61004 #endif
61005 
61006 #ifdef __LITTLE_ENDIAN__
61007 #define vshrn_high_n_u16(__p0_234, __p1_234, __p2_234) __extension__ ({ \
61008  uint8x8_t __s0_234 = __p0_234; \
61009  uint16x8_t __s1_234 = __p1_234; \
61010  uint8x16_t __ret_234; \
61011  __ret_234 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_234), (uint8x8_t)(vshrn_n_u16(__s1_234, __p2_234)))); \
61012  __ret_234; \
61013 })
61014 #else
61015 #define vshrn_high_n_u16(__p0_235, __p1_235, __p2_235) __extension__ ({ \
61016  uint8x8_t __s0_235 = __p0_235; \
61017  uint16x8_t __s1_235 = __p1_235; \
61018  uint8x8_t __rev0_235; __rev0_235 = __builtin_shufflevector(__s0_235, __s0_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61019  uint16x8_t __rev1_235; __rev1_235 = __builtin_shufflevector(__s1_235, __s1_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61020  uint8x16_t __ret_235; \
61021  __ret_235 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_235), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_235, __p2_235)))); \
61022  __ret_235 = __builtin_shufflevector(__ret_235, __ret_235, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61023  __ret_235; \
61024 })
61025 #endif
61026 
61027 #ifdef __LITTLE_ENDIAN__
61028 #define vshrn_high_n_s32(__p0_236, __p1_236, __p2_236) __extension__ ({ \
61029  int16x4_t __s0_236 = __p0_236; \
61030  int32x4_t __s1_236 = __p1_236; \
61031  int16x8_t __ret_236; \
61032  __ret_236 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_236), (int16x4_t)(vshrn_n_s32(__s1_236, __p2_236)))); \
61033  __ret_236; \
61034 })
61035 #else
61036 #define vshrn_high_n_s32(__p0_237, __p1_237, __p2_237) __extension__ ({ \
61037  int16x4_t __s0_237 = __p0_237; \
61038  int32x4_t __s1_237 = __p1_237; \
61039  int16x4_t __rev0_237; __rev0_237 = __builtin_shufflevector(__s0_237, __s0_237, 3, 2, 1, 0); \
61040  int32x4_t __rev1_237; __rev1_237 = __builtin_shufflevector(__s1_237, __s1_237, 3, 2, 1, 0); \
61041  int16x8_t __ret_237; \
61042  __ret_237 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_237), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_237, __p2_237)))); \
61043  __ret_237 = __builtin_shufflevector(__ret_237, __ret_237, 7, 6, 5, 4, 3, 2, 1, 0); \
61044  __ret_237; \
61045 })
61046 #endif
61047 
61048 #ifdef __LITTLE_ENDIAN__
61049 #define vshrn_high_n_s64(__p0_238, __p1_238, __p2_238) __extension__ ({ \
61050  int32x2_t __s0_238 = __p0_238; \
61051  int64x2_t __s1_238 = __p1_238; \
61052  int32x4_t __ret_238; \
61053  __ret_238 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_238), (int32x2_t)(vshrn_n_s64(__s1_238, __p2_238)))); \
61054  __ret_238; \
61055 })
61056 #else
61057 #define vshrn_high_n_s64(__p0_239, __p1_239, __p2_239) __extension__ ({ \
61058  int32x2_t __s0_239 = __p0_239; \
61059  int64x2_t __s1_239 = __p1_239; \
61060  int32x2_t __rev0_239; __rev0_239 = __builtin_shufflevector(__s0_239, __s0_239, 1, 0); \
61061  int64x2_t __rev1_239; __rev1_239 = __builtin_shufflevector(__s1_239, __s1_239, 1, 0); \
61062  int32x4_t __ret_239; \
61063  __ret_239 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_239), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_239, __p2_239)))); \
61064  __ret_239 = __builtin_shufflevector(__ret_239, __ret_239, 3, 2, 1, 0); \
61065  __ret_239; \
61066 })
61067 #endif
61068 
61069 #ifdef __LITTLE_ENDIAN__
61070 #define vshrn_high_n_s16(__p0_240, __p1_240, __p2_240) __extension__ ({ \
61071  int8x8_t __s0_240 = __p0_240; \
61072  int16x8_t __s1_240 = __p1_240; \
61073  int8x16_t __ret_240; \
61074  __ret_240 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_240), (int8x8_t)(vshrn_n_s16(__s1_240, __p2_240)))); \
61075  __ret_240; \
61076 })
61077 #else
61078 #define vshrn_high_n_s16(__p0_241, __p1_241, __p2_241) __extension__ ({ \
61079  int8x8_t __s0_241 = __p0_241; \
61080  int16x8_t __s1_241 = __p1_241; \
61081  int8x8_t __rev0_241; __rev0_241 = __builtin_shufflevector(__s0_241, __s0_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61082  int16x8_t __rev1_241; __rev1_241 = __builtin_shufflevector(__s1_241, __s1_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61083  int8x16_t __ret_241; \
61084  __ret_241 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_241), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_241, __p2_241)))); \
61085  __ret_241 = __builtin_shufflevector(__ret_241, __ret_241, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61086  __ret_241; \
61087 })
61088 #endif
61089 
61090 #ifdef __LITTLE_ENDIAN__
61091 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61092  uint64_t __s0 = __p0; \
61093  uint64_t __s1 = __p1; \
61094  uint64_t __ret; \
61095  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61096  __ret; \
61097 })
61098 #else
61099 #define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61100  uint64_t __s0 = __p0; \
61101  uint64_t __s1 = __p1; \
61102  uint64_t __ret; \
61103  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61104  __ret; \
61105 })
61106 #endif
61107 
61108 #ifdef __LITTLE_ENDIAN__
61109 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61110  int64_t __s0 = __p0; \
61111  int64_t __s1 = __p1; \
61112  int64_t __ret; \
61113  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61114  __ret; \
61115 })
61116 #else
61117 #define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61118  int64_t __s0 = __p0; \
61119  int64_t __s1 = __p1; \
61120  int64_t __ret; \
61121  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61122  __ret; \
61123 })
61124 #endif
61125 
61126 #ifdef __LITTLE_ENDIAN__
61127 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61128  poly64x1_t __s0 = __p0; \
61129  poly64x1_t __s1 = __p1; \
61130  poly64x1_t __ret; \
61131  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61132  __ret; \
61133 })
61134 #else
61135 #define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61136  poly64x1_t __s0 = __p0; \
61137  poly64x1_t __s1 = __p1; \
61138  poly64x1_t __ret; \
61139  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61140  __ret; \
61141 })
61142 #endif
61143 
61144 #ifdef __LITTLE_ENDIAN__
61145 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61146  poly64x2_t __s0 = __p0; \
61147  poly64x2_t __s1 = __p1; \
61148  poly64x2_t __ret; \
61149  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61150  __ret; \
61151 })
61152 #else
61153 #define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61154  poly64x2_t __s0 = __p0; \
61155  poly64x2_t __s1 = __p1; \
61156  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61157  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61158  poly64x2_t __ret; \
61159  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61160  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61161  __ret; \
61162 })
61163 #endif
61164 
61165 #ifdef __LITTLE_ENDIAN__
61166 __ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61167  uint8_t __ret;
61168  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61169  return __ret;
61170 }
61171 #else
61172 __ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61173  uint8_t __ret;
61174  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61175  return __ret;
61176 }
61177 #endif
61178 
61179 #ifdef __LITTLE_ENDIAN__
61180 __ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61181  uint32_t __ret;
61182  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61183  return __ret;
61184 }
61185 #else
61186 __ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61187  uint32_t __ret;
61188  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61189  return __ret;
61190 }
61191 #endif
61192 
61193 #ifdef __LITTLE_ENDIAN__
61194 __ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61195  uint64_t __ret;
61196  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61197  return __ret;
61198 }
61199 #else
61200 __ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61201  uint64_t __ret;
61202  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61203  return __ret;
61204 }
61205 #endif
61206 
61207 #ifdef __LITTLE_ENDIAN__
61208 __ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61209  uint16_t __ret;
61210  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61211  return __ret;
61212 }
61213 #else
61214 __ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61215  uint16_t __ret;
61216  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61217  return __ret;
61218 }
61219 #endif
61220 
61221 #ifdef __LITTLE_ENDIAN__
61222 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61223  uint8x16_t __ret;
61224  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
61225  return __ret;
61226 }
61227 #else
61228 __ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61229  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61230  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61231  uint8x16_t __ret;
61232  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
61233  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61234  return __ret;
61235 }
61236 #endif
61237 
61238 #ifdef __LITTLE_ENDIAN__
61239 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61240  uint32x4_t __ret;
61241  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
61242  return __ret;
61243 }
61244 #else
61245 __ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61246  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61247  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61248  uint32x4_t __ret;
61249  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
61250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61251  return __ret;
61252 }
61253 #endif
61254 
61255 #ifdef __LITTLE_ENDIAN__
61256 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61257  uint64x2_t __ret;
61258  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
61259  return __ret;
61260 }
61261 #else
61262 __ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61263  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61264  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61265  uint64x2_t __ret;
61266  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
61267  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61268  return __ret;
61269 }
61270 #endif
61271 
61272 #ifdef __LITTLE_ENDIAN__
61273 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61274  uint16x8_t __ret;
61275  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
61276  return __ret;
61277 }
61278 #else
61279 __ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61280  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61281  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61282  uint16x8_t __ret;
61283  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
61284  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61285  return __ret;
61286 }
61287 #endif
61288 
61289 #ifdef __LITTLE_ENDIAN__
61290 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61291  uint8x8_t __ret;
61292  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
61293  return __ret;
61294 }
61295 #else
61296 __ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61297  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61298  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61299  uint8x8_t __ret;
61300  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
61301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61302  return __ret;
61303 }
61304 #endif
61305 
61306 #ifdef __LITTLE_ENDIAN__
61307 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61308  uint32x2_t __ret;
61309  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
61310  return __ret;
61311 }
61312 #else
61313 __ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61314  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61315  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61316  uint32x2_t __ret;
61317  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
61318  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61319  return __ret;
61320 }
61321 #endif
61322 
61323 #ifdef __LITTLE_ENDIAN__
61324 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61325  uint64x1_t __ret;
61326  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61327  return __ret;
61328 }
61329 #else
61330 __ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61331  uint64x1_t __ret;
61332  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61333  return __ret;
61334 }
61335 #endif
61336 
61337 #ifdef __LITTLE_ENDIAN__
61338 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61339  uint16x4_t __ret;
61340  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
61341  return __ret;
61342 }
61343 #else
61344 __ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61345  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61346  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61347  uint16x4_t __ret;
61348  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
61349  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61350  return __ret;
61351 }
61352 #endif
61353 
61354 #ifdef __LITTLE_ENDIAN__
61355 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61356  float64x2_t __ret;
61357  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
61358  return __ret;
61359 }
61360 #else
61361 __ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61362  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61363  float64x2_t __ret;
61364  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
61365  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61366  return __ret;
61367 }
61368 #endif
61369 
61370 #ifdef __LITTLE_ENDIAN__
61371 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61372  float32x4_t __ret;
61373  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
61374  return __ret;
61375 }
61376 #else
61377 __ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61378  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61379  float32x4_t __ret;
61380  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
61381  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61382  return __ret;
61383 }
61384 #endif
61385 
61386 #ifdef __LITTLE_ENDIAN__
61387 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61388  float64x1_t __ret;
61389  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61390  return __ret;
61391 }
61392 #else
61393 __ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61394  float64x1_t __ret;
61395  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61396  return __ret;
61397 }
61398 #endif
61399 
61400 #ifdef __LITTLE_ENDIAN__
61401 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61402  float32x2_t __ret;
61403  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
61404  return __ret;
61405 }
61406 #else
61407 __ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61408  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61409  float32x2_t __ret;
61410  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
61411  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61412  return __ret;
61413 }
61414 #endif
61415 
61416 #ifdef __LITTLE_ENDIAN__
61417 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61418  uint64_t __s0 = __p0; \
61419  uint64_t __s1 = __p1; \
61420  uint64_t __ret; \
61421  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61422  __ret; \
61423 })
61424 #else
61425 #define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61426  uint64_t __s0 = __p0; \
61427  uint64_t __s1 = __p1; \
61428  uint64_t __ret; \
61429  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61430  __ret; \
61431 })
61432 #endif
61433 
61434 #ifdef __LITTLE_ENDIAN__
61435 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61436  int64_t __s0 = __p0; \
61437  int64_t __s1 = __p1; \
61438  int64_t __ret; \
61439  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61440  __ret; \
61441 })
61442 #else
61443 #define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61444  int64_t __s0 = __p0; \
61445  int64_t __s1 = __p1; \
61446  int64_t __ret; \
61447  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61448  __ret; \
61449 })
61450 #endif
61451 
61452 #ifdef __LITTLE_ENDIAN__
61453 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61454  uint64_t __s0 = __p0; \
61455  uint64_t __s1 = __p1; \
61456  uint64_t __ret; \
61457  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61458  __ret; \
61459 })
61460 #else
61461 #define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61462  uint64_t __s0 = __p0; \
61463  uint64_t __s1 = __p1; \
61464  uint64_t __ret; \
61465  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61466  __ret; \
61467 })
61468 #endif
61469 
61470 #ifdef __LITTLE_ENDIAN__
61471 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61472  int64_t __s0 = __p0; \
61473  int64_t __s1 = __p1; \
61474  int64_t __ret; \
61475  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61476  __ret; \
61477 })
61478 #else
61479 #define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61480  int64_t __s0 = __p0; \
61481  int64_t __s1 = __p1; \
61482  int64_t __ret; \
61483  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61484  __ret; \
61485 })
61486 #endif
61487 
61488 #ifdef __LITTLE_ENDIAN__
61489 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61490  poly64x1_t __s0 = __p0; \
61491  poly64x1_t __s1 = __p1; \
61492  poly64x1_t __ret; \
61493  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61494  __ret; \
61495 })
61496 #else
61497 #define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61498  poly64x1_t __s0 = __p0; \
61499  poly64x1_t __s1 = __p1; \
61500  poly64x1_t __ret; \
61501  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61502  __ret; \
61503 })
61504 #endif
61505 
61506 #ifdef __LITTLE_ENDIAN__
61507 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61508  poly64x2_t __s0 = __p0; \
61509  poly64x2_t __s1 = __p1; \
61510  poly64x2_t __ret; \
61511  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61512  __ret; \
61513 })
61514 #else
61515 #define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61516  poly64x2_t __s0 = __p0; \
61517  poly64x2_t __s1 = __p1; \
61518  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61519  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61520  poly64x2_t __ret; \
61521  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61522  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61523  __ret; \
61524 })
61525 #endif
61526 
61527 #ifdef __LITTLE_ENDIAN__
61528 #define vst1_p64(__p0, __p1) __extension__ ({ \
61529  poly64x1_t __s1 = __p1; \
61530  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61531 })
61532 #else
61533 #define vst1_p64(__p0, __p1) __extension__ ({ \
61534  poly64x1_t __s1 = __p1; \
61535  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61536 })
61537 #endif
61538 
61539 #ifdef __LITTLE_ENDIAN__
61540 #define vst1q_p64(__p0, __p1) __extension__ ({ \
61541  poly64x2_t __s1 = __p1; \
61542  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
61543 })
61544 #else
61545 #define vst1q_p64(__p0, __p1) __extension__ ({ \
61546  poly64x2_t __s1 = __p1; \
61547  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61548  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
61549 })
61550 #endif
61551 
61552 #ifdef __LITTLE_ENDIAN__
61553 #define vst1q_f64(__p0, __p1) __extension__ ({ \
61554  float64x2_t __s1 = __p1; \
61555  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
61556 })
61557 #else
61558 #define vst1q_f64(__p0, __p1) __extension__ ({ \
61559  float64x2_t __s1 = __p1; \
61560  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61561  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
61562 })
61563 #endif
61564 
61565 #ifdef __LITTLE_ENDIAN__
61566 #define vst1_f64(__p0, __p1) __extension__ ({ \
61567  float64x1_t __s1 = __p1; \
61568  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61569 })
61570 #else
61571 #define vst1_f64(__p0, __p1) __extension__ ({ \
61572  float64x1_t __s1 = __p1; \
61573  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61574 })
61575 #endif
61576 
61577 #ifdef __LITTLE_ENDIAN__
61578 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61579  poly64x1_t __s1 = __p1; \
61580  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61581 })
61582 #else
61583 #define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61584  poly64x1_t __s1 = __p1; \
61585  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61586 })
61587 #endif
61588 
61589 #ifdef __LITTLE_ENDIAN__
61590 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61591  poly64x2_t __s1 = __p1; \
61592  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
61593 })
61594 #else
61595 #define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61596  poly64x2_t __s1 = __p1; \
61597  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61598  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
61599 })
61600 #endif
61601 
61602 #ifdef __LITTLE_ENDIAN__
61603 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61604  float64x2_t __s1 = __p1; \
61605  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
61606 })
61607 #else
61608 #define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61609  float64x2_t __s1 = __p1; \
61610  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61611  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
61612 })
61613 #endif
61614 
61615 #ifdef __LITTLE_ENDIAN__
61616 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61617  float64x1_t __s1 = __p1; \
61618  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61619 })
61620 #else
61621 #define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61622  float64x1_t __s1 = __p1; \
61623  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61624 })
61625 #endif
61626 
61627 #ifdef __LITTLE_ENDIAN__
61628 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61629  poly8x8x2_t __s1 = __p1; \
61630  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
61631 })
61632 #else
61633 #define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61634  poly8x8x2_t __s1 = __p1; \
61635  poly8x8x2_t __rev1; \
61636  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61637  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61638  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
61639 })
61640 #endif
61641 
61642 #ifdef __LITTLE_ENDIAN__
61643 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61644  poly64x1x2_t __s1 = __p1; \
61645  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61646 })
61647 #else
61648 #define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61649  poly64x1x2_t __s1 = __p1; \
61650  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61651 })
61652 #endif
61653 
61654 #ifdef __LITTLE_ENDIAN__
61655 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61656  poly16x4x2_t __s1 = __p1; \
61657  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
61658 })
61659 #else
61660 #define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61661  poly16x4x2_t __s1 = __p1; \
61662  poly16x4x2_t __rev1; \
61663  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61664  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61665  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
61666 })
61667 #endif
61668 
61669 #ifdef __LITTLE_ENDIAN__
61670 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61671  poly8x16x2_t __s1 = __p1; \
61672  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
61673 })
61674 #else
61675 #define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61676  poly8x16x2_t __s1 = __p1; \
61677  poly8x16x2_t __rev1; \
61678  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61679  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61680  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
61681 })
61682 #endif
61683 
61684 #ifdef __LITTLE_ENDIAN__
61685 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61686  poly64x2x2_t __s1 = __p1; \
61687  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
61688 })
61689 #else
61690 #define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61691  poly64x2x2_t __s1 = __p1; \
61692  poly64x2x2_t __rev1; \
61693  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61694  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61695  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
61696 })
61697 #endif
61698 
61699 #ifdef __LITTLE_ENDIAN__
61700 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61701  poly16x8x2_t __s1 = __p1; \
61702  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
61703 })
61704 #else
61705 #define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61706  poly16x8x2_t __s1 = __p1; \
61707  poly16x8x2_t __rev1; \
61708  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61709  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61710  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
61711 })
61712 #endif
61713 
61714 #ifdef __LITTLE_ENDIAN__
61715 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61716  uint8x16x2_t __s1 = __p1; \
61717  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
61718 })
61719 #else
61720 #define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61721  uint8x16x2_t __s1 = __p1; \
61722  uint8x16x2_t __rev1; \
61723  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61724  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61725  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
61726 })
61727 #endif
61728 
61729 #ifdef __LITTLE_ENDIAN__
61730 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61731  uint32x4x2_t __s1 = __p1; \
61732  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
61733 })
61734 #else
61735 #define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61736  uint32x4x2_t __s1 = __p1; \
61737  uint32x4x2_t __rev1; \
61738  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61739  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61740  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
61741 })
61742 #endif
61743 
61744 #ifdef __LITTLE_ENDIAN__
61745 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61746  uint64x2x2_t __s1 = __p1; \
61747  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
61748 })
61749 #else
61750 #define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61751  uint64x2x2_t __s1 = __p1; \
61752  uint64x2x2_t __rev1; \
61753  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61754  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61755  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
61756 })
61757 #endif
61758 
61759 #ifdef __LITTLE_ENDIAN__
61760 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61761  uint16x8x2_t __s1 = __p1; \
61762  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
61763 })
61764 #else
61765 #define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61766  uint16x8x2_t __s1 = __p1; \
61767  uint16x8x2_t __rev1; \
61768  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61769  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61770  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
61771 })
61772 #endif
61773 
61774 #ifdef __LITTLE_ENDIAN__
61775 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61776  int8x16x2_t __s1 = __p1; \
61777  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
61778 })
61779 #else
61780 #define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61781  int8x16x2_t __s1 = __p1; \
61782  int8x16x2_t __rev1; \
61783  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61784  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61785  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
61786 })
61787 #endif
61788 
61789 #ifdef __LITTLE_ENDIAN__
61790 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61791  float64x2x2_t __s1 = __p1; \
61792  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 42); \
61793 })
61794 #else
61795 #define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61796  float64x2x2_t __s1 = __p1; \
61797  float64x2x2_t __rev1; \
61798  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61799  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61800  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
61801 })
61802 #endif
61803 
61804 #ifdef __LITTLE_ENDIAN__
61805 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61806  float32x4x2_t __s1 = __p1; \
61807  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 41); \
61808 })
61809 #else
61810 #define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61811  float32x4x2_t __s1 = __p1; \
61812  float32x4x2_t __rev1; \
61813  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61814  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61815  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
61816 })
61817 #endif
61818 
61819 #ifdef __LITTLE_ENDIAN__
61820 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61821  float16x8x2_t __s1 = __p1; \
61822  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 40); \
61823 })
61824 #else
61825 #define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61826  float16x8x2_t __s1 = __p1; \
61827  float16x8x2_t __rev1; \
61828  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61829  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61830  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
61831 })
61832 #endif
61833 
61834 #ifdef __LITTLE_ENDIAN__
61835 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61836  int32x4x2_t __s1 = __p1; \
61837  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 34); \
61838 })
61839 #else
61840 #define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61841  int32x4x2_t __s1 = __p1; \
61842  int32x4x2_t __rev1; \
61843  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61844  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61845  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
61846 })
61847 #endif
61848 
61849 #ifdef __LITTLE_ENDIAN__
61850 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61851  int64x2x2_t __s1 = __p1; \
61852  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 35); \
61853 })
61854 #else
61855 #define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61856  int64x2x2_t __s1 = __p1; \
61857  int64x2x2_t __rev1; \
61858  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61859  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61860  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
61861 })
61862 #endif
61863 
61864 #ifdef __LITTLE_ENDIAN__
61865 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61866  int16x8x2_t __s1 = __p1; \
61867  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 33); \
61868 })
61869 #else
61870 #define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61871  int16x8x2_t __s1 = __p1; \
61872  int16x8x2_t __rev1; \
61873  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61874  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61875  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
61876 })
61877 #endif
61878 
61879 #ifdef __LITTLE_ENDIAN__
61880 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61881  uint8x8x2_t __s1 = __p1; \
61882  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
61883 })
61884 #else
61885 #define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61886  uint8x8x2_t __s1 = __p1; \
61887  uint8x8x2_t __rev1; \
61888  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61889  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61890  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
61891 })
61892 #endif
61893 
61894 #ifdef __LITTLE_ENDIAN__
61895 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61896  uint32x2x2_t __s1 = __p1; \
61897  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
61898 })
61899 #else
61900 #define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61901  uint32x2x2_t __s1 = __p1; \
61902  uint32x2x2_t __rev1; \
61903  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61904  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61905  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
61906 })
61907 #endif
61908 
61909 #ifdef __LITTLE_ENDIAN__
61910 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61911  uint64x1x2_t __s1 = __p1; \
61912  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61913 })
61914 #else
61915 #define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61916  uint64x1x2_t __s1 = __p1; \
61917  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61918 })
61919 #endif
61920 
61921 #ifdef __LITTLE_ENDIAN__
61922 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61923  uint16x4x2_t __s1 = __p1; \
61924  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
61925 })
61926 #else
61927 #define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61928  uint16x4x2_t __s1 = __p1; \
61929  uint16x4x2_t __rev1; \
61930  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61931  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61932  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
61933 })
61934 #endif
61935 
61936 #ifdef __LITTLE_ENDIAN__
61937 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61938  int8x8x2_t __s1 = __p1; \
61939  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
61940 })
61941 #else
61942 #define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61943  int8x8x2_t __s1 = __p1; \
61944  int8x8x2_t __rev1; \
61945  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61946  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61947  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
61948 })
61949 #endif
61950 
61951 #ifdef __LITTLE_ENDIAN__
61952 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61953  float64x1x2_t __s1 = __p1; \
61954  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61955 })
61956 #else
61957 #define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61958  float64x1x2_t __s1 = __p1; \
61959  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61960 })
61961 #endif
61962 
61963 #ifdef __LITTLE_ENDIAN__
61964 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61965  float32x2x2_t __s1 = __p1; \
61966  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 9); \
61967 })
61968 #else
61969 #define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61970  float32x2x2_t __s1 = __p1; \
61971  float32x2x2_t __rev1; \
61972  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61973  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61974  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
61975 })
61976 #endif
61977 
61978 #ifdef __LITTLE_ENDIAN__
61979 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61980  float16x4x2_t __s1 = __p1; \
61981  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 8); \
61982 })
61983 #else
61984 #define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61985  float16x4x2_t __s1 = __p1; \
61986  float16x4x2_t __rev1; \
61987  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61988  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61989  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
61990 })
61991 #endif
61992 
61993 #ifdef __LITTLE_ENDIAN__
61994 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61995  int32x2x2_t __s1 = __p1; \
61996  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 2); \
61997 })
61998 #else
61999 #define vst1_s32_x2(__p0, __p1) __extension__ ({ \
62000  int32x2x2_t __s1 = __p1; \
62001  int32x2x2_t __rev1; \
62002  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62003  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62004  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
62005 })
62006 #endif
62007 
62008 #ifdef __LITTLE_ENDIAN__
62009 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62010  int64x1x2_t __s1 = __p1; \
62011  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62012 })
62013 #else
62014 #define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62015  int64x1x2_t __s1 = __p1; \
62016  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62017 })
62018 #endif
62019 
62020 #ifdef __LITTLE_ENDIAN__
62021 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62022  int16x4x2_t __s1 = __p1; \
62023  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 1); \
62024 })
62025 #else
62026 #define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62027  int16x4x2_t __s1 = __p1; \
62028  int16x4x2_t __rev1; \
62029  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62030  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62031  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
62032 })
62033 #endif
62034 
62035 #ifdef __LITTLE_ENDIAN__
62036 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62037  poly8x8x3_t __s1 = __p1; \
62038  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
62039 })
62040 #else
62041 #define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62042  poly8x8x3_t __s1 = __p1; \
62043  poly8x8x3_t __rev1; \
62044  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62045  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62046  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62047  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
62048 })
62049 #endif
62050 
62051 #ifdef __LITTLE_ENDIAN__
62052 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62053  poly64x1x3_t __s1 = __p1; \
62054  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62055 })
62056 #else
62057 #define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62058  poly64x1x3_t __s1 = __p1; \
62059  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62060 })
62061 #endif
62062 
62063 #ifdef __LITTLE_ENDIAN__
62064 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62065  poly16x4x3_t __s1 = __p1; \
62066  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
62067 })
62068 #else
62069 #define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62070  poly16x4x3_t __s1 = __p1; \
62071  poly16x4x3_t __rev1; \
62072  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62073  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62074  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62075  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
62076 })
62077 #endif
62078 
62079 #ifdef __LITTLE_ENDIAN__
62080 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62081  poly8x16x3_t __s1 = __p1; \
62082  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
62083 })
62084 #else
62085 #define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62086  poly8x16x3_t __s1 = __p1; \
62087  poly8x16x3_t __rev1; \
62088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62090  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62091  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
62092 })
62093 #endif
62094 
62095 #ifdef __LITTLE_ENDIAN__
62096 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62097  poly64x2x3_t __s1 = __p1; \
62098  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
62099 })
62100 #else
62101 #define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62102  poly64x2x3_t __s1 = __p1; \
62103  poly64x2x3_t __rev1; \
62104  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62105  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62106  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62107  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
62108 })
62109 #endif
62110 
62111 #ifdef __LITTLE_ENDIAN__
62112 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62113  poly16x8x3_t __s1 = __p1; \
62114  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
62115 })
62116 #else
62117 #define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62118  poly16x8x3_t __s1 = __p1; \
62119  poly16x8x3_t __rev1; \
62120  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62121  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62122  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62123  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
62124 })
62125 #endif
62126 
62127 #ifdef __LITTLE_ENDIAN__
62128 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62129  uint8x16x3_t __s1 = __p1; \
62130  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
62131 })
62132 #else
62133 #define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62134  uint8x16x3_t __s1 = __p1; \
62135  uint8x16x3_t __rev1; \
62136  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62137  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62138  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62139  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
62140 })
62141 #endif
62142 
62143 #ifdef __LITTLE_ENDIAN__
62144 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62145  uint32x4x3_t __s1 = __p1; \
62146  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
62147 })
62148 #else
62149 #define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62150  uint32x4x3_t __s1 = __p1; \
62151  uint32x4x3_t __rev1; \
62152  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62153  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62154  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62155  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
62156 })
62157 #endif
62158 
62159 #ifdef __LITTLE_ENDIAN__
62160 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62161  uint64x2x3_t __s1 = __p1; \
62162  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
62163 })
62164 #else
62165 #define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62166  uint64x2x3_t __s1 = __p1; \
62167  uint64x2x3_t __rev1; \
62168  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62169  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62170  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62171  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
62172 })
62173 #endif
62174 
62175 #ifdef __LITTLE_ENDIAN__
62176 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62177  uint16x8x3_t __s1 = __p1; \
62178  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
62179 })
62180 #else
62181 #define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62182  uint16x8x3_t __s1 = __p1; \
62183  uint16x8x3_t __rev1; \
62184  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62185  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62186  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62187  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
62188 })
62189 #endif
62190 
62191 #ifdef __LITTLE_ENDIAN__
62192 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62193  int8x16x3_t __s1 = __p1; \
62194  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
62195 })
62196 #else
62197 #define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62198  int8x16x3_t __s1 = __p1; \
62199  int8x16x3_t __rev1; \
62200  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62201  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62202  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62203  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
62204 })
62205 #endif
62206 
62207 #ifdef __LITTLE_ENDIAN__
62208 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62209  float64x2x3_t __s1 = __p1; \
62210  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
62211 })
62212 #else
62213 #define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62214  float64x2x3_t __s1 = __p1; \
62215  float64x2x3_t __rev1; \
62216  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62217  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62218  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62219  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
62220 })
62221 #endif
62222 
62223 #ifdef __LITTLE_ENDIAN__
62224 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62225  float32x4x3_t __s1 = __p1; \
62226  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
62227 })
62228 #else
62229 #define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62230  float32x4x3_t __s1 = __p1; \
62231  float32x4x3_t __rev1; \
62232  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62233  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62234  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62235  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
62236 })
62237 #endif
62238 
62239 #ifdef __LITTLE_ENDIAN__
62240 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62241  float16x8x3_t __s1 = __p1; \
62242  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
62243 })
62244 #else
62245 #define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62246  float16x8x3_t __s1 = __p1; \
62247  float16x8x3_t __rev1; \
62248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62251  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
62252 })
62253 #endif
62254 
62255 #ifdef __LITTLE_ENDIAN__
62256 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62257  int32x4x3_t __s1 = __p1; \
62258  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
62259 })
62260 #else
62261 #define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62262  int32x4x3_t __s1 = __p1; \
62263  int32x4x3_t __rev1; \
62264  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62265  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62266  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62267  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
62268 })
62269 #endif
62270 
62271 #ifdef __LITTLE_ENDIAN__
62272 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62273  int64x2x3_t __s1 = __p1; \
62274  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
62275 })
62276 #else
62277 #define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62278  int64x2x3_t __s1 = __p1; \
62279  int64x2x3_t __rev1; \
62280  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62281  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62282  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62283  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
62284 })
62285 #endif
62286 
62287 #ifdef __LITTLE_ENDIAN__
62288 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62289  int16x8x3_t __s1 = __p1; \
62290  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
62291 })
62292 #else
62293 #define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62294  int16x8x3_t __s1 = __p1; \
62295  int16x8x3_t __rev1; \
62296  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62297  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62298  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62299  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
62300 })
62301 #endif
62302 
62303 #ifdef __LITTLE_ENDIAN__
62304 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62305  uint8x8x3_t __s1 = __p1; \
62306  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
62307 })
62308 #else
62309 #define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62310  uint8x8x3_t __s1 = __p1; \
62311  uint8x8x3_t __rev1; \
62312  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62313  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62314  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62315  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
62316 })
62317 #endif
62318 
62319 #ifdef __LITTLE_ENDIAN__
62320 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62321  uint32x2x3_t __s1 = __p1; \
62322  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
62323 })
62324 #else
62325 #define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62326  uint32x2x3_t __s1 = __p1; \
62327  uint32x2x3_t __rev1; \
62328  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62329  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62330  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62331  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
62332 })
62333 #endif
62334 
62335 #ifdef __LITTLE_ENDIAN__
62336 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62337  uint64x1x3_t __s1 = __p1; \
62338  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62339 })
62340 #else
62341 #define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62342  uint64x1x3_t __s1 = __p1; \
62343  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62344 })
62345 #endif
62346 
62347 #ifdef __LITTLE_ENDIAN__
62348 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62349  uint16x4x3_t __s1 = __p1; \
62350  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
62351 })
62352 #else
62353 #define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62354  uint16x4x3_t __s1 = __p1; \
62355  uint16x4x3_t __rev1; \
62356  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62357  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62358  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62359  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
62360 })
62361 #endif
62362 
62363 #ifdef __LITTLE_ENDIAN__
62364 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62365  int8x8x3_t __s1 = __p1; \
62366  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
62367 })
62368 #else
62369 #define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62370  int8x8x3_t __s1 = __p1; \
62371  int8x8x3_t __rev1; \
62372  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62373  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62374  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62375  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
62376 })
62377 #endif
62378 
62379 #ifdef __LITTLE_ENDIAN__
62380 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62381  float64x1x3_t __s1 = __p1; \
62382  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62383 })
62384 #else
62385 #define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62386  float64x1x3_t __s1 = __p1; \
62387  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62388 })
62389 #endif
62390 
62391 #ifdef __LITTLE_ENDIAN__
62392 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62393  float32x2x3_t __s1 = __p1; \
62394  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
62395 })
62396 #else
62397 #define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62398  float32x2x3_t __s1 = __p1; \
62399  float32x2x3_t __rev1; \
62400  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62401  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62402  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62403  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
62404 })
62405 #endif
62406 
62407 #ifdef __LITTLE_ENDIAN__
62408 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62409  float16x4x3_t __s1 = __p1; \
62410  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
62411 })
62412 #else
62413 #define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62414  float16x4x3_t __s1 = __p1; \
62415  float16x4x3_t __rev1; \
62416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62419  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
62420 })
62421 #endif
62422 
62423 #ifdef __LITTLE_ENDIAN__
62424 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62425  int32x2x3_t __s1 = __p1; \
62426  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
62427 })
62428 #else
62429 #define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62430  int32x2x3_t __s1 = __p1; \
62431  int32x2x3_t __rev1; \
62432  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62433  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62434  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62435  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
62436 })
62437 #endif
62438 
62439 #ifdef __LITTLE_ENDIAN__
62440 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62441  int64x1x3_t __s1 = __p1; \
62442  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62443 })
62444 #else
62445 #define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62446  int64x1x3_t __s1 = __p1; \
62447  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62448 })
62449 #endif
62450 
62451 #ifdef __LITTLE_ENDIAN__
62452 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62453  int16x4x3_t __s1 = __p1; \
62454  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
62455 })
62456 #else
62457 #define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62458  int16x4x3_t __s1 = __p1; \
62459  int16x4x3_t __rev1; \
62460  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62461  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62462  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62463  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
62464 })
62465 #endif
62466 
62467 #ifdef __LITTLE_ENDIAN__
62468 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62469  poly8x8x4_t __s1 = __p1; \
62470  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
62471 })
62472 #else
62473 #define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62474  poly8x8x4_t __s1 = __p1; \
62475  poly8x8x4_t __rev1; \
62476  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62477  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62478  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62479  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62480  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
62481 })
62482 #endif
62483 
62484 #ifdef __LITTLE_ENDIAN__
62485 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62486  poly64x1x4_t __s1 = __p1; \
62487  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62488 })
62489 #else
62490 #define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62491  poly64x1x4_t __s1 = __p1; \
62492  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62493 })
62494 #endif
62495 
62496 #ifdef __LITTLE_ENDIAN__
62497 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62498  poly16x4x4_t __s1 = __p1; \
62499  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
62500 })
62501 #else
62502 #define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62503  poly16x4x4_t __s1 = __p1; \
62504  poly16x4x4_t __rev1; \
62505  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62506  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62507  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62508  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62509  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
62510 })
62511 #endif
62512 
62513 #ifdef __LITTLE_ENDIAN__
62514 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62515  poly8x16x4_t __s1 = __p1; \
62516  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
62517 })
62518 #else
62519 #define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62520  poly8x16x4_t __s1 = __p1; \
62521  poly8x16x4_t __rev1; \
62522  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62523  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62524  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62525  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62526  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
62527 })
62528 #endif
62529 
62530 #ifdef __LITTLE_ENDIAN__
62531 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62532  poly64x2x4_t __s1 = __p1; \
62533  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
62534 })
62535 #else
62536 #define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62537  poly64x2x4_t __s1 = __p1; \
62538  poly64x2x4_t __rev1; \
62539  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62540  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62541  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62542  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62543  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
62544 })
62545 #endif
62546 
62547 #ifdef __LITTLE_ENDIAN__
62548 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62549  poly16x8x4_t __s1 = __p1; \
62550  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
62551 })
62552 #else
62553 #define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62554  poly16x8x4_t __s1 = __p1; \
62555  poly16x8x4_t __rev1; \
62556  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62557  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62558  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62559  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62560  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
62561 })
62562 #endif
62563 
62564 #ifdef __LITTLE_ENDIAN__
62565 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62566  uint8x16x4_t __s1 = __p1; \
62567  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
62568 })
62569 #else
62570 #define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62571  uint8x16x4_t __s1 = __p1; \
62572  uint8x16x4_t __rev1; \
62573  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62574  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62575  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62576  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62577  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
62578 })
62579 #endif
62580 
62581 #ifdef __LITTLE_ENDIAN__
62582 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62583  uint32x4x4_t __s1 = __p1; \
62584  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
62585 })
62586 #else
62587 #define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62588  uint32x4x4_t __s1 = __p1; \
62589  uint32x4x4_t __rev1; \
62590  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62591  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62592  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62593  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62594  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
62595 })
62596 #endif
62597 
62598 #ifdef __LITTLE_ENDIAN__
62599 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62600  uint64x2x4_t __s1 = __p1; \
62601  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
62602 })
62603 #else
62604 #define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62605  uint64x2x4_t __s1 = __p1; \
62606  uint64x2x4_t __rev1; \
62607  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62608  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62609  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62610  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62611  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
62612 })
62613 #endif
62614 
62615 #ifdef __LITTLE_ENDIAN__
62616 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62617  uint16x8x4_t __s1 = __p1; \
62618  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
62619 })
62620 #else
62621 #define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62622  uint16x8x4_t __s1 = __p1; \
62623  uint16x8x4_t __rev1; \
62624  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62625  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62626  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62627  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62628  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
62629 })
62630 #endif
62631 
62632 #ifdef __LITTLE_ENDIAN__
62633 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62634  int8x16x4_t __s1 = __p1; \
62635  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
62636 })
62637 #else
62638 #define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62639  int8x16x4_t __s1 = __p1; \
62640  int8x16x4_t __rev1; \
62641  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62642  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62643  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62644  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62645  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
62646 })
62647 #endif
62648 
62649 #ifdef __LITTLE_ENDIAN__
62650 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62651  float64x2x4_t __s1 = __p1; \
62652  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
62653 })
62654 #else
62655 #define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62656  float64x2x4_t __s1 = __p1; \
62657  float64x2x4_t __rev1; \
62658  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62659  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62660  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62661  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62662  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
62663 })
62664 #endif
62665 
62666 #ifdef __LITTLE_ENDIAN__
62667 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62668  float32x4x4_t __s1 = __p1; \
62669  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
62670 })
62671 #else
62672 #define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62673  float32x4x4_t __s1 = __p1; \
62674  float32x4x4_t __rev1; \
62675  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62676  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62677  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62678  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62679  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
62680 })
62681 #endif
62682 
62683 #ifdef __LITTLE_ENDIAN__
62684 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62685  float16x8x4_t __s1 = __p1; \
62686  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
62687 })
62688 #else
62689 #define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62690  float16x8x4_t __s1 = __p1; \
62691  float16x8x4_t __rev1; \
62692  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62693  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62694  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62695  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62696  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
62697 })
62698 #endif
62699 
62700 #ifdef __LITTLE_ENDIAN__
62701 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62702  int32x4x4_t __s1 = __p1; \
62703  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
62704 })
62705 #else
62706 #define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62707  int32x4x4_t __s1 = __p1; \
62708  int32x4x4_t __rev1; \
62709  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62710  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62711  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62712  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62713  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
62714 })
62715 #endif
62716 
62717 #ifdef __LITTLE_ENDIAN__
62718 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62719  int64x2x4_t __s1 = __p1; \
62720  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
62721 })
62722 #else
62723 #define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62724  int64x2x4_t __s1 = __p1; \
62725  int64x2x4_t __rev1; \
62726  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62727  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62728  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62729  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62730  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
62731 })
62732 #endif
62733 
62734 #ifdef __LITTLE_ENDIAN__
62735 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62736  int16x8x4_t __s1 = __p1; \
62737  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
62738 })
62739 #else
62740 #define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62741  int16x8x4_t __s1 = __p1; \
62742  int16x8x4_t __rev1; \
62743  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62744  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62745  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62746  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62747  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
62748 })
62749 #endif
62750 
62751 #ifdef __LITTLE_ENDIAN__
62752 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62753  uint8x8x4_t __s1 = __p1; \
62754  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
62755 })
62756 #else
62757 #define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62758  uint8x8x4_t __s1 = __p1; \
62759  uint8x8x4_t __rev1; \
62760  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62761  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62762  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62763  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62764  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
62765 })
62766 #endif
62767 
62768 #ifdef __LITTLE_ENDIAN__
62769 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62770  uint32x2x4_t __s1 = __p1; \
62771  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
62772 })
62773 #else
62774 #define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62775  uint32x2x4_t __s1 = __p1; \
62776  uint32x2x4_t __rev1; \
62777  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62778  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62779  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62780  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62781  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
62782 })
62783 #endif
62784 
62785 #ifdef __LITTLE_ENDIAN__
62786 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62787  uint64x1x4_t __s1 = __p1; \
62788  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62789 })
62790 #else
62791 #define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62792  uint64x1x4_t __s1 = __p1; \
62793  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62794 })
62795 #endif
62796 
62797 #ifdef __LITTLE_ENDIAN__
62798 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62799  uint16x4x4_t __s1 = __p1; \
62800  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
62801 })
62802 #else
62803 #define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62804  uint16x4x4_t __s1 = __p1; \
62805  uint16x4x4_t __rev1; \
62806  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62807  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62808  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62809  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62810  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
62811 })
62812 #endif
62813 
62814 #ifdef __LITTLE_ENDIAN__
62815 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62816  int8x8x4_t __s1 = __p1; \
62817  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
62818 })
62819 #else
62820 #define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62821  int8x8x4_t __s1 = __p1; \
62822  int8x8x4_t __rev1; \
62823  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62824  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62825  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62826  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62827  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
62828 })
62829 #endif
62830 
62831 #ifdef __LITTLE_ENDIAN__
62832 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62833  float64x1x4_t __s1 = __p1; \
62834  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62835 })
62836 #else
62837 #define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62838  float64x1x4_t __s1 = __p1; \
62839  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62840 })
62841 #endif
62842 
62843 #ifdef __LITTLE_ENDIAN__
62844 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62845  float32x2x4_t __s1 = __p1; \
62846  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
62847 })
62848 #else
62849 #define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62850  float32x2x4_t __s1 = __p1; \
62851  float32x2x4_t __rev1; \
62852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62854  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62855  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62856  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
62857 })
62858 #endif
62859 
62860 #ifdef __LITTLE_ENDIAN__
62861 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62862  float16x4x4_t __s1 = __p1; \
62863  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
62864 })
62865 #else
62866 #define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62867  float16x4x4_t __s1 = __p1; \
62868  float16x4x4_t __rev1; \
62869  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62870  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62871  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62872  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62873  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
62874 })
62875 #endif
62876 
62877 #ifdef __LITTLE_ENDIAN__
62878 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62879  int32x2x4_t __s1 = __p1; \
62880  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
62881 })
62882 #else
62883 #define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62884  int32x2x4_t __s1 = __p1; \
62885  int32x2x4_t __rev1; \
62886  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62887  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62888  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62889  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62890  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
62891 })
62892 #endif
62893 
62894 #ifdef __LITTLE_ENDIAN__
62895 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62896  int64x1x4_t __s1 = __p1; \
62897  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62898 })
62899 #else
62900 #define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62901  int64x1x4_t __s1 = __p1; \
62902  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62903 })
62904 #endif
62905 
62906 #ifdef __LITTLE_ENDIAN__
62907 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62908  int16x4x4_t __s1 = __p1; \
62909  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
62910 })
62911 #else
62912 #define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62913  int16x4x4_t __s1 = __p1; \
62914  int16x4x4_t __rev1; \
62915  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62916  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62917  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62918  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62919  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
62920 })
62921 #endif
62922 
62923 #ifdef __LITTLE_ENDIAN__
62924 #define vst2_p64(__p0, __p1) __extension__ ({ \
62925  poly64x1x2_t __s1 = __p1; \
62926  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62927 })
62928 #else
62929 #define vst2_p64(__p0, __p1) __extension__ ({ \
62930  poly64x1x2_t __s1 = __p1; \
62931  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62932 })
62933 #endif
62934 
62935 #ifdef __LITTLE_ENDIAN__
62936 #define vst2q_p64(__p0, __p1) __extension__ ({ \
62937  poly64x2x2_t __s1 = __p1; \
62938  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
62939 })
62940 #else
62941 #define vst2q_p64(__p0, __p1) __extension__ ({ \
62942  poly64x2x2_t __s1 = __p1; \
62943  poly64x2x2_t __rev1; \
62944  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62945  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62946  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
62947 })
62948 #endif
62949 
62950 #ifdef __LITTLE_ENDIAN__
62951 #define vst2q_u64(__p0, __p1) __extension__ ({ \
62952  uint64x2x2_t __s1 = __p1; \
62953  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
62954 })
62955 #else
62956 #define vst2q_u64(__p0, __p1) __extension__ ({ \
62957  uint64x2x2_t __s1 = __p1; \
62958  uint64x2x2_t __rev1; \
62959  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62960  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62961  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
62962 })
62963 #endif
62964 
62965 #ifdef __LITTLE_ENDIAN__
62966 #define vst2q_f64(__p0, __p1) __extension__ ({ \
62967  float64x2x2_t __s1 = __p1; \
62968  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 42); \
62969 })
62970 #else
62971 #define vst2q_f64(__p0, __p1) __extension__ ({ \
62972  float64x2x2_t __s1 = __p1; \
62973  float64x2x2_t __rev1; \
62974  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62975  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62976  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
62977 })
62978 #endif
62979 
62980 #ifdef __LITTLE_ENDIAN__
62981 #define vst2q_s64(__p0, __p1) __extension__ ({ \
62982  int64x2x2_t __s1 = __p1; \
62983  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 35); \
62984 })
62985 #else
62986 #define vst2q_s64(__p0, __p1) __extension__ ({ \
62987  int64x2x2_t __s1 = __p1; \
62988  int64x2x2_t __rev1; \
62989  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62990  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62991  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
62992 })
62993 #endif
62994 
62995 #ifdef __LITTLE_ENDIAN__
62996 #define vst2_f64(__p0, __p1) __extension__ ({ \
62997  float64x1x2_t __s1 = __p1; \
62998  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62999 })
63000 #else
63001 #define vst2_f64(__p0, __p1) __extension__ ({ \
63002  float64x1x2_t __s1 = __p1; \
63003  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
63004 })
63005 #endif
63006 
63007 #ifdef __LITTLE_ENDIAN__
63008 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63009  poly64x1x2_t __s1 = __p1; \
63010  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63011 })
63012 #else
63013 #define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63014  poly64x1x2_t __s1 = __p1; \
63015  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63016 })
63017 #endif
63018 
63019 #ifdef __LITTLE_ENDIAN__
63020 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63021  poly8x16x2_t __s1 = __p1; \
63022  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
63023 })
63024 #else
63025 #define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63026  poly8x16x2_t __s1 = __p1; \
63027  poly8x16x2_t __rev1; \
63028  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63029  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63030  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
63031 })
63032 #endif
63033 
63034 #ifdef __LITTLE_ENDIAN__
63035 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63036  poly64x2x2_t __s1 = __p1; \
63037  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
63038 })
63039 #else
63040 #define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63041  poly64x2x2_t __s1 = __p1; \
63042  poly64x2x2_t __rev1; \
63043  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63044  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63045  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
63046 })
63047 #endif
63048 
63049 #ifdef __LITTLE_ENDIAN__
63050 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63051  uint8x16x2_t __s1 = __p1; \
63052  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
63053 })
63054 #else
63055 #define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63056  uint8x16x2_t __s1 = __p1; \
63057  uint8x16x2_t __rev1; \
63058  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63059  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63060  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
63061 })
63062 #endif
63063 
63064 #ifdef __LITTLE_ENDIAN__
63065 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63066  uint64x2x2_t __s1 = __p1; \
63067  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
63068 })
63069 #else
63070 #define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63071  uint64x2x2_t __s1 = __p1; \
63072  uint64x2x2_t __rev1; \
63073  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63074  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63075  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
63076 })
63077 #endif
63078 
63079 #ifdef __LITTLE_ENDIAN__
63080 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63081  int8x16x2_t __s1 = __p1; \
63082  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
63083 })
63084 #else
63085 #define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63086  int8x16x2_t __s1 = __p1; \
63087  int8x16x2_t __rev1; \
63088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63090  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
63091 })
63092 #endif
63093 
63094 #ifdef __LITTLE_ENDIAN__
63095 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63096  float64x2x2_t __s1 = __p1; \
63097  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 42); \
63098 })
63099 #else
63100 #define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63101  float64x2x2_t __s1 = __p1; \
63102  float64x2x2_t __rev1; \
63103  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63104  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63105  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
63106 })
63107 #endif
63108 
63109 #ifdef __LITTLE_ENDIAN__
63110 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63111  int64x2x2_t __s1 = __p1; \
63112  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 35); \
63113 })
63114 #else
63115 #define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63116  int64x2x2_t __s1 = __p1; \
63117  int64x2x2_t __rev1; \
63118  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63119  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63120  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
63121 })
63122 #endif
63123 
63124 #ifdef __LITTLE_ENDIAN__
63125 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63126  uint64x1x2_t __s1 = __p1; \
63127  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63128 })
63129 #else
63130 #define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63131  uint64x1x2_t __s1 = __p1; \
63132  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63133 })
63134 #endif
63135 
63136 #ifdef __LITTLE_ENDIAN__
63137 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63138  float64x1x2_t __s1 = __p1; \
63139  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63140 })
63141 #else
63142 #define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63143  float64x1x2_t __s1 = __p1; \
63144  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63145 })
63146 #endif
63147 
63148 #ifdef __LITTLE_ENDIAN__
63149 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63150  int64x1x2_t __s1 = __p1; \
63151  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63152 })
63153 #else
63154 #define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63155  int64x1x2_t __s1 = __p1; \
63156  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63157 })
63158 #endif
63159 
63160 #ifdef __LITTLE_ENDIAN__
63161 #define vst3_p64(__p0, __p1) __extension__ ({ \
63162  poly64x1x3_t __s1 = __p1; \
63163  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63164 })
63165 #else
63166 #define vst3_p64(__p0, __p1) __extension__ ({ \
63167  poly64x1x3_t __s1 = __p1; \
63168  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63169 })
63170 #endif
63171 
63172 #ifdef __LITTLE_ENDIAN__
63173 #define vst3q_p64(__p0, __p1) __extension__ ({ \
63174  poly64x2x3_t __s1 = __p1; \
63175  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
63176 })
63177 #else
63178 #define vst3q_p64(__p0, __p1) __extension__ ({ \
63179  poly64x2x3_t __s1 = __p1; \
63180  poly64x2x3_t __rev1; \
63181  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63182  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63183  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63184  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
63185 })
63186 #endif
63187 
63188 #ifdef __LITTLE_ENDIAN__
63189 #define vst3q_u64(__p0, __p1) __extension__ ({ \
63190  uint64x2x3_t __s1 = __p1; \
63191  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
63192 })
63193 #else
63194 #define vst3q_u64(__p0, __p1) __extension__ ({ \
63195  uint64x2x3_t __s1 = __p1; \
63196  uint64x2x3_t __rev1; \
63197  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63198  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63199  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63200  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
63201 })
63202 #endif
63203 
63204 #ifdef __LITTLE_ENDIAN__
63205 #define vst3q_f64(__p0, __p1) __extension__ ({ \
63206  float64x2x3_t __s1 = __p1; \
63207  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
63208 })
63209 #else
63210 #define vst3q_f64(__p0, __p1) __extension__ ({ \
63211  float64x2x3_t __s1 = __p1; \
63212  float64x2x3_t __rev1; \
63213  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63214  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63215  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63216  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
63217 })
63218 #endif
63219 
63220 #ifdef __LITTLE_ENDIAN__
63221 #define vst3q_s64(__p0, __p1) __extension__ ({ \
63222  int64x2x3_t __s1 = __p1; \
63223  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
63224 })
63225 #else
63226 #define vst3q_s64(__p0, __p1) __extension__ ({ \
63227  int64x2x3_t __s1 = __p1; \
63228  int64x2x3_t __rev1; \
63229  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63230  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63231  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63232  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
63233 })
63234 #endif
63235 
63236 #ifdef __LITTLE_ENDIAN__
63237 #define vst3_f64(__p0, __p1) __extension__ ({ \
63238  float64x1x3_t __s1 = __p1; \
63239  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63240 })
63241 #else
63242 #define vst3_f64(__p0, __p1) __extension__ ({ \
63243  float64x1x3_t __s1 = __p1; \
63244  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63245 })
63246 #endif
63247 
63248 #ifdef __LITTLE_ENDIAN__
63249 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63250  poly64x1x3_t __s1 = __p1; \
63251  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63252 })
63253 #else
63254 #define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63255  poly64x1x3_t __s1 = __p1; \
63256  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63257 })
63258 #endif
63259 
63260 #ifdef __LITTLE_ENDIAN__
63261 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63262  poly8x16x3_t __s1 = __p1; \
63263  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
63264 })
63265 #else
63266 #define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63267  poly8x16x3_t __s1 = __p1; \
63268  poly8x16x3_t __rev1; \
63269  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63270  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63271  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63272  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
63273 })
63274 #endif
63275 
63276 #ifdef __LITTLE_ENDIAN__
63277 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63278  poly64x2x3_t __s1 = __p1; \
63279  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
63280 })
63281 #else
63282 #define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63283  poly64x2x3_t __s1 = __p1; \
63284  poly64x2x3_t __rev1; \
63285  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63286  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63287  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63288  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
63289 })
63290 #endif
63291 
63292 #ifdef __LITTLE_ENDIAN__
63293 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63294  uint8x16x3_t __s1 = __p1; \
63295  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
63296 })
63297 #else
63298 #define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63299  uint8x16x3_t __s1 = __p1; \
63300  uint8x16x3_t __rev1; \
63301  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63302  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63303  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63304  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
63305 })
63306 #endif
63307 
63308 #ifdef __LITTLE_ENDIAN__
63309 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63310  uint64x2x3_t __s1 = __p1; \
63311  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
63312 })
63313 #else
63314 #define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63315  uint64x2x3_t __s1 = __p1; \
63316  uint64x2x3_t __rev1; \
63317  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63318  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63319  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63320  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
63321 })
63322 #endif
63323 
63324 #ifdef __LITTLE_ENDIAN__
63325 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63326  int8x16x3_t __s1 = __p1; \
63327  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
63328 })
63329 #else
63330 #define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63331  int8x16x3_t __s1 = __p1; \
63332  int8x16x3_t __rev1; \
63333  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63334  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63335  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63336  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
63337 })
63338 #endif
63339 
63340 #ifdef __LITTLE_ENDIAN__
63341 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63342  float64x2x3_t __s1 = __p1; \
63343  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
63344 })
63345 #else
63346 #define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63347  float64x2x3_t __s1 = __p1; \
63348  float64x2x3_t __rev1; \
63349  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63350  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63351  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63352  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
63353 })
63354 #endif
63355 
63356 #ifdef __LITTLE_ENDIAN__
63357 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63358  int64x2x3_t __s1 = __p1; \
63359  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
63360 })
63361 #else
63362 #define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63363  int64x2x3_t __s1 = __p1; \
63364  int64x2x3_t __rev1; \
63365  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63366  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63367  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63368  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
63369 })
63370 #endif
63371 
63372 #ifdef __LITTLE_ENDIAN__
63373 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63374  uint64x1x3_t __s1 = __p1; \
63375  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63376 })
63377 #else
63378 #define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63379  uint64x1x3_t __s1 = __p1; \
63380  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63381 })
63382 #endif
63383 
63384 #ifdef __LITTLE_ENDIAN__
63385 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63386  float64x1x3_t __s1 = __p1; \
63387  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63388 })
63389 #else
63390 #define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63391  float64x1x3_t __s1 = __p1; \
63392  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63393 })
63394 #endif
63395 
63396 #ifdef __LITTLE_ENDIAN__
63397 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63398  int64x1x3_t __s1 = __p1; \
63399  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63400 })
63401 #else
63402 #define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63403  int64x1x3_t __s1 = __p1; \
63404  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63405 })
63406 #endif
63407 
63408 #ifdef __LITTLE_ENDIAN__
63409 #define vst4_p64(__p0, __p1) __extension__ ({ \
63410  poly64x1x4_t __s1 = __p1; \
63411  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63412 })
63413 #else
63414 #define vst4_p64(__p0, __p1) __extension__ ({ \
63415  poly64x1x4_t __s1 = __p1; \
63416  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63417 })
63418 #endif
63419 
63420 #ifdef __LITTLE_ENDIAN__
63421 #define vst4q_p64(__p0, __p1) __extension__ ({ \
63422  poly64x2x4_t __s1 = __p1; \
63423  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
63424 })
63425 #else
63426 #define vst4q_p64(__p0, __p1) __extension__ ({ \
63427  poly64x2x4_t __s1 = __p1; \
63428  poly64x2x4_t __rev1; \
63429  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63430  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63431  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63432  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63433  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
63434 })
63435 #endif
63436 
63437 #ifdef __LITTLE_ENDIAN__
63438 #define vst4q_u64(__p0, __p1) __extension__ ({ \
63439  uint64x2x4_t __s1 = __p1; \
63440  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
63441 })
63442 #else
63443 #define vst4q_u64(__p0, __p1) __extension__ ({ \
63444  uint64x2x4_t __s1 = __p1; \
63445  uint64x2x4_t __rev1; \
63446  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63447  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63448  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63449  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63450  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
63451 })
63452 #endif
63453 
63454 #ifdef __LITTLE_ENDIAN__
63455 #define vst4q_f64(__p0, __p1) __extension__ ({ \
63456  float64x2x4_t __s1 = __p1; \
63457  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
63458 })
63459 #else
63460 #define vst4q_f64(__p0, __p1) __extension__ ({ \
63461  float64x2x4_t __s1 = __p1; \
63462  float64x2x4_t __rev1; \
63463  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63464  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63465  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63466  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63467  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
63468 })
63469 #endif
63470 
63471 #ifdef __LITTLE_ENDIAN__
63472 #define vst4q_s64(__p0, __p1) __extension__ ({ \
63473  int64x2x4_t __s1 = __p1; \
63474  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
63475 })
63476 #else
63477 #define vst4q_s64(__p0, __p1) __extension__ ({ \
63478  int64x2x4_t __s1 = __p1; \
63479  int64x2x4_t __rev1; \
63480  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63481  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63482  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63483  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63484  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
63485 })
63486 #endif
63487 
63488 #ifdef __LITTLE_ENDIAN__
63489 #define vst4_f64(__p0, __p1) __extension__ ({ \
63490  float64x1x4_t __s1 = __p1; \
63491  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63492 })
63493 #else
63494 #define vst4_f64(__p0, __p1) __extension__ ({ \
63495  float64x1x4_t __s1 = __p1; \
63496  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63497 })
63498 #endif
63499 
63500 #ifdef __LITTLE_ENDIAN__
63501 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63502  poly64x1x4_t __s1 = __p1; \
63503  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63504 })
63505 #else
63506 #define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63507  poly64x1x4_t __s1 = __p1; \
63508  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63509 })
63510 #endif
63511 
63512 #ifdef __LITTLE_ENDIAN__
63513 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63514  poly8x16x4_t __s1 = __p1; \
63515  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
63516 })
63517 #else
63518 #define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63519  poly8x16x4_t __s1 = __p1; \
63520  poly8x16x4_t __rev1; \
63521  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63522  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63523  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63524  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63525  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
63526 })
63527 #endif
63528 
63529 #ifdef __LITTLE_ENDIAN__
63530 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63531  poly64x2x4_t __s1 = __p1; \
63532  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
63533 })
63534 #else
63535 #define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63536  poly64x2x4_t __s1 = __p1; \
63537  poly64x2x4_t __rev1; \
63538  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63539  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63540  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63541  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63542  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
63543 })
63544 #endif
63545 
63546 #ifdef __LITTLE_ENDIAN__
63547 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63548  uint8x16x4_t __s1 = __p1; \
63549  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
63550 })
63551 #else
63552 #define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63553  uint8x16x4_t __s1 = __p1; \
63554  uint8x16x4_t __rev1; \
63555  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63556  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63557  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63558  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63559  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
63560 })
63561 #endif
63562 
63563 #ifdef __LITTLE_ENDIAN__
63564 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63565  uint64x2x4_t __s1 = __p1; \
63566  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
63567 })
63568 #else
63569 #define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63570  uint64x2x4_t __s1 = __p1; \
63571  uint64x2x4_t __rev1; \
63572  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63573  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63574  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63575  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63576  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
63577 })
63578 #endif
63579 
63580 #ifdef __LITTLE_ENDIAN__
63581 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63582  int8x16x4_t __s1 = __p1; \
63583  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
63584 })
63585 #else
63586 #define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63587  int8x16x4_t __s1 = __p1; \
63588  int8x16x4_t __rev1; \
63589  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63590  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63591  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63592  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63593  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
63594 })
63595 #endif
63596 
63597 #ifdef __LITTLE_ENDIAN__
63598 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63599  float64x2x4_t __s1 = __p1; \
63600  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
63601 })
63602 #else
63603 #define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63604  float64x2x4_t __s1 = __p1; \
63605  float64x2x4_t __rev1; \
63606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63608  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63609  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63610  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
63611 })
63612 #endif
63613 
63614 #ifdef __LITTLE_ENDIAN__
63615 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63616  int64x2x4_t __s1 = __p1; \
63617  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
63618 })
63619 #else
63620 #define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63621  int64x2x4_t __s1 = __p1; \
63622  int64x2x4_t __rev1; \
63623  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63624  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63625  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63626  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63627  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
63628 })
63629 #endif
63630 
63631 #ifdef __LITTLE_ENDIAN__
63632 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63633  uint64x1x4_t __s1 = __p1; \
63634  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63635 })
63636 #else
63637 #define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63638  uint64x1x4_t __s1 = __p1; \
63639  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63640 })
63641 #endif
63642 
63643 #ifdef __LITTLE_ENDIAN__
63644 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63645  float64x1x4_t __s1 = __p1; \
63646  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63647 })
63648 #else
63649 #define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63650  float64x1x4_t __s1 = __p1; \
63651  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63652 })
63653 #endif
63654 
63655 #ifdef __LITTLE_ENDIAN__
63656 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63657  int64x1x4_t __s1 = __p1; \
63658  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63659 })
63660 #else
63661 #define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63662  int64x1x4_t __s1 = __p1; \
63663  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63664 })
63665 #endif
63666 
63667 #ifdef __LITTLE_ENDIAN__
63668 #define vstrq_p128(__p0, __p1) __extension__ ({ \
63669  poly128_t __s1 = __p1; \
63670  __builtin_neon_vstrq_p128(__p0, __s1); \
63671 })
63672 #else
63673 #define vstrq_p128(__p0, __p1) __extension__ ({ \
63674  poly128_t __s1 = __p1; \
63675  __builtin_neon_vstrq_p128(__p0, __s1); \
63676 })
63677 #endif
63678 
63679 #ifdef __LITTLE_ENDIAN__
63680 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63681  uint64_t __ret;
63682  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63683  return __ret;
63684 }
63685 #else
63686 __ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63687  uint64_t __ret;
63688  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63689  return __ret;
63690 }
63691 #endif
63692 
63693 #ifdef __LITTLE_ENDIAN__
63694 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63695  int64_t __ret;
63696  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63697  return __ret;
63698 }
63699 #else
63700 __ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63701  int64_t __ret;
63702  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63703  return __ret;
63704 }
63705 #endif
63706 
63707 #ifdef __LITTLE_ENDIAN__
63708 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63709  float64x2_t __ret;
63710  __ret = __p0 - __p1;
63711  return __ret;
63712 }
63713 #else
63714 __ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63715  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63716  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63717  float64x2_t __ret;
63718  __ret = __rev0 - __rev1;
63719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63720  return __ret;
63721 }
63722 #endif
63723 
63724 #ifdef __LITTLE_ENDIAN__
63725 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63726  float64x1_t __ret;
63727  __ret = __p0 - __p1;
63728  return __ret;
63729 }
63730 #else
63731 __ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63732  float64x1_t __ret;
63733  __ret = __p0 - __p1;
63734  return __ret;
63735 }
63736 #endif
63737 
63738 #ifdef __LITTLE_ENDIAN__
63739 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63740  uint16x8_t __ret;
63741  __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
63742  return __ret;
63743 }
63744 #else
63745 __ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63746  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63747  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63748  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63749  uint16x8_t __ret;
63750  __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
63751  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63752  return __ret;
63753 }
63754 #endif
63755 
63756 #ifdef __LITTLE_ENDIAN__
63757 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63758  uint32x4_t __ret;
63759  __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
63760  return __ret;
63761 }
63762 #else
63763 __ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63764  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63765  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63766  uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63767  uint32x4_t __ret;
63768  __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
63769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63770  return __ret;
63771 }
63772 #endif
63773 
63774 #ifdef __LITTLE_ENDIAN__
63775 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63776  uint8x16_t __ret;
63777  __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
63778  return __ret;
63779 }
63780 #else
63781 __ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63782  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63783  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63784  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63785  uint8x16_t __ret;
63786  __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
63787  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63788  return __ret;
63789 }
63790 #endif
63791 
63792 #ifdef __LITTLE_ENDIAN__
63793 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63794  int16x8_t __ret;
63795  __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
63796  return __ret;
63797 }
63798 #else
63799 __ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63800  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63801  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63802  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63803  int16x8_t __ret;
63804  __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
63805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63806  return __ret;
63807 }
63808 #endif
63809 
63810 #ifdef __LITTLE_ENDIAN__
63811 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63812  int32x4_t __ret;
63813  __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
63814  return __ret;
63815 }
63816 #else
63817 __ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63818  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63819  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63820  int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63821  int32x4_t __ret;
63822  __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
63823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63824  return __ret;
63825 }
63826 #endif
63827 
63828 #ifdef __LITTLE_ENDIAN__
63829 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63830  int8x16_t __ret;
63831  __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
63832  return __ret;
63833 }
63834 #else
63835 __ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63836  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63837  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63838  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63839  int8x16_t __ret;
63840  __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
63841  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63842  return __ret;
63843 }
63844 #endif
63845 
63846 #ifdef __LITTLE_ENDIAN__
63847 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63848  uint16x8_t __ret;
63849  __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
63850  return __ret;
63851 }
63852 #else
63853 __ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63854  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63855  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63856  uint16x8_t __ret;
63857  __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
63858  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63859  return __ret;
63860 }
63861 #endif
63862 
63863 #ifdef __LITTLE_ENDIAN__
63864 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63865  uint64x2_t __ret;
63866  __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
63867  return __ret;
63868 }
63869 #else
63870 __ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63871  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63872  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63873  uint64x2_t __ret;
63874  __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
63875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63876  return __ret;
63877 }
63878 #endif
63879 
63880 #ifdef __LITTLE_ENDIAN__
63881 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63882  uint32x4_t __ret;
63883  __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
63884  return __ret;
63885 }
63886 #else
63887 __ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63888  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63889  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63890  uint32x4_t __ret;
63891  __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
63892  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63893  return __ret;
63894 }
63895 #endif
63896 
63897 #ifdef __LITTLE_ENDIAN__
63898 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63899  int16x8_t __ret;
63900  __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
63901  return __ret;
63902 }
63903 #else
63904 __ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63905  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63906  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63907  int16x8_t __ret;
63908  __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
63909  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63910  return __ret;
63911 }
63912 #endif
63913 
63914 #ifdef __LITTLE_ENDIAN__
63915 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63916  int64x2_t __ret;
63917  __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
63918  return __ret;
63919 }
63920 #else
63921 __ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63922  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63923  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63924  int64x2_t __ret;
63925  __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
63926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63927  return __ret;
63928 }
63929 #endif
63930 
63931 #ifdef __LITTLE_ENDIAN__
63932 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63933  int32x4_t __ret;
63934  __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
63935  return __ret;
63936 }
63937 #else
63938 __ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63939  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63940  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63941  int32x4_t __ret;
63942  __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
63943  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63944  return __ret;
63945 }
63946 #endif
63947 
63948 #ifdef __LITTLE_ENDIAN__
63949 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63950  uint16x8_t __ret;
63951  __ret = __p0 - vmovl_high_u8(__p1);
63952  return __ret;
63953 }
63954 #else
63955 __ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63956  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63957  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63958  uint16x8_t __ret;
63959  __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
63960  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63961  return __ret;
63962 }
63963 #endif
63964 
63965 #ifdef __LITTLE_ENDIAN__
63966 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63967  uint64x2_t __ret;
63968  __ret = __p0 - vmovl_high_u32(__p1);
63969  return __ret;
63970 }
63971 #else
63972 __ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63973  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63974  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63975  uint64x2_t __ret;
63976  __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
63977  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63978  return __ret;
63979 }
63980 #endif
63981 
63982 #ifdef __LITTLE_ENDIAN__
63983 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63984  uint32x4_t __ret;
63985  __ret = __p0 - vmovl_high_u16(__p1);
63986  return __ret;
63987 }
63988 #else
63989 __ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63990  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63991  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63992  uint32x4_t __ret;
63993  __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
63994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63995  return __ret;
63996 }
63997 #endif
63998 
63999 #ifdef __LITTLE_ENDIAN__
64000 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
64001  int16x8_t __ret;
64002  __ret = __p0 - vmovl_high_s8(__p1);
64003  return __ret;
64004 }
64005 #else
64006 __ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
64007  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64008  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64009  int16x8_t __ret;
64010  __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
64011  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64012  return __ret;
64013 }
64014 #endif
64015 
64016 #ifdef __LITTLE_ENDIAN__
64017 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64018  int64x2_t __ret;
64019  __ret = __p0 - vmovl_high_s32(__p1);
64020  return __ret;
64021 }
64022 #else
64023 __ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64024  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64025  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64026  int64x2_t __ret;
64027  __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
64028  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64029  return __ret;
64030 }
64031 #endif
64032 
64033 #ifdef __LITTLE_ENDIAN__
64034 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64035  int32x4_t __ret;
64036  __ret = __p0 - vmovl_high_s16(__p1);
64037  return __ret;
64038 }
64039 #else
64040 __ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64041  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64042  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64043  int32x4_t __ret;
64044  __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
64045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64046  return __ret;
64047 }
64048 #endif
64049 
64050 #ifdef __LITTLE_ENDIAN__
64051 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64052  poly8x8_t __ret;
64053  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64054  return __ret;
64055 }
64056 #else
64057 __ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64058  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64059  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64060  poly8x8_t __ret;
64061  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64062  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64063  return __ret;
64064 }
64065 #endif
64066 
64067 #ifdef __LITTLE_ENDIAN__
64068 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64069  poly16x4_t __ret;
64070  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64071  return __ret;
64072 }
64073 #else
64074 __ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64075  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64076  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64077  poly16x4_t __ret;
64078  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64079  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64080  return __ret;
64081 }
64082 #endif
64083 
64084 #ifdef __LITTLE_ENDIAN__
64085 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64086  poly8x16_t __ret;
64087  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64088  return __ret;
64089 }
64090 #else
64091 __ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64092  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64093  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64094  poly8x16_t __ret;
64095  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64096  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64097  return __ret;
64098 }
64099 #endif
64100 
64101 #ifdef __LITTLE_ENDIAN__
64102 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64103  poly64x2_t __ret;
64104  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64105  return __ret;
64106 }
64107 #else
64108 __ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64109  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64110  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64111  poly64x2_t __ret;
64112  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64114  return __ret;
64115 }
64116 #endif
64117 
64118 #ifdef __LITTLE_ENDIAN__
64119 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64120  poly16x8_t __ret;
64121  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64122  return __ret;
64123 }
64124 #else
64125 __ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64126  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64127  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64128  poly16x8_t __ret;
64129  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64131  return __ret;
64132 }
64133 #endif
64134 
64135 #ifdef __LITTLE_ENDIAN__
64136 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64137  uint8x16_t __ret;
64138  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64139  return __ret;
64140 }
64141 #else
64142 __ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64143  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64144  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64145  uint8x16_t __ret;
64146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64147  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64148  return __ret;
64149 }
64150 #endif
64151 
64152 #ifdef __LITTLE_ENDIAN__
64153 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64154  uint32x4_t __ret;
64155  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64156  return __ret;
64157 }
64158 #else
64159 __ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64160  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64161  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64162  uint32x4_t __ret;
64163  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64164  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64165  return __ret;
64166 }
64167 #endif
64168 
64169 #ifdef __LITTLE_ENDIAN__
64170 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64171  uint64x2_t __ret;
64172  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64173  return __ret;
64174 }
64175 #else
64176 __ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64177  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64178  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64179  uint64x2_t __ret;
64180  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64181  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64182  return __ret;
64183 }
64184 #endif
64185 
64186 #ifdef __LITTLE_ENDIAN__
64187 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64188  uint16x8_t __ret;
64189  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64190  return __ret;
64191 }
64192 #else
64193 __ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64194  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64195  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64196  uint16x8_t __ret;
64197  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64199  return __ret;
64200 }
64201 #endif
64202 
64203 #ifdef __LITTLE_ENDIAN__
64204 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64205  int8x16_t __ret;
64206  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64207  return __ret;
64208 }
64209 #else
64210 __ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64211  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64212  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64213  int8x16_t __ret;
64214  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64216  return __ret;
64217 }
64218 #endif
64219 
64220 #ifdef __LITTLE_ENDIAN__
64221 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64222  float64x2_t __ret;
64223  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64224  return __ret;
64225 }
64226 #else
64227 __ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64228  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64229  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64230  float64x2_t __ret;
64231  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64232  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64233  return __ret;
64234 }
64235 #endif
64236 
64237 #ifdef __LITTLE_ENDIAN__
64238 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64239  float32x4_t __ret;
64240  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64241  return __ret;
64242 }
64243 #else
64244 __ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64245  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64246  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64247  float32x4_t __ret;
64248  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64249  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64250  return __ret;
64251 }
64252 #endif
64253 
64254 #ifdef __LITTLE_ENDIAN__
64255 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64256  int32x4_t __ret;
64257  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64258  return __ret;
64259 }
64260 #else
64261 __ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64262  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64263  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64264  int32x4_t __ret;
64265  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64266  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64267  return __ret;
64268 }
64269 #endif
64270 
64271 #ifdef __LITTLE_ENDIAN__
64272 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64273  int64x2_t __ret;
64274  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64275  return __ret;
64276 }
64277 #else
64278 __ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64279  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64280  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64281  int64x2_t __ret;
64282  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64284  return __ret;
64285 }
64286 #endif
64287 
64288 #ifdef __LITTLE_ENDIAN__
64289 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64290  int16x8_t __ret;
64291  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64292  return __ret;
64293 }
64294 #else
64295 __ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64296  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64297  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64298  int16x8_t __ret;
64299  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64300  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64301  return __ret;
64302 }
64303 #endif
64304 
64305 #ifdef __LITTLE_ENDIAN__
64306 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64307  uint8x8_t __ret;
64308  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64309  return __ret;
64310 }
64311 #else
64312 __ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64313  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64314  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64315  uint8x8_t __ret;
64316  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64317  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64318  return __ret;
64319 }
64320 #endif
64321 
64322 #ifdef __LITTLE_ENDIAN__
64323 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64324  uint32x2_t __ret;
64325  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64326  return __ret;
64327 }
64328 #else
64329 __ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64330  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64331  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64332  uint32x2_t __ret;
64333  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64334  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64335  return __ret;
64336 }
64337 #endif
64338 
64339 #ifdef __LITTLE_ENDIAN__
64340 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64341  uint16x4_t __ret;
64342  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64343  return __ret;
64344 }
64345 #else
64346 __ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64347  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64348  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64349  uint16x4_t __ret;
64350  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64351  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64352  return __ret;
64353 }
64354 #endif
64355 
64356 #ifdef __LITTLE_ENDIAN__
64357 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64358  int8x8_t __ret;
64359  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64360  return __ret;
64361 }
64362 #else
64363 __ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64364  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64365  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64366  int8x8_t __ret;
64367  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64368  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64369  return __ret;
64370 }
64371 #endif
64372 
64373 #ifdef __LITTLE_ENDIAN__
64374 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64375  float32x2_t __ret;
64376  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64377  return __ret;
64378 }
64379 #else
64380 __ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64381  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64382  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64383  float32x2_t __ret;
64384  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64386  return __ret;
64387 }
64388 #endif
64389 
64390 #ifdef __LITTLE_ENDIAN__
64391 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64392  int32x2_t __ret;
64393  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64394  return __ret;
64395 }
64396 #else
64397 __ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64398  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64399  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64400  int32x2_t __ret;
64401  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64402  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64403  return __ret;
64404 }
64405 #endif
64406 
64407 #ifdef __LITTLE_ENDIAN__
64408 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64409  int16x4_t __ret;
64410  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64411  return __ret;
64412 }
64413 #else
64414 __ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64415  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64416  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64417  int16x4_t __ret;
64418  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64419  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64420  return __ret;
64421 }
64422 #endif
64423 
64424 #ifdef __LITTLE_ENDIAN__
64425 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64426  poly8x8_t __ret;
64427  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64428  return __ret;
64429 }
64430 #else
64431 __ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64432  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64433  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64434  poly8x8_t __ret;
64435  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64436  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64437  return __ret;
64438 }
64439 #endif
64440 
64441 #ifdef __LITTLE_ENDIAN__
64442 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64443  poly16x4_t __ret;
64444  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64445  return __ret;
64446 }
64447 #else
64448 __ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64449  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64450  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64451  poly16x4_t __ret;
64452  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64453  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64454  return __ret;
64455 }
64456 #endif
64457 
64458 #ifdef __LITTLE_ENDIAN__
64459 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64460  poly8x16_t __ret;
64461  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64462  return __ret;
64463 }
64464 #else
64465 __ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64466  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64467  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64468  poly8x16_t __ret;
64469  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64470  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64471  return __ret;
64472 }
64473 #endif
64474 
64475 #ifdef __LITTLE_ENDIAN__
64476 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64477  poly64x2_t __ret;
64478  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64479  return __ret;
64480 }
64481 #else
64482 __ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64483  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64484  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64485  poly64x2_t __ret;
64486  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64487  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64488  return __ret;
64489 }
64490 #endif
64491 
64492 #ifdef __LITTLE_ENDIAN__
64493 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64494  poly16x8_t __ret;
64495  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64496  return __ret;
64497 }
64498 #else
64499 __ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64500  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64501  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64502  poly16x8_t __ret;
64503  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64504  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64505  return __ret;
64506 }
64507 #endif
64508 
64509 #ifdef __LITTLE_ENDIAN__
64510 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64511  uint8x16_t __ret;
64512  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64513  return __ret;
64514 }
64515 #else
64516 __ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64517  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64518  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64519  uint8x16_t __ret;
64520  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64521  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64522  return __ret;
64523 }
64524 #endif
64525 
64526 #ifdef __LITTLE_ENDIAN__
64527 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64528  uint32x4_t __ret;
64529  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64530  return __ret;
64531 }
64532 #else
64533 __ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64534  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64535  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64536  uint32x4_t __ret;
64537  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64538  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64539  return __ret;
64540 }
64541 #endif
64542 
64543 #ifdef __LITTLE_ENDIAN__
64544 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64545  uint64x2_t __ret;
64546  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64547  return __ret;
64548 }
64549 #else
64550 __ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64551  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64552  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64553  uint64x2_t __ret;
64554  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64555  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64556  return __ret;
64557 }
64558 #endif
64559 
64560 #ifdef __LITTLE_ENDIAN__
64561 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64562  uint16x8_t __ret;
64563  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64564  return __ret;
64565 }
64566 #else
64567 __ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64568  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64569  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64570  uint16x8_t __ret;
64571  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64572  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64573  return __ret;
64574 }
64575 #endif
64576 
64577 #ifdef __LITTLE_ENDIAN__
64578 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64579  int8x16_t __ret;
64580  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64581  return __ret;
64582 }
64583 #else
64584 __ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64585  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64586  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64587  int8x16_t __ret;
64588  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64589  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64590  return __ret;
64591 }
64592 #endif
64593 
64594 #ifdef __LITTLE_ENDIAN__
64595 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64596  float64x2_t __ret;
64597  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64598  return __ret;
64599 }
64600 #else
64601 __ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64602  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64603  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64604  float64x2_t __ret;
64605  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64606  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64607  return __ret;
64608 }
64609 #endif
64610 
64611 #ifdef __LITTLE_ENDIAN__
64612 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64613  float32x4_t __ret;
64614  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64615  return __ret;
64616 }
64617 #else
64618 __ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64619  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64620  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64621  float32x4_t __ret;
64622  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64623  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64624  return __ret;
64625 }
64626 #endif
64627 
64628 #ifdef __LITTLE_ENDIAN__
64629 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64630  int32x4_t __ret;
64631  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64632  return __ret;
64633 }
64634 #else
64635 __ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64636  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64637  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64638  int32x4_t __ret;
64639  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64641  return __ret;
64642 }
64643 #endif
64644 
64645 #ifdef __LITTLE_ENDIAN__
64646 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64647  int64x2_t __ret;
64648  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64649  return __ret;
64650 }
64651 #else
64652 __ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64653  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64654  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64655  int64x2_t __ret;
64656  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64657  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64658  return __ret;
64659 }
64660 #endif
64661 
64662 #ifdef __LITTLE_ENDIAN__
64663 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64664  int16x8_t __ret;
64665  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64666  return __ret;
64667 }
64668 #else
64669 __ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64670  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64671  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64672  int16x8_t __ret;
64673  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64674  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64675  return __ret;
64676 }
64677 #endif
64678 
64679 #ifdef __LITTLE_ENDIAN__
64680 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64681  uint8x8_t __ret;
64682  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64683  return __ret;
64684 }
64685 #else
64686 __ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64687  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64688  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64689  uint8x8_t __ret;
64690  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64691  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64692  return __ret;
64693 }
64694 #endif
64695 
64696 #ifdef __LITTLE_ENDIAN__
64697 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64698  uint32x2_t __ret;
64699  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64700  return __ret;
64701 }
64702 #else
64703 __ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64704  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64705  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64706  uint32x2_t __ret;
64707  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64709  return __ret;
64710 }
64711 #endif
64712 
64713 #ifdef __LITTLE_ENDIAN__
64714 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64715  uint16x4_t __ret;
64716  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64717  return __ret;
64718 }
64719 #else
64720 __ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64721  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64722  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64723  uint16x4_t __ret;
64724  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64726  return __ret;
64727 }
64728 #endif
64729 
64730 #ifdef __LITTLE_ENDIAN__
64731 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64732  int8x8_t __ret;
64733  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64734  return __ret;
64735 }
64736 #else
64737 __ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64738  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64739  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64740  int8x8_t __ret;
64741  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64742  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64743  return __ret;
64744 }
64745 #endif
64746 
64747 #ifdef __LITTLE_ENDIAN__
64748 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64749  float32x2_t __ret;
64750  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64751  return __ret;
64752 }
64753 #else
64754 __ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64755  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64756  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64757  float32x2_t __ret;
64758  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64759  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64760  return __ret;
64761 }
64762 #endif
64763 
64764 #ifdef __LITTLE_ENDIAN__
64765 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64766  int32x2_t __ret;
64767  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64768  return __ret;
64769 }
64770 #else
64771 __ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64772  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64773  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64774  int32x2_t __ret;
64775  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64776  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64777  return __ret;
64778 }
64779 #endif
64780 
64781 #ifdef __LITTLE_ENDIAN__
64782 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64783  int16x4_t __ret;
64784  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64785  return __ret;
64786 }
64787 #else
64788 __ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64789  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64790  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64791  int16x4_t __ret;
64792  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64794  return __ret;
64795 }
64796 #endif
64797 
64798 #ifdef __LITTLE_ENDIAN__
64799 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64800  uint64x1_t __ret;
64801  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64802  return __ret;
64803 }
64804 #else
64805 __ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64806  uint64x1_t __ret;
64807  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64808  return __ret;
64809 }
64810 #endif
64811 
64812 #ifdef __LITTLE_ENDIAN__
64813 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64814  uint64x2_t __ret;
64815  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64816  return __ret;
64817 }
64818 #else
64819 __ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64820  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64821  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64822  uint64x2_t __ret;
64823  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64824  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64825  return __ret;
64826 }
64827 #endif
64828 
64829 #ifdef __LITTLE_ENDIAN__
64830 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64831  uint64x2_t __ret;
64832  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64833  return __ret;
64834 }
64835 #else
64836 __ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64837  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64838  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64839  uint64x2_t __ret;
64840  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64842  return __ret;
64843 }
64844 #endif
64845 
64846 #ifdef __LITTLE_ENDIAN__
64847 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64848  uint64x2_t __ret;
64849  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64850  return __ret;
64851 }
64852 #else
64853 __ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64854  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64855  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64856  uint64x2_t __ret;
64857  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64859  return __ret;
64860 }
64861 #endif
64862 
64863 #ifdef __LITTLE_ENDIAN__
64864 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64865  uint64x1_t __ret;
64866  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64867  return __ret;
64868 }
64869 #else
64870 __ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64871  uint64x1_t __ret;
64872  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64873  return __ret;
64874 }
64875 #endif
64876 
64877 #ifdef __LITTLE_ENDIAN__
64878 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64879  uint64x1_t __ret;
64880  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64881  return __ret;
64882 }
64883 #else
64884 __ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64885  uint64x1_t __ret;
64886  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64887  return __ret;
64888 }
64889 #endif
64890 
64891 #ifdef __LITTLE_ENDIAN__
64892 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64893  uint64_t __ret;
64894  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64895  return __ret;
64896 }
64897 #else
64898 __ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64899  uint64_t __ret;
64900  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64901  return __ret;
64902 }
64903 #endif
64904 
64905 #ifdef __LITTLE_ENDIAN__
64906 __ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64907  int64_t __ret;
64908  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64909  return __ret;
64910 }
64911 #else
64912 __ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64913  int64_t __ret;
64914  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64915  return __ret;
64916 }
64917 #endif
64918 
64919 #ifdef __LITTLE_ENDIAN__
64920 __ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64921  int8_t __ret;
64922  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64923  return __ret;
64924 }
64925 #else
64926 __ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64927  int8_t __ret;
64928  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64929  return __ret;
64930 }
64931 #endif
64932 
64933 #ifdef __LITTLE_ENDIAN__
64934 __ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64935  int32_t __ret;
64936  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64937  return __ret;
64938 }
64939 #else
64940 __ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64941  int32_t __ret;
64942  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64943  return __ret;
64944 }
64945 #endif
64946 
64947 #ifdef __LITTLE_ENDIAN__
64948 __ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64949  int64_t __ret;
64950  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64951  return __ret;
64952 }
64953 #else
64954 __ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64955  int64_t __ret;
64956  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64957  return __ret;
64958 }
64959 #endif
64960 
64961 #ifdef __LITTLE_ENDIAN__
64962 __ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64963  int16_t __ret;
64964  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64965  return __ret;
64966 }
64967 #else
64968 __ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64969  int16_t __ret;
64970  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64971  return __ret;
64972 }
64973 #endif
64974 
64975 #ifdef __LITTLE_ENDIAN__
64976 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64977  int8x16_t __ret;
64978  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
64979  return __ret;
64980 }
64981 #else
64982 __ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64983  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64984  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64985  int8x16_t __ret;
64986  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
64987  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64988  return __ret;
64989 }
64990 #endif
64991 
64992 #ifdef __LITTLE_ENDIAN__
64993 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64994  int32x4_t __ret;
64995  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
64996  return __ret;
64997 }
64998 #else
64999 __ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
65000  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65001  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65002  int32x4_t __ret;
65003  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
65004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65005  return __ret;
65006 }
65007 #endif
65008 
65009 #ifdef __LITTLE_ENDIAN__
65010 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65011  int64x2_t __ret;
65012  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
65013  return __ret;
65014 }
65015 #else
65016 __ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65017  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65018  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65019  int64x2_t __ret;
65020  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
65021  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65022  return __ret;
65023 }
65024 #endif
65025 
65026 #ifdef __LITTLE_ENDIAN__
65027 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65028  int16x8_t __ret;
65029  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
65030  return __ret;
65031 }
65032 #else
65033 __ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65034  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65035  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65036  int16x8_t __ret;
65037  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
65038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65039  return __ret;
65040 }
65041 #endif
65042 
65043 #ifdef __LITTLE_ENDIAN__
65044 __ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65045  int8x8_t __ret;
65046  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
65047  return __ret;
65048 }
65049 #else
65050 __ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65051  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65052  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65053  int8x8_t __ret;
65054  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
65055  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65056  return __ret;
65057 }
65058 #endif
65059 
65060 #ifdef __LITTLE_ENDIAN__
65061 __ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65062  int32x2_t __ret;
65063  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
65064  return __ret;
65065 }
65066 #else
65067 __ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65068  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65069  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65070  int32x2_t __ret;
65071  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
65072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65073  return __ret;
65074 }
65075 #endif
65076 
65077 #ifdef __LITTLE_ENDIAN__
65078 __ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65079  int64x1_t __ret;
65080  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65081  return __ret;
65082 }
65083 #else
65084 __ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65085  int64x1_t __ret;
65086  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65087  return __ret;
65088 }
65089 #endif
65090 
65091 #ifdef __LITTLE_ENDIAN__
65092 __ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65093  int16x4_t __ret;
65094  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
65095  return __ret;
65096 }
65097 #else
65098 __ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65099  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65100  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65101  int16x4_t __ret;
65102  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
65103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65104  return __ret;
65105 }
65106 #endif
65107 
65108 #ifdef __LITTLE_ENDIAN__
65109 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65110  poly8x8_t __ret;
65111  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65112  return __ret;
65113 }
65114 #else
65115 __ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65116  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65117  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65118  poly8x8_t __ret;
65119  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65120  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65121  return __ret;
65122 }
65123 #endif
65124 
65125 #ifdef __LITTLE_ENDIAN__
65126 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65127  poly16x4_t __ret;
65128  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65129  return __ret;
65130 }
65131 #else
65132 __ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65133  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65134  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65135  poly16x4_t __ret;
65136  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65137  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65138  return __ret;
65139 }
65140 #endif
65141 
65142 #ifdef __LITTLE_ENDIAN__
65143 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65144  poly8x16_t __ret;
65145  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65146  return __ret;
65147 }
65148 #else
65149 __ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65150  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65151  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65152  poly8x16_t __ret;
65153  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65154  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65155  return __ret;
65156 }
65157 #endif
65158 
65159 #ifdef __LITTLE_ENDIAN__
65160 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65161  poly64x2_t __ret;
65162  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65163  return __ret;
65164 }
65165 #else
65166 __ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65167  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65168  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65169  poly64x2_t __ret;
65170  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65171  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65172  return __ret;
65173 }
65174 #endif
65175 
65176 #ifdef __LITTLE_ENDIAN__
65177 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65178  poly16x8_t __ret;
65179  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65180  return __ret;
65181 }
65182 #else
65183 __ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65184  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65185  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65186  poly16x8_t __ret;
65187  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65188  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65189  return __ret;
65190 }
65191 #endif
65192 
65193 #ifdef __LITTLE_ENDIAN__
65194 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65195  uint8x16_t __ret;
65196  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65197  return __ret;
65198 }
65199 #else
65200 __ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65201  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65202  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65203  uint8x16_t __ret;
65204  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65205  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65206  return __ret;
65207 }
65208 #endif
65209 
65210 #ifdef __LITTLE_ENDIAN__
65211 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65212  uint32x4_t __ret;
65213  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65214  return __ret;
65215 }
65216 #else
65217 __ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65218  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65219  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65220  uint32x4_t __ret;
65221  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65223  return __ret;
65224 }
65225 #endif
65226 
65227 #ifdef __LITTLE_ENDIAN__
65228 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65229  uint64x2_t __ret;
65230  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65231  return __ret;
65232 }
65233 #else
65234 __ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65235  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65236  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65237  uint64x2_t __ret;
65238  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65240  return __ret;
65241 }
65242 #endif
65243 
65244 #ifdef __LITTLE_ENDIAN__
65245 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65246  uint16x8_t __ret;
65247  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65248  return __ret;
65249 }
65250 #else
65251 __ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65252  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65253  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65254  uint16x8_t __ret;
65255  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65256  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65257  return __ret;
65258 }
65259 #endif
65260 
65261 #ifdef __LITTLE_ENDIAN__
65262 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65263  int8x16_t __ret;
65264  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65265  return __ret;
65266 }
65267 #else
65268 __ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65269  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65270  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65271  int8x16_t __ret;
65272  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65273  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65274  return __ret;
65275 }
65276 #endif
65277 
65278 #ifdef __LITTLE_ENDIAN__
65279 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65280  float64x2_t __ret;
65281  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65282  return __ret;
65283 }
65284 #else
65285 __ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65286  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65287  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65288  float64x2_t __ret;
65289  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65291  return __ret;
65292 }
65293 #endif
65294 
65295 #ifdef __LITTLE_ENDIAN__
65296 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65297  float32x4_t __ret;
65298  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65299  return __ret;
65300 }
65301 #else
65302 __ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65303  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65304  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65305  float32x4_t __ret;
65306  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65307  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65308  return __ret;
65309 }
65310 #endif
65311 
65312 #ifdef __LITTLE_ENDIAN__
65313 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65314  int32x4_t __ret;
65315  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65316  return __ret;
65317 }
65318 #else
65319 __ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65320  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65321  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65322  int32x4_t __ret;
65323  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65325  return __ret;
65326 }
65327 #endif
65328 
65329 #ifdef __LITTLE_ENDIAN__
65330 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65331  int64x2_t __ret;
65332  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65333  return __ret;
65334 }
65335 #else
65336 __ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65337  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65338  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65339  int64x2_t __ret;
65340  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65342  return __ret;
65343 }
65344 #endif
65345 
65346 #ifdef __LITTLE_ENDIAN__
65347 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65348  int16x8_t __ret;
65349  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65350  return __ret;
65351 }
65352 #else
65353 __ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65354  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65355  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65356  int16x8_t __ret;
65357  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65359  return __ret;
65360 }
65361 #endif
65362 
65363 #ifdef __LITTLE_ENDIAN__
65364 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65365  uint8x8_t __ret;
65366  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65367  return __ret;
65368 }
65369 #else
65370 __ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65371  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65372  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65373  uint8x8_t __ret;
65374  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65376  return __ret;
65377 }
65378 #endif
65379 
65380 #ifdef __LITTLE_ENDIAN__
65381 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65382  uint32x2_t __ret;
65383  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65384  return __ret;
65385 }
65386 #else
65387 __ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65388  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65389  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65390  uint32x2_t __ret;
65391  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65392  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65393  return __ret;
65394 }
65395 #endif
65396 
65397 #ifdef __LITTLE_ENDIAN__
65398 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65399  uint16x4_t __ret;
65400  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65401  return __ret;
65402 }
65403 #else
65404 __ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65405  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65406  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65407  uint16x4_t __ret;
65408  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65410  return __ret;
65411 }
65412 #endif
65413 
65414 #ifdef __LITTLE_ENDIAN__
65415 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65416  int8x8_t __ret;
65417  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65418  return __ret;
65419 }
65420 #else
65421 __ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65422  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65423  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65424  int8x8_t __ret;
65425  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65427  return __ret;
65428 }
65429 #endif
65430 
65431 #ifdef __LITTLE_ENDIAN__
65432 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65433  float32x2_t __ret;
65434  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65435  return __ret;
65436 }
65437 #else
65438 __ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65439  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65440  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65441  float32x2_t __ret;
65442  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65443  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65444  return __ret;
65445 }
65446 #endif
65447 
65448 #ifdef __LITTLE_ENDIAN__
65449 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65450  int32x2_t __ret;
65451  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65452  return __ret;
65453 }
65454 #else
65455 __ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65456  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65457  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65458  int32x2_t __ret;
65459  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65460  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65461  return __ret;
65462 }
65463 #endif
65464 
65465 #ifdef __LITTLE_ENDIAN__
65466 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65467  int16x4_t __ret;
65468  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65469  return __ret;
65470 }
65471 #else
65472 __ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65473  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65474  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65475  int16x4_t __ret;
65476  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65478  return __ret;
65479 }
65480 #endif
65481 
65482 #ifdef __LITTLE_ENDIAN__
65483 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65484  poly8x8_t __ret;
65485  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65486  return __ret;
65487 }
65488 #else
65489 __ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65490  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65491  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65492  poly8x8_t __ret;
65493  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65495  return __ret;
65496 }
65497 #endif
65498 
65499 #ifdef __LITTLE_ENDIAN__
65500 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65501  poly16x4_t __ret;
65502  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65503  return __ret;
65504 }
65505 #else
65506 __ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65507  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65508  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65509  poly16x4_t __ret;
65510  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65512  return __ret;
65513 }
65514 #endif
65515 
65516 #ifdef __LITTLE_ENDIAN__
65517 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65518  poly8x16_t __ret;
65519  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65520  return __ret;
65521 }
65522 #else
65523 __ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65524  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65525  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65526  poly8x16_t __ret;
65527  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65528  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65529  return __ret;
65530 }
65531 #endif
65532 
65533 #ifdef __LITTLE_ENDIAN__
65534 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65535  poly64x2_t __ret;
65536  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65537  return __ret;
65538 }
65539 #else
65540 __ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65541  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65542  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65543  poly64x2_t __ret;
65544  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65546  return __ret;
65547 }
65548 #endif
65549 
65550 #ifdef __LITTLE_ENDIAN__
65551 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65552  poly16x8_t __ret;
65553  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65554  return __ret;
65555 }
65556 #else
65557 __ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65558  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65559  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65560  poly16x8_t __ret;
65561  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65563  return __ret;
65564 }
65565 #endif
65566 
65567 #ifdef __LITTLE_ENDIAN__
65568 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65569  uint8x16_t __ret;
65570  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65571  return __ret;
65572 }
65573 #else
65574 __ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65575  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65576  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65577  uint8x16_t __ret;
65578  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65579  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65580  return __ret;
65581 }
65582 #endif
65583 
65584 #ifdef __LITTLE_ENDIAN__
65585 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65586  uint32x4_t __ret;
65587  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65588  return __ret;
65589 }
65590 #else
65591 __ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65592  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65593  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65594  uint32x4_t __ret;
65595  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65597  return __ret;
65598 }
65599 #endif
65600 
65601 #ifdef __LITTLE_ENDIAN__
65602 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65603  uint64x2_t __ret;
65604  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65605  return __ret;
65606 }
65607 #else
65608 __ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65609  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65610  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65611  uint64x2_t __ret;
65612  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65614  return __ret;
65615 }
65616 #endif
65617 
65618 #ifdef __LITTLE_ENDIAN__
65619 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65620  uint16x8_t __ret;
65621  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65622  return __ret;
65623 }
65624 #else
65625 __ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65626  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65627  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65628  uint16x8_t __ret;
65629  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65630  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65631  return __ret;
65632 }
65633 #endif
65634 
65635 #ifdef __LITTLE_ENDIAN__
65636 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65637  int8x16_t __ret;
65638  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65639  return __ret;
65640 }
65641 #else
65642 __ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65643  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65644  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65645  int8x16_t __ret;
65646  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65648  return __ret;
65649 }
65650 #endif
65651 
65652 #ifdef __LITTLE_ENDIAN__
65653 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65654  float64x2_t __ret;
65655  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65656  return __ret;
65657 }
65658 #else
65659 __ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65660  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65661  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65662  float64x2_t __ret;
65663  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65664  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65665  return __ret;
65666 }
65667 #endif
65668 
65669 #ifdef __LITTLE_ENDIAN__
65670 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65671  float32x4_t __ret;
65672  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65673  return __ret;
65674 }
65675 #else
65676 __ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65677  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65678  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65679  float32x4_t __ret;
65680  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65682  return __ret;
65683 }
65684 #endif
65685 
65686 #ifdef __LITTLE_ENDIAN__
65687 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65688  int32x4_t __ret;
65689  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65690  return __ret;
65691 }
65692 #else
65693 __ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65694  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65695  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65696  int32x4_t __ret;
65697  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65699  return __ret;
65700 }
65701 #endif
65702 
65703 #ifdef __LITTLE_ENDIAN__
65704 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65705  int64x2_t __ret;
65706  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65707  return __ret;
65708 }
65709 #else
65710 __ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65711  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65712  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65713  int64x2_t __ret;
65714  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65715  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65716  return __ret;
65717 }
65718 #endif
65719 
65720 #ifdef __LITTLE_ENDIAN__
65721 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65722  int16x8_t __ret;
65723  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65724  return __ret;
65725 }
65726 #else
65727 __ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65728  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65729  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65730  int16x8_t __ret;
65731  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65733  return __ret;
65734 }
65735 #endif
65736 
65737 #ifdef __LITTLE_ENDIAN__
65738 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65739  uint8x8_t __ret;
65740  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65741  return __ret;
65742 }
65743 #else
65744 __ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65745  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65746  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65747  uint8x8_t __ret;
65748  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65750  return __ret;
65751 }
65752 #endif
65753 
65754 #ifdef __LITTLE_ENDIAN__
65755 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65756  uint32x2_t __ret;
65757  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65758  return __ret;
65759 }
65760 #else
65761 __ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65762  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65763  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65764  uint32x2_t __ret;
65765  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65767  return __ret;
65768 }
65769 #endif
65770 
65771 #ifdef __LITTLE_ENDIAN__
65772 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65773  uint16x4_t __ret;
65774  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65775  return __ret;
65776 }
65777 #else
65778 __ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65779  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65780  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65781  uint16x4_t __ret;
65782  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65784  return __ret;
65785 }
65786 #endif
65787 
65788 #ifdef __LITTLE_ENDIAN__
65789 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65790  int8x8_t __ret;
65791  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65792  return __ret;
65793 }
65794 #else
65795 __ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65796  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65797  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65798  int8x8_t __ret;
65799  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65801  return __ret;
65802 }
65803 #endif
65804 
65805 #ifdef __LITTLE_ENDIAN__
65806 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65807  float32x2_t __ret;
65808  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65809  return __ret;
65810 }
65811 #else
65812 __ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65813  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65814  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65815  float32x2_t __ret;
65816  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65818  return __ret;
65819 }
65820 #endif
65821 
65822 #ifdef __LITTLE_ENDIAN__
65823 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65824  int32x2_t __ret;
65825  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65826  return __ret;
65827 }
65828 #else
65829 __ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65830  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65831  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65832  int32x2_t __ret;
65833  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65835  return __ret;
65836 }
65837 #endif
65838 
65839 #ifdef __LITTLE_ENDIAN__
65840 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65841  int16x4_t __ret;
65842  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65843  return __ret;
65844 }
65845 #else
65846 __ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65847  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65848  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65849  int16x4_t __ret;
65850  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65852  return __ret;
65853 }
65854 #endif
65855 
65856 #ifdef __LITTLE_ENDIAN__
65857 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65858  poly8x8_t __ret;
65859  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65860  return __ret;
65861 }
65862 #else
65863 __ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65864  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65865  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65866  poly8x8_t __ret;
65867  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65868  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65869  return __ret;
65870 }
65871 #endif
65872 
65873 #ifdef __LITTLE_ENDIAN__
65874 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65875  poly16x4_t __ret;
65876  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65877  return __ret;
65878 }
65879 #else
65880 __ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65881  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65882  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65883  poly16x4_t __ret;
65884  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65886  return __ret;
65887 }
65888 #endif
65889 
65890 #ifdef __LITTLE_ENDIAN__
65891 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65892  poly8x16_t __ret;
65893  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65894  return __ret;
65895 }
65896 #else
65897 __ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65898  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65899  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65900  poly8x16_t __ret;
65901  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65902  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65903  return __ret;
65904 }
65905 #endif
65906 
65907 #ifdef __LITTLE_ENDIAN__
65908 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65909  poly64x2_t __ret;
65910  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65911  return __ret;
65912 }
65913 #else
65914 __ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65915  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65916  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65917  poly64x2_t __ret;
65918  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65920  return __ret;
65921 }
65922 #endif
65923 
65924 #ifdef __LITTLE_ENDIAN__
65925 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65926  poly16x8_t __ret;
65927  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65928  return __ret;
65929 }
65930 #else
65931 __ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65932  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65933  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65934  poly16x8_t __ret;
65935  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65936  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65937  return __ret;
65938 }
65939 #endif
65940 
65941 #ifdef __LITTLE_ENDIAN__
65942 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65943  uint8x16_t __ret;
65944  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65945  return __ret;
65946 }
65947 #else
65948 __ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65949  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65950  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65951  uint8x16_t __ret;
65952  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65953  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65954  return __ret;
65955 }
65956 #endif
65957 
65958 #ifdef __LITTLE_ENDIAN__
65959 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65960  uint32x4_t __ret;
65961  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65962  return __ret;
65963 }
65964 #else
65965 __ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65966  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65967  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65968  uint32x4_t __ret;
65969  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65970  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65971  return __ret;
65972 }
65973 #endif
65974 
65975 #ifdef __LITTLE_ENDIAN__
65976 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65977  uint64x2_t __ret;
65978  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65979  return __ret;
65980 }
65981 #else
65982 __ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65983  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65984  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65985  uint64x2_t __ret;
65986  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65987  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65988  return __ret;
65989 }
65990 #endif
65991 
65992 #ifdef __LITTLE_ENDIAN__
65993 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65994  uint16x8_t __ret;
65995  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65996  return __ret;
65997 }
65998 #else
65999 __ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66000  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66001  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66002  uint16x8_t __ret;
66003  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66004  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66005  return __ret;
66006 }
66007 #endif
66008 
66009 #ifdef __LITTLE_ENDIAN__
66010 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66011  int8x16_t __ret;
66012  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66013  return __ret;
66014 }
66015 #else
66016 __ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66017  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66018  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66019  int8x16_t __ret;
66020  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66021  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66022  return __ret;
66023 }
66024 #endif
66025 
66026 #ifdef __LITTLE_ENDIAN__
66027 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66028  float64x2_t __ret;
66029  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66030  return __ret;
66031 }
66032 #else
66033 __ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66034  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66035  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66036  float64x2_t __ret;
66037  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66039  return __ret;
66040 }
66041 #endif
66042 
66043 #ifdef __LITTLE_ENDIAN__
66044 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66045  float32x4_t __ret;
66046  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66047  return __ret;
66048 }
66049 #else
66050 __ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66051  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66052  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66053  float32x4_t __ret;
66054  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66055  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66056  return __ret;
66057 }
66058 #endif
66059 
66060 #ifdef __LITTLE_ENDIAN__
66061 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66062  int32x4_t __ret;
66063  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66064  return __ret;
66065 }
66066 #else
66067 __ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66068  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66069  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66070  int32x4_t __ret;
66071  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66072  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66073  return __ret;
66074 }
66075 #endif
66076 
66077 #ifdef __LITTLE_ENDIAN__
66078 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66079  int64x2_t __ret;
66080  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66081  return __ret;
66082 }
66083 #else
66084 __ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66085  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66086  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66087  int64x2_t __ret;
66088  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66089  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66090  return __ret;
66091 }
66092 #endif
66093 
66094 #ifdef __LITTLE_ENDIAN__
66095 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66096  int16x8_t __ret;
66097  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66098  return __ret;
66099 }
66100 #else
66101 __ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66102  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66103  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66104  int16x8_t __ret;
66105  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66107  return __ret;
66108 }
66109 #endif
66110 
66111 #ifdef __LITTLE_ENDIAN__
66112 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66113  uint8x8_t __ret;
66114  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66115  return __ret;
66116 }
66117 #else
66118 __ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66119  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66120  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66121  uint8x8_t __ret;
66122  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66123  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66124  return __ret;
66125 }
66126 #endif
66127 
66128 #ifdef __LITTLE_ENDIAN__
66129 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66130  uint32x2_t __ret;
66131  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66132  return __ret;
66133 }
66134 #else
66135 __ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66136  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66137  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66138  uint32x2_t __ret;
66139  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66140  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66141  return __ret;
66142 }
66143 #endif
66144 
66145 #ifdef __LITTLE_ENDIAN__
66146 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66147  uint16x4_t __ret;
66148  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66149  return __ret;
66150 }
66151 #else
66152 __ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66153  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66154  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66155  uint16x4_t __ret;
66156  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66157  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66158  return __ret;
66159 }
66160 #endif
66161 
66162 #ifdef __LITTLE_ENDIAN__
66163 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66164  int8x8_t __ret;
66165  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66166  return __ret;
66167 }
66168 #else
66169 __ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66170  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66171  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66172  int8x8_t __ret;
66173  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66174  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66175  return __ret;
66176 }
66177 #endif
66178 
66179 #ifdef __LITTLE_ENDIAN__
66180 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66181  float32x2_t __ret;
66182  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66183  return __ret;
66184 }
66185 #else
66186 __ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66187  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66188  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66189  float32x2_t __ret;
66190  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66191  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66192  return __ret;
66193 }
66194 #endif
66195 
66196 #ifdef __LITTLE_ENDIAN__
66197 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66198  int32x2_t __ret;
66199  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66200  return __ret;
66201 }
66202 #else
66203 __ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66204  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66205  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66206  int32x2_t __ret;
66207  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66208  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66209  return __ret;
66210 }
66211 #endif
66212 
66213 #ifdef __LITTLE_ENDIAN__
66214 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66215  int16x4_t __ret;
66216  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66217  return __ret;
66218 }
66219 #else
66220 __ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66221  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66222  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66223  int16x4_t __ret;
66224  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66225  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66226  return __ret;
66227 }
66228 #endif
66229 
66230 #ifdef __LITTLE_ENDIAN__
66231 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66232  poly8x8_t __ret;
66233  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66234  return __ret;
66235 }
66236 #else
66237 __ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66238  poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66239  poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66240  poly8x8_t __ret;
66241  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66243  return __ret;
66244 }
66245 #endif
66246 
66247 #ifdef __LITTLE_ENDIAN__
66248 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66249  poly16x4_t __ret;
66250  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66251  return __ret;
66252 }
66253 #else
66254 __ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66255  poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66256  poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66257  poly16x4_t __ret;
66258  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66260  return __ret;
66261 }
66262 #endif
66263 
66264 #ifdef __LITTLE_ENDIAN__
66265 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66266  poly8x16_t __ret;
66267  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66268  return __ret;
66269 }
66270 #else
66271 __ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66272  poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66273  poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66274  poly8x16_t __ret;
66275  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66276  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66277  return __ret;
66278 }
66279 #endif
66280 
66281 #ifdef __LITTLE_ENDIAN__
66282 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66283  poly64x2_t __ret;
66284  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66285  return __ret;
66286 }
66287 #else
66288 __ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66289  poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66290  poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66291  poly64x2_t __ret;
66292  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66294  return __ret;
66295 }
66296 #endif
66297 
66298 #ifdef __LITTLE_ENDIAN__
66299 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66300  poly16x8_t __ret;
66301  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66302  return __ret;
66303 }
66304 #else
66305 __ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66306  poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66307  poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66308  poly16x8_t __ret;
66309  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66311  return __ret;
66312 }
66313 #endif
66314 
66315 #ifdef __LITTLE_ENDIAN__
66316 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66317  uint8x16_t __ret;
66318  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66319  return __ret;
66320 }
66321 #else
66322 __ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66323  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66324  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66325  uint8x16_t __ret;
66326  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66327  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66328  return __ret;
66329 }
66330 #endif
66331 
66332 #ifdef __LITTLE_ENDIAN__
66333 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66334  uint32x4_t __ret;
66335  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66336  return __ret;
66337 }
66338 #else
66339 __ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66340  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66341  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66342  uint32x4_t __ret;
66343  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66344  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66345  return __ret;
66346 }
66347 #endif
66348 
66349 #ifdef __LITTLE_ENDIAN__
66350 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66351  uint64x2_t __ret;
66352  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66353  return __ret;
66354 }
66355 #else
66356 __ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66357  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66358  uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66359  uint64x2_t __ret;
66360  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66361  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66362  return __ret;
66363 }
66364 #endif
66365 
66366 #ifdef __LITTLE_ENDIAN__
66367 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66368  uint16x8_t __ret;
66369  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66370  return __ret;
66371 }
66372 #else
66373 __ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66374  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66375  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66376  uint16x8_t __ret;
66377  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66378  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66379  return __ret;
66380 }
66381 #endif
66382 
66383 #ifdef __LITTLE_ENDIAN__
66384 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66385  int8x16_t __ret;
66386  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66387  return __ret;
66388 }
66389 #else
66390 __ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66391  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66392  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66393  int8x16_t __ret;
66394  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66395  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66396  return __ret;
66397 }
66398 #endif
66399 
66400 #ifdef __LITTLE_ENDIAN__
66401 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66402  float64x2_t __ret;
66403  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66404  return __ret;
66405 }
66406 #else
66407 __ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66408  float64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66409  float64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66410  float64x2_t __ret;
66411  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66412  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66413  return __ret;
66414 }
66415 #endif
66416 
66417 #ifdef __LITTLE_ENDIAN__
66418 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66419  float32x4_t __ret;
66420  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66421  return __ret;
66422 }
66423 #else
66424 __ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66425  float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66426  float32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66427  float32x4_t __ret;
66428  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66429  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66430  return __ret;
66431 }
66432 #endif
66433 
66434 #ifdef __LITTLE_ENDIAN__
66435 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66436  int32x4_t __ret;
66437  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66438  return __ret;
66439 }
66440 #else
66441 __ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66442  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66443  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66444  int32x4_t __ret;
66445  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66446  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66447  return __ret;
66448 }
66449 #endif
66450 
66451 #ifdef __LITTLE_ENDIAN__
66452 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66453  int64x2_t __ret;
66454  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66455  return __ret;
66456 }
66457 #else
66458 __ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66459  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66460  int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66461  int64x2_t __ret;
66462  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66463  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66464  return __ret;
66465 }
66466 #endif
66467 
66468 #ifdef __LITTLE_ENDIAN__
66469 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66470  int16x8_t __ret;
66471  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66472  return __ret;
66473 }
66474 #else
66475 __ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66476  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66477  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66478  int16x8_t __ret;
66479  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66481  return __ret;
66482 }
66483 #endif
66484 
66485 #ifdef __LITTLE_ENDIAN__
66486 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66487  uint8x8_t __ret;
66488  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66489  return __ret;
66490 }
66491 #else
66492 __ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66493  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66494  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66495  uint8x8_t __ret;
66496  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66497  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66498  return __ret;
66499 }
66500 #endif
66501 
66502 #ifdef __LITTLE_ENDIAN__
66503 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66504  uint32x2_t __ret;
66505  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66506  return __ret;
66507 }
66508 #else
66509 __ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66510  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66511  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66512  uint32x2_t __ret;
66513  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66515  return __ret;
66516 }
66517 #endif
66518 
66519 #ifdef __LITTLE_ENDIAN__
66520 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66521  uint16x4_t __ret;
66522  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66523  return __ret;
66524 }
66525 #else
66526 __ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66527  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66528  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66529  uint16x4_t __ret;
66530  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66531  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66532  return __ret;
66533 }
66534 #endif
66535 
66536 #ifdef __LITTLE_ENDIAN__
66537 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66538  int8x8_t __ret;
66539  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66540  return __ret;
66541 }
66542 #else
66543 __ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66544  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66545  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66546  int8x8_t __ret;
66547  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66548  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66549  return __ret;
66550 }
66551 #endif
66552 
66553 #ifdef __LITTLE_ENDIAN__
66554 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66555  float32x2_t __ret;
66556  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66557  return __ret;
66558 }
66559 #else
66560 __ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66561  float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66562  float32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66563  float32x2_t __ret;
66564  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66565  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66566  return __ret;
66567 }
66568 #endif
66569 
66570 #ifdef __LITTLE_ENDIAN__
66571 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66572  int32x2_t __ret;
66573  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66574  return __ret;
66575 }
66576 #else
66577 __ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66578  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66579  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66580  int32x2_t __ret;
66581  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66582  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66583  return __ret;
66584 }
66585 #endif
66586 
66587 #ifdef __LITTLE_ENDIAN__
66588 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66589  int16x4_t __ret;
66590  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66591  return __ret;
66592 }
66593 #else
66594 __ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66595  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66596  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66597  int16x4_t __ret;
66598  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66599  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66600  return __ret;
66601 }
66602 #endif
66603 
66604 #endif
66605 #ifdef __LITTLE_ENDIAN__
66606 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66607  uint8x16_t __ret;
66608  __ret = __p0 + vabdq_u8(__p1, __p2);
66609  return __ret;
66610 }
66611 #else
66612 __ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66613  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66614  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66615  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66616  uint8x16_t __ret;
66617  __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
66618  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66619  return __ret;
66620 }
66621 #endif
66622 
66623 #ifdef __LITTLE_ENDIAN__
66624 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66625  uint32x4_t __ret;
66626  __ret = __p0 + vabdq_u32(__p1, __p2);
66627  return __ret;
66628 }
66629 #else
66630 __ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66631  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66632  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66633  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66634  uint32x4_t __ret;
66635  __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
66636  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66637  return __ret;
66638 }
66639 #endif
66640 
66641 #ifdef __LITTLE_ENDIAN__
66642 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66643  uint16x8_t __ret;
66644  __ret = __p0 + vabdq_u16(__p1, __p2);
66645  return __ret;
66646 }
66647 #else
66648 __ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66649  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66650  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66651  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66652  uint16x8_t __ret;
66653  __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
66654  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66655  return __ret;
66656 }
66657 #endif
66658 
66659 #ifdef __LITTLE_ENDIAN__
66660 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66661  int8x16_t __ret;
66662  __ret = __p0 + vabdq_s8(__p1, __p2);
66663  return __ret;
66664 }
66665 #else
66666 __ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66667  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66668  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66669  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66670  int8x16_t __ret;
66671  __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
66672  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66673  return __ret;
66674 }
66675 #endif
66676 
66677 #ifdef __LITTLE_ENDIAN__
66678 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66679  int32x4_t __ret;
66680  __ret = __p0 + vabdq_s32(__p1, __p2);
66681  return __ret;
66682 }
66683 #else
66684 __ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66685  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66686  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66687  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66688  int32x4_t __ret;
66689  __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
66690  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66691  return __ret;
66692 }
66693 #endif
66694 
66695 #ifdef __LITTLE_ENDIAN__
66696 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66697  int16x8_t __ret;
66698  __ret = __p0 + vabdq_s16(__p1, __p2);
66699  return __ret;
66700 }
66701 #else
66702 __ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66703  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66704  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66705  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66706  int16x8_t __ret;
66707  __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
66708  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66709  return __ret;
66710 }
66711 #endif
66712 
66713 #ifdef __LITTLE_ENDIAN__
66714 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66715  uint8x8_t __ret;
66716  __ret = __p0 + vabd_u8(__p1, __p2);
66717  return __ret;
66718 }
66719 #else
66720 __ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66721  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66722  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66723  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66724  uint8x8_t __ret;
66725  __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
66726  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66727  return __ret;
66728 }
66729 #endif
66730 
66731 #ifdef __LITTLE_ENDIAN__
66732 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66733  uint32x2_t __ret;
66734  __ret = __p0 + vabd_u32(__p1, __p2);
66735  return __ret;
66736 }
66737 #else
66738 __ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66739  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66740  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66741  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66742  uint32x2_t __ret;
66743  __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
66744  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66745  return __ret;
66746 }
66747 #endif
66748 
66749 #ifdef __LITTLE_ENDIAN__
66750 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66751  uint16x4_t __ret;
66752  __ret = __p0 + vabd_u16(__p1, __p2);
66753  return __ret;
66754 }
66755 #else
66756 __ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66757  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66758  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66759  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66760  uint16x4_t __ret;
66761  __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
66762  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66763  return __ret;
66764 }
66765 #endif
66766 
66767 #ifdef __LITTLE_ENDIAN__
66768 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66769  int8x8_t __ret;
66770  __ret = __p0 + vabd_s8(__p1, __p2);
66771  return __ret;
66772 }
66773 #else
66774 __ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66775  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66776  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66777  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66778  int8x8_t __ret;
66779  __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
66780  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66781  return __ret;
66782 }
66783 #endif
66784 
66785 #ifdef __LITTLE_ENDIAN__
66786 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66787  int32x2_t __ret;
66788  __ret = __p0 + vabd_s32(__p1, __p2);
66789  return __ret;
66790 }
66791 #else
66792 __ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66793  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66794  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66795  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66796  int32x2_t __ret;
66797  __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
66798  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66799  return __ret;
66800 }
66801 #endif
66802 
66803 #ifdef __LITTLE_ENDIAN__
66804 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66805  int16x4_t __ret;
66806  __ret = __p0 + vabd_s16(__p1, __p2);
66807  return __ret;
66808 }
66809 #else
66810 __ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66811  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66812  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66813  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66814  int16x4_t __ret;
66815  __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
66816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66817  return __ret;
66818 }
66819 #endif
66820 
66821 #ifdef __LITTLE_ENDIAN__
66822 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66823  uint16x8_t __ret;
66824  __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
66825  return __ret;
66826 }
66827 #else
66828 __ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66829  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66830  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66831  uint16x8_t __ret;
66832  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
66833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66834  return __ret;
66835 }
66836 __ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66837  uint16x8_t __ret;
66838  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
66839  return __ret;
66840 }
66841 #endif
66842 
66843 #ifdef __LITTLE_ENDIAN__
66844 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66845  uint64x2_t __ret;
66846  __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
66847  return __ret;
66848 }
66849 #else
66850 __ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66851  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66852  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66853  uint64x2_t __ret;
66854  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
66855  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66856  return __ret;
66857 }
66858 __ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66859  uint64x2_t __ret;
66860  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
66861  return __ret;
66862 }
66863 #endif
66864 
66865 #ifdef __LITTLE_ENDIAN__
66866 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66867  uint32x4_t __ret;
66868  __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
66869  return __ret;
66870 }
66871 #else
66872 __ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66873  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66874  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66875  uint32x4_t __ret;
66876  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
66877  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66878  return __ret;
66879 }
66880 __ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66881  uint32x4_t __ret;
66882  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
66883  return __ret;
66884 }
66885 #endif
66886 
66887 #ifdef __LITTLE_ENDIAN__
66888 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66889  int16x8_t __ret;
66890  __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
66891  return __ret;
66892 }
66893 #else
66894 __ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66895  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66896  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66897  int16x8_t __ret;
66898  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
66899  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66900  return __ret;
66901 }
66902 __ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66903  int16x8_t __ret;
66904  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
66905  return __ret;
66906 }
66907 #endif
66908 
66909 #ifdef __LITTLE_ENDIAN__
66910 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66911  int64x2_t __ret;
66912  __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
66913  return __ret;
66914 }
66915 #else
66916 __ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66917  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66918  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66919  int64x2_t __ret;
66920  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
66921  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66922  return __ret;
66923 }
66924 __ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66925  int64x2_t __ret;
66926  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
66927  return __ret;
66928 }
66929 #endif
66930 
66931 #ifdef __LITTLE_ENDIAN__
66932 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66933  int32x4_t __ret;
66934  __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
66935  return __ret;
66936 }
66937 #else
66938 __ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66939  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66940  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66941  int32x4_t __ret;
66942  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
66943  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66944  return __ret;
66945 }
66946 __ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66947  int32x4_t __ret;
66948  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
66949  return __ret;
66950 }
66951 #endif
66952 
66953 #ifdef __LITTLE_ENDIAN__
66954 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66955  uint16x8_t __ret;
66956  __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
66957  return __ret;
66958 }
66959 #else
66960 __ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66961  uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66962  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66963  uint16x8_t __ret;
66964  __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
66965  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66966  return __ret;
66967 }
66968 #endif
66969 
66970 #ifdef __LITTLE_ENDIAN__
66971 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66972  uint64x2_t __ret;
66973  __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
66974  return __ret;
66975 }
66976 #else
66977 __ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66978  uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66979  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66980  uint64x2_t __ret;
66981  __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
66982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66983  return __ret;
66984 }
66985 #endif
66986 
66987 #ifdef __LITTLE_ENDIAN__
66988 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66989  uint32x4_t __ret;
66990  __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
66991  return __ret;
66992 }
66993 #else
66994 __ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66995  uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66996  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66997  uint32x4_t __ret;
66998  __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
66999  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67000  return __ret;
67001 }
67002 #endif
67003 
67004 #ifdef __LITTLE_ENDIAN__
67005 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67006  int16x8_t __ret;
67007  __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
67008  return __ret;
67009 }
67010 #else
67011 __ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67012  int8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67013  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67014  int16x8_t __ret;
67015  __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
67016  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67017  return __ret;
67018 }
67019 #endif
67020 
67021 #ifdef __LITTLE_ENDIAN__
67022 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67023  int64x2_t __ret;
67024  __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
67025  return __ret;
67026 }
67027 #else
67028 __ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67029  int32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67030  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67031  int64x2_t __ret;
67032  __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
67033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67034  return __ret;
67035 }
67036 #endif
67037 
67038 #ifdef __LITTLE_ENDIAN__
67039 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67040  int32x4_t __ret;
67041  __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
67042  return __ret;
67043 }
67044 #else
67045 __ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67046  int16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67047  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67048  int32x4_t __ret;
67049  __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
67050  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67051  return __ret;
67052 }
67053 #endif
67054 
67055 #ifdef __LITTLE_ENDIAN__
67056 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67057  uint16x8_t __ret;
67058  __ret = __p0 + vmovl_u8(__p1);
67059  return __ret;
67060 }
67061 #else
67062 __ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67063  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67064  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67065  uint16x8_t __ret;
67066  __ret = __rev0 + __noswap_vmovl_u8(__rev1);
67067  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67068  return __ret;
67069 }
67070 #endif
67071 
67072 #ifdef __LITTLE_ENDIAN__
67073 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67074  uint64x2_t __ret;
67075  __ret = __p0 + vmovl_u32(__p1);
67076  return __ret;
67077 }
67078 #else
67079 __ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67080  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67081  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67082  uint64x2_t __ret;
67083  __ret = __rev0 + __noswap_vmovl_u32(__rev1);
67084  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67085  return __ret;
67086 }
67087 #endif
67088 
67089 #ifdef __LITTLE_ENDIAN__
67090 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67091  uint32x4_t __ret;
67092  __ret = __p0 + vmovl_u16(__p1);
67093  return __ret;
67094 }
67095 #else
67096 __ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67097  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67098  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67099  uint32x4_t __ret;
67100  __ret = __rev0 + __noswap_vmovl_u16(__rev1);
67101  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67102  return __ret;
67103 }
67104 #endif
67105 
67106 #ifdef __LITTLE_ENDIAN__
67107 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67108  int16x8_t __ret;
67109  __ret = __p0 + vmovl_s8(__p1);
67110  return __ret;
67111 }
67112 #else
67113 __ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67114  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67115  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67116  int16x8_t __ret;
67117  __ret = __rev0 + __noswap_vmovl_s8(__rev1);
67118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67119  return __ret;
67120 }
67121 #endif
67122 
67123 #ifdef __LITTLE_ENDIAN__
67124 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67125  int64x2_t __ret;
67126  __ret = __p0 + vmovl_s32(__p1);
67127  return __ret;
67128 }
67129 #else
67130 __ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67131  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67132  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67133  int64x2_t __ret;
67134  __ret = __rev0 + __noswap_vmovl_s32(__rev1);
67135  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67136  return __ret;
67137 }
67138 #endif
67139 
67140 #ifdef __LITTLE_ENDIAN__
67141 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67142  int32x4_t __ret;
67143  __ret = __p0 + vmovl_s16(__p1);
67144  return __ret;
67145 }
67146 #else
67147 __ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67148  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67149  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67150  int32x4_t __ret;
67151  __ret = __rev0 + __noswap_vmovl_s16(__rev1);
67152  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67153  return __ret;
67154 }
67155 #endif
67156 
67157 #ifdef __LITTLE_ENDIAN__
67158 #define vget_lane_f16(__p0_242, __p1_242) __extension__ ({ \
67159  float16x4_t __s0_242 = __p0_242; \
67160  float16_t __ret_242; \
67161 float16x4_t __reint_242 = __s0_242; \
67162 int16_t __reint1_242 = vget_lane_s16(*(int16x4_t *) &__reint_242, __p1_242); \
67163  __ret_242 = *(float16_t *) &__reint1_242; \
67164  __ret_242; \
67165 })
67166 #else
67167 #define vget_lane_f16(__p0_243, __p1_243) __extension__ ({ \
67168  float16x4_t __s0_243 = __p0_243; \
67169  float16x4_t __rev0_243; __rev0_243 = __builtin_shufflevector(__s0_243, __s0_243, 3, 2, 1, 0); \
67170  float16_t __ret_243; \
67171 float16x4_t __reint_243 = __rev0_243; \
67172 int16_t __reint1_243 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_243, __p1_243); \
67173  __ret_243 = *(float16_t *) &__reint1_243; \
67174  __ret_243; \
67175 })
67176 #endif
67177 
67178 #ifdef __LITTLE_ENDIAN__
67179 #define vgetq_lane_f16(__p0_244, __p1_244) __extension__ ({ \
67180  float16x8_t __s0_244 = __p0_244; \
67181  float16_t __ret_244; \
67182 float16x8_t __reint_244 = __s0_244; \
67183 int16_t __reint1_244 = vgetq_lane_s16(*(int16x8_t *) &__reint_244, __p1_244); \
67184  __ret_244 = *(float16_t *) &__reint1_244; \
67185  __ret_244; \
67186 })
67187 #else
67188 #define vgetq_lane_f16(__p0_245, __p1_245) __extension__ ({ \
67189  float16x8_t __s0_245 = __p0_245; \
67190  float16x8_t __rev0_245; __rev0_245 = __builtin_shufflevector(__s0_245, __s0_245, 7, 6, 5, 4, 3, 2, 1, 0); \
67191  float16_t __ret_245; \
67192 float16x8_t __reint_245 = __rev0_245; \
67193 int16_t __reint1_245 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_245, __p1_245); \
67194  __ret_245 = *(float16_t *) &__reint1_245; \
67195  __ret_245; \
67196 })
67197 #endif
67198 
67199 #ifdef __LITTLE_ENDIAN__
67200 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67201  uint16x8_t __ret;
67202  __ret = __p0 + vmull_u8(__p1, __p2);
67203  return __ret;
67204 }
67205 #else
67206 __ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67207  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67208  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67209  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67210  uint16x8_t __ret;
67211  __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
67212  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67213  return __ret;
67214 }
67215 __ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67216  uint16x8_t __ret;
67217  __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
67218  return __ret;
67219 }
67220 #endif
67221 
67222 #ifdef __LITTLE_ENDIAN__
67223 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67224  uint64x2_t __ret;
67225  __ret = __p0 + vmull_u32(__p1, __p2);
67226  return __ret;
67227 }
67228 #else
67229 __ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67230  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67231  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67232  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67233  uint64x2_t __ret;
67234  __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
67235  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67236  return __ret;
67237 }
67238 __ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67239  uint64x2_t __ret;
67240  __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
67241  return __ret;
67242 }
67243 #endif
67244 
67245 #ifdef __LITTLE_ENDIAN__
67246 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67247  uint32x4_t __ret;
67248  __ret = __p0 + vmull_u16(__p1, __p2);
67249  return __ret;
67250 }
67251 #else
67252 __ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67253  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67254  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67255  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67256  uint32x4_t __ret;
67257  __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
67258  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67259  return __ret;
67260 }
67261 __ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67262  uint32x4_t __ret;
67263  __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
67264  return __ret;
67265 }
67266 #endif
67267 
67268 #ifdef __LITTLE_ENDIAN__
67269 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67270  int16x8_t __ret;
67271  __ret = __p0 + vmull_s8(__p1, __p2);
67272  return __ret;
67273 }
67274 #else
67275 __ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67276  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67277  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67278  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67279  int16x8_t __ret;
67280  __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
67281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67282  return __ret;
67283 }
67284 __ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67285  int16x8_t __ret;
67286  __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
67287  return __ret;
67288 }
67289 #endif
67290 
67291 #ifdef __LITTLE_ENDIAN__
67292 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67293  int64x2_t __ret;
67294  __ret = __p0 + vmull_s32(__p1, __p2);
67295  return __ret;
67296 }
67297 #else
67298 __ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67299  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67300  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67301  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67302  int64x2_t __ret;
67303  __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
67304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67305  return __ret;
67306 }
67307 __ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67308  int64x2_t __ret;
67309  __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
67310  return __ret;
67311 }
67312 #endif
67313 
67314 #ifdef __LITTLE_ENDIAN__
67315 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67316  int32x4_t __ret;
67317  __ret = __p0 + vmull_s16(__p1, __p2);
67318  return __ret;
67319 }
67320 #else
67321 __ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67322  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67323  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67324  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67325  int32x4_t __ret;
67326  __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
67327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67328  return __ret;
67329 }
67330 __ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67331  int32x4_t __ret;
67332  __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
67333  return __ret;
67334 }
67335 #endif
67336 
67337 #ifdef __LITTLE_ENDIAN__
67338 #define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67339  uint64x2_t __s0 = __p0; \
67340  uint32x2_t __s1 = __p1; \
67341  uint32x2_t __s2 = __p2; \
67342  uint64x2_t __ret; \
67343  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67344  __ret; \
67345 })
67346 #else
67347 #define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67348  uint64x2_t __s0 = __p0; \
67349  uint32x2_t __s1 = __p1; \
67350  uint32x2_t __s2 = __p2; \
67351  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67352  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67353  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67354  uint64x2_t __ret; \
67355  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67356  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67357  __ret; \
67358 })
67359 #endif
67360 
67361 #ifdef __LITTLE_ENDIAN__
67362 #define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67363  uint32x4_t __s0 = __p0; \
67364  uint16x4_t __s1 = __p1; \
67365  uint16x4_t __s2 = __p2; \
67366  uint32x4_t __ret; \
67367  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67368  __ret; \
67369 })
67370 #else
67371 #define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67372  uint32x4_t __s0 = __p0; \
67373  uint16x4_t __s1 = __p1; \
67374  uint16x4_t __s2 = __p2; \
67375  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67376  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67377  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67378  uint32x4_t __ret; \
67379  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67380  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67381  __ret; \
67382 })
67383 #endif
67384 
67385 #ifdef __LITTLE_ENDIAN__
67386 #define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67387  int64x2_t __s0 = __p0; \
67388  int32x2_t __s1 = __p1; \
67389  int32x2_t __s2 = __p2; \
67390  int64x2_t __ret; \
67391  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67392  __ret; \
67393 })
67394 #else
67395 #define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67396  int64x2_t __s0 = __p0; \
67397  int32x2_t __s1 = __p1; \
67398  int32x2_t __s2 = __p2; \
67399  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67400  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67401  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67402  int64x2_t __ret; \
67403  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67405  __ret; \
67406 })
67407 #endif
67408 
67409 #ifdef __LITTLE_ENDIAN__
67410 #define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67411  int32x4_t __s0 = __p0; \
67412  int16x4_t __s1 = __p1; \
67413  int16x4_t __s2 = __p2; \
67414  int32x4_t __ret; \
67415  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67416  __ret; \
67417 })
67418 #else
67419 #define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67420  int32x4_t __s0 = __p0; \
67421  int16x4_t __s1 = __p1; \
67422  int16x4_t __s2 = __p2; \
67423  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67424  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67425  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67426  int32x4_t __ret; \
67427  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67429  __ret; \
67430 })
67431 #endif
67432 
67433 #ifdef __LITTLE_ENDIAN__
67434 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67435  uint64x2_t __ret;
67436  __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67437  return __ret;
67438 }
67439 #else
67440 __ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67441  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67442  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67443  uint64x2_t __ret;
67444  __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67445  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67446  return __ret;
67447 }
67448 __ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67449  uint64x2_t __ret;
67450  __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67451  return __ret;
67452 }
67453 #endif
67454 
67455 #ifdef __LITTLE_ENDIAN__
67456 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67457  uint32x4_t __ret;
67458  __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67459  return __ret;
67460 }
67461 #else
67462 __ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67463  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67464  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67465  uint32x4_t __ret;
67466  __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67467  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67468  return __ret;
67469 }
67470 __ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67471  uint32x4_t __ret;
67472  __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67473  return __ret;
67474 }
67475 #endif
67476 
67477 #ifdef __LITTLE_ENDIAN__
67478 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67479  int64x2_t __ret;
67480  __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67481  return __ret;
67482 }
67483 #else
67484 __ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67485  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67486  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67487  int64x2_t __ret;
67488  __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67490  return __ret;
67491 }
67492 __ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67493  int64x2_t __ret;
67494  __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67495  return __ret;
67496 }
67497 #endif
67498 
67499 #ifdef __LITTLE_ENDIAN__
67500 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67501  int32x4_t __ret;
67502  __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67503  return __ret;
67504 }
67505 #else
67506 __ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67507  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67508  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67509  int32x4_t __ret;
67510  __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67512  return __ret;
67513 }
67514 __ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67515  int32x4_t __ret;
67516  __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67517  return __ret;
67518 }
67519 #endif
67520 
67521 #ifdef __LITTLE_ENDIAN__
67522 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67523  uint16x8_t __ret;
67524  __ret = __p0 - vmull_u8(__p1, __p2);
67525  return __ret;
67526 }
67527 #else
67528 __ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67529  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67530  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67531  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67532  uint16x8_t __ret;
67533  __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
67534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67535  return __ret;
67536 }
67537 __ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67538  uint16x8_t __ret;
67539  __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
67540  return __ret;
67541 }
67542 #endif
67543 
67544 #ifdef __LITTLE_ENDIAN__
67545 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67546  uint64x2_t __ret;
67547  __ret = __p0 - vmull_u32(__p1, __p2);
67548  return __ret;
67549 }
67550 #else
67551 __ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67552  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67553  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67554  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67555  uint64x2_t __ret;
67556  __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
67557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67558  return __ret;
67559 }
67560 __ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67561  uint64x2_t __ret;
67562  __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
67563  return __ret;
67564 }
67565 #endif
67566 
67567 #ifdef __LITTLE_ENDIAN__
67568 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67569  uint32x4_t __ret;
67570  __ret = __p0 - vmull_u16(__p1, __p2);
67571  return __ret;
67572 }
67573 #else
67574 __ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67575  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67576  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67577  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67578  uint32x4_t __ret;
67579  __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
67580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67581  return __ret;
67582 }
67583 __ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67584  uint32x4_t __ret;
67585  __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
67586  return __ret;
67587 }
67588 #endif
67589 
67590 #ifdef __LITTLE_ENDIAN__
67591 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67592  int16x8_t __ret;
67593  __ret = __p0 - vmull_s8(__p1, __p2);
67594  return __ret;
67595 }
67596 #else
67597 __ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67598  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67599  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67600  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67601  int16x8_t __ret;
67602  __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
67603  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67604  return __ret;
67605 }
67606 __ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67607  int16x8_t __ret;
67608  __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
67609  return __ret;
67610 }
67611 #endif
67612 
67613 #ifdef __LITTLE_ENDIAN__
67614 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67615  int64x2_t __ret;
67616  __ret = __p0 - vmull_s32(__p1, __p2);
67617  return __ret;
67618 }
67619 #else
67620 __ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67621  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67622  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67623  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67624  int64x2_t __ret;
67625  __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
67626  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67627  return __ret;
67628 }
67629 __ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67630  int64x2_t __ret;
67631  __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
67632  return __ret;
67633 }
67634 #endif
67635 
67636 #ifdef __LITTLE_ENDIAN__
67637 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67638  int32x4_t __ret;
67639  __ret = __p0 - vmull_s16(__p1, __p2);
67640  return __ret;
67641 }
67642 #else
67643 __ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67644  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67645  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67646  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67647  int32x4_t __ret;
67648  __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
67649  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67650  return __ret;
67651 }
67652 __ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67653  int32x4_t __ret;
67654  __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
67655  return __ret;
67656 }
67657 #endif
67658 
67659 #ifdef __LITTLE_ENDIAN__
67660 #define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67661  uint64x2_t __s0 = __p0; \
67662  uint32x2_t __s1 = __p1; \
67663  uint32x2_t __s2 = __p2; \
67664  uint64x2_t __ret; \
67665  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67666  __ret; \
67667 })
67668 #else
67669 #define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67670  uint64x2_t __s0 = __p0; \
67671  uint32x2_t __s1 = __p1; \
67672  uint32x2_t __s2 = __p2; \
67673  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67674  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67675  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67676  uint64x2_t __ret; \
67677  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67678  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67679  __ret; \
67680 })
67681 #endif
67682 
67683 #ifdef __LITTLE_ENDIAN__
67684 #define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67685  uint32x4_t __s0 = __p0; \
67686  uint16x4_t __s1 = __p1; \
67687  uint16x4_t __s2 = __p2; \
67688  uint32x4_t __ret; \
67689  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67690  __ret; \
67691 })
67692 #else
67693 #define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67694  uint32x4_t __s0 = __p0; \
67695  uint16x4_t __s1 = __p1; \
67696  uint16x4_t __s2 = __p2; \
67697  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67698  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67699  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67700  uint32x4_t __ret; \
67701  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67703  __ret; \
67704 })
67705 #endif
67706 
67707 #ifdef __LITTLE_ENDIAN__
67708 #define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67709  int64x2_t __s0 = __p0; \
67710  int32x2_t __s1 = __p1; \
67711  int32x2_t __s2 = __p2; \
67712  int64x2_t __ret; \
67713  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67714  __ret; \
67715 })
67716 #else
67717 #define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67718  int64x2_t __s0 = __p0; \
67719  int32x2_t __s1 = __p1; \
67720  int32x2_t __s2 = __p2; \
67721  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67722  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67723  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67724  int64x2_t __ret; \
67725  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67727  __ret; \
67728 })
67729 #endif
67730 
67731 #ifdef __LITTLE_ENDIAN__
67732 #define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67733  int32x4_t __s0 = __p0; \
67734  int16x4_t __s1 = __p1; \
67735  int16x4_t __s2 = __p2; \
67736  int32x4_t __ret; \
67737  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67738  __ret; \
67739 })
67740 #else
67741 #define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67742  int32x4_t __s0 = __p0; \
67743  int16x4_t __s1 = __p1; \
67744  int16x4_t __s2 = __p2; \
67745  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67746  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67747  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67748  int32x4_t __ret; \
67749  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67750  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67751  __ret; \
67752 })
67753 #endif
67754 
67755 #ifdef __LITTLE_ENDIAN__
67756 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67757  uint64x2_t __ret;
67758  __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67759  return __ret;
67760 }
67761 #else
67762 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67763  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67764  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67765  uint64x2_t __ret;
67766  __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67767  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67768  return __ret;
67769 }
67770 __ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67771  uint64x2_t __ret;
67772  __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67773  return __ret;
67774 }
67775 #endif
67776 
67777 #ifdef __LITTLE_ENDIAN__
67778 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67779  uint32x4_t __ret;
67780  __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67781  return __ret;
67782 }
67783 #else
67784 __ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67785  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67786  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67787  uint32x4_t __ret;
67788  __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67790  return __ret;
67791 }
67792 __ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67793  uint32x4_t __ret;
67794  __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67795  return __ret;
67796 }
67797 #endif
67798 
67799 #ifdef __LITTLE_ENDIAN__
67800 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67801  int64x2_t __ret;
67802  __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67803  return __ret;
67804 }
67805 #else
67806 __ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67807  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67808  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67809  int64x2_t __ret;
67810  __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67812  return __ret;
67813 }
67814 __ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67815  int64x2_t __ret;
67816  __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67817  return __ret;
67818 }
67819 #endif
67820 
67821 #ifdef __LITTLE_ENDIAN__
67822 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67823  int32x4_t __ret;
67824  __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67825  return __ret;
67826 }
67827 #else
67828 __ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67829  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67830  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67831  int32x4_t __ret;
67832  __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67833  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67834  return __ret;
67835 }
67836 __ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67837  int32x4_t __ret;
67838  __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67839  return __ret;
67840 }
67841 #endif
67842 
67843 #ifdef __LITTLE_ENDIAN__
67844 #define vset_lane_f16(__p0_246, __p1_246, __p2_246) __extension__ ({ \
67845  float16_t __s0_246 = __p0_246; \
67846  float16x4_t __s1_246 = __p1_246; \
67847  float16x4_t __ret_246; \
67848 float16_t __reint_246 = __s0_246; \
67849 float16x4_t __reint1_246 = __s1_246; \
67850 int16x4_t __reint2_246 = vset_lane_s16(*(int16_t *) &__reint_246, *(int16x4_t *) &__reint1_246, __p2_246); \
67851  __ret_246 = *(float16x4_t *) &__reint2_246; \
67852  __ret_246; \
67853 })
67854 #else
67855 #define vset_lane_f16(__p0_247, __p1_247, __p2_247) __extension__ ({ \
67856  float16_t __s0_247 = __p0_247; \
67857  float16x4_t __s1_247 = __p1_247; \
67858  float16x4_t __rev1_247; __rev1_247 = __builtin_shufflevector(__s1_247, __s1_247, 3, 2, 1, 0); \
67859  float16x4_t __ret_247; \
67860 float16_t __reint_247 = __s0_247; \
67861 float16x4_t __reint1_247 = __rev1_247; \
67862 int16x4_t __reint2_247 = __noswap_vset_lane_s16(*(int16_t *) &__reint_247, *(int16x4_t *) &__reint1_247, __p2_247); \
67863  __ret_247 = *(float16x4_t *) &__reint2_247; \
67864  __ret_247 = __builtin_shufflevector(__ret_247, __ret_247, 3, 2, 1, 0); \
67865  __ret_247; \
67866 })
67867 #endif
67868 
67869 #ifdef __LITTLE_ENDIAN__
67870 #define vsetq_lane_f16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
67871  float16_t __s0_248 = __p0_248; \
67872  float16x8_t __s1_248 = __p1_248; \
67873  float16x8_t __ret_248; \
67874 float16_t __reint_248 = __s0_248; \
67875 float16x8_t __reint1_248 = __s1_248; \
67876 int16x8_t __reint2_248 = vsetq_lane_s16(*(int16_t *) &__reint_248, *(int16x8_t *) &__reint1_248, __p2_248); \
67877  __ret_248 = *(float16x8_t *) &__reint2_248; \
67878  __ret_248; \
67879 })
67880 #else
67881 #define vsetq_lane_f16(__p0_249, __p1_249, __p2_249) __extension__ ({ \
67882  float16_t __s0_249 = __p0_249; \
67883  float16x8_t __s1_249 = __p1_249; \
67884  float16x8_t __rev1_249; __rev1_249 = __builtin_shufflevector(__s1_249, __s1_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67885  float16x8_t __ret_249; \
67886 float16_t __reint_249 = __s0_249; \
67887 float16x8_t __reint1_249 = __rev1_249; \
67888 int16x8_t __reint2_249 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_249, *(int16x8_t *) &__reint1_249, __p2_249); \
67889  __ret_249 = *(float16x8_t *) &__reint2_249; \
67890  __ret_249 = __builtin_shufflevector(__ret_249, __ret_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67891  __ret_249; \
67892 })
67893 #endif
67894 
67895 #if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
67896 #ifdef __LITTLE_ENDIAN__
67897 __ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67898  int32_t __ret;
67899  __ret = vqadds_s32(__p0, vqrdmulhs_s32(__p1, __p2));
67900  return __ret;
67901 }
67902 #else
67903 __ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67904  int32_t __ret;
67905  __ret = __noswap_vqadds_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
67906  return __ret;
67907 }
67908 #endif
67909 
67910 #ifdef __LITTLE_ENDIAN__
67911 __ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67912  int16_t __ret;
67913  __ret = vqaddh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
67914  return __ret;
67915 }
67916 #else
67917 __ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67918  int16_t __ret;
67919  __ret = __noswap_vqaddh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
67920  return __ret;
67921 }
67922 #endif
67923 
67924 #ifdef __LITTLE_ENDIAN__
67925 #define vqrdmlahs_lane_s32(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
67926  int32_t __s0_250 = __p0_250; \
67927  int32_t __s1_250 = __p1_250; \
67928  int32x2_t __s2_250 = __p2_250; \
67929  int32_t __ret_250; \
67930  __ret_250 = vqadds_s32(__s0_250, vqrdmulhs_s32(__s1_250, vget_lane_s32(__s2_250, __p3_250))); \
67931  __ret_250; \
67932 })
67933 #else
67934 #define vqrdmlahs_lane_s32(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
67935  int32_t __s0_251 = __p0_251; \
67936  int32_t __s1_251 = __p1_251; \
67937  int32x2_t __s2_251 = __p2_251; \
67938  int32x2_t __rev2_251; __rev2_251 = __builtin_shufflevector(__s2_251, __s2_251, 1, 0); \
67939  int32_t __ret_251; \
67940  __ret_251 = __noswap_vqadds_s32(__s0_251, __noswap_vqrdmulhs_s32(__s1_251, __noswap_vget_lane_s32(__rev2_251, __p3_251))); \
67941  __ret_251; \
67942 })
67943 #endif
67944 
67945 #ifdef __LITTLE_ENDIAN__
67946 #define vqrdmlahh_lane_s16(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
67947  int16_t __s0_252 = __p0_252; \
67948  int16_t __s1_252 = __p1_252; \
67949  int16x4_t __s2_252 = __p2_252; \
67950  int16_t __ret_252; \
67951  __ret_252 = vqaddh_s16(__s0_252, vqrdmulhh_s16(__s1_252, vget_lane_s16(__s2_252, __p3_252))); \
67952  __ret_252; \
67953 })
67954 #else
67955 #define vqrdmlahh_lane_s16(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
67956  int16_t __s0_253 = __p0_253; \
67957  int16_t __s1_253 = __p1_253; \
67958  int16x4_t __s2_253 = __p2_253; \
67959  int16x4_t __rev2_253; __rev2_253 = __builtin_shufflevector(__s2_253, __s2_253, 3, 2, 1, 0); \
67960  int16_t __ret_253; \
67961  __ret_253 = __noswap_vqaddh_s16(__s0_253, __noswap_vqrdmulhh_s16(__s1_253, __noswap_vget_lane_s16(__rev2_253, __p3_253))); \
67962  __ret_253; \
67963 })
67964 #endif
67965 
67966 #ifdef __LITTLE_ENDIAN__
67967 #define vqrdmlahs_laneq_s32(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
67968  int32_t __s0_254 = __p0_254; \
67969  int32_t __s1_254 = __p1_254; \
67970  int32x4_t __s2_254 = __p2_254; \
67971  int32_t __ret_254; \
67972  __ret_254 = vqadds_s32(__s0_254, vqrdmulhs_s32(__s1_254, vgetq_lane_s32(__s2_254, __p3_254))); \
67973  __ret_254; \
67974 })
67975 #else
67976 #define vqrdmlahs_laneq_s32(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
67977  int32_t __s0_255 = __p0_255; \
67978  int32_t __s1_255 = __p1_255; \
67979  int32x4_t __s2_255 = __p2_255; \
67980  int32x4_t __rev2_255; __rev2_255 = __builtin_shufflevector(__s2_255, __s2_255, 3, 2, 1, 0); \
67981  int32_t __ret_255; \
67982  __ret_255 = __noswap_vqadds_s32(__s0_255, __noswap_vqrdmulhs_s32(__s1_255, __noswap_vgetq_lane_s32(__rev2_255, __p3_255))); \
67983  __ret_255; \
67984 })
67985 #endif
67986 
67987 #ifdef __LITTLE_ENDIAN__
67988 #define vqrdmlahh_laneq_s16(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
67989  int16_t __s0_256 = __p0_256; \
67990  int16_t __s1_256 = __p1_256; \
67991  int16x8_t __s2_256 = __p2_256; \
67992  int16_t __ret_256; \
67993  __ret_256 = vqaddh_s16(__s0_256, vqrdmulhh_s16(__s1_256, vgetq_lane_s16(__s2_256, __p3_256))); \
67994  __ret_256; \
67995 })
67996 #else
67997 #define vqrdmlahh_laneq_s16(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
67998  int16_t __s0_257 = __p0_257; \
67999  int16_t __s1_257 = __p1_257; \
68000  int16x8_t __s2_257 = __p2_257; \
68001  int16x8_t __rev2_257; __rev2_257 = __builtin_shufflevector(__s2_257, __s2_257, 7, 6, 5, 4, 3, 2, 1, 0); \
68002  int16_t __ret_257; \
68003  __ret_257 = __noswap_vqaddh_s16(__s0_257, __noswap_vqrdmulhh_s16(__s1_257, __noswap_vgetq_lane_s16(__rev2_257, __p3_257))); \
68004  __ret_257; \
68005 })
68006 #endif
68007 
68008 #ifdef __LITTLE_ENDIAN__
68009 __ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68010  int32_t __ret;
68011  __ret = vqsubs_s32(__p0, vqrdmulhs_s32(__p1, __p2));
68012  return __ret;
68013 }
68014 #else
68015 __ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68016  int32_t __ret;
68017  __ret = __noswap_vqsubs_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
68018  return __ret;
68019 }
68020 #endif
68021 
68022 #ifdef __LITTLE_ENDIAN__
68023 __ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68024  int16_t __ret;
68025  __ret = vqsubh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
68026  return __ret;
68027 }
68028 #else
68029 __ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68030  int16_t __ret;
68031  __ret = __noswap_vqsubh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
68032  return __ret;
68033 }
68034 #endif
68035 
68036 #ifdef __LITTLE_ENDIAN__
68037 #define vqrdmlshs_lane_s32(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
68038  int32_t __s0_258 = __p0_258; \
68039  int32_t __s1_258 = __p1_258; \
68040  int32x2_t __s2_258 = __p2_258; \
68041  int32_t __ret_258; \
68042  __ret_258 = vqsubs_s32(__s0_258, vqrdmulhs_s32(__s1_258, vget_lane_s32(__s2_258, __p3_258))); \
68043  __ret_258; \
68044 })
68045 #else
68046 #define vqrdmlshs_lane_s32(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
68047  int32_t __s0_259 = __p0_259; \
68048  int32_t __s1_259 = __p1_259; \
68049  int32x2_t __s2_259 = __p2_259; \
68050  int32x2_t __rev2_259; __rev2_259 = __builtin_shufflevector(__s2_259, __s2_259, 1, 0); \
68051  int32_t __ret_259; \
68052  __ret_259 = __noswap_vqsubs_s32(__s0_259, __noswap_vqrdmulhs_s32(__s1_259, __noswap_vget_lane_s32(__rev2_259, __p3_259))); \
68053  __ret_259; \
68054 })
68055 #endif
68056 
68057 #ifdef __LITTLE_ENDIAN__
68058 #define vqrdmlshh_lane_s16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
68059  int16_t __s0_260 = __p0_260; \
68060  int16_t __s1_260 = __p1_260; \
68061  int16x4_t __s2_260 = __p2_260; \
68062  int16_t __ret_260; \
68063  __ret_260 = vqsubh_s16(__s0_260, vqrdmulhh_s16(__s1_260, vget_lane_s16(__s2_260, __p3_260))); \
68064  __ret_260; \
68065 })
68066 #else
68067 #define vqrdmlshh_lane_s16(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
68068  int16_t __s0_261 = __p0_261; \
68069  int16_t __s1_261 = __p1_261; \
68070  int16x4_t __s2_261 = __p2_261; \
68071  int16x4_t __rev2_261; __rev2_261 = __builtin_shufflevector(__s2_261, __s2_261, 3, 2, 1, 0); \
68072  int16_t __ret_261; \
68073  __ret_261 = __noswap_vqsubh_s16(__s0_261, __noswap_vqrdmulhh_s16(__s1_261, __noswap_vget_lane_s16(__rev2_261, __p3_261))); \
68074  __ret_261; \
68075 })
68076 #endif
68077 
68078 #ifdef __LITTLE_ENDIAN__
68079 #define vqrdmlshs_laneq_s32(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
68080  int32_t __s0_262 = __p0_262; \
68081  int32_t __s1_262 = __p1_262; \
68082  int32x4_t __s2_262 = __p2_262; \
68083  int32_t __ret_262; \
68084  __ret_262 = vqsubs_s32(__s0_262, vqrdmulhs_s32(__s1_262, vgetq_lane_s32(__s2_262, __p3_262))); \
68085  __ret_262; \
68086 })
68087 #else
68088 #define vqrdmlshs_laneq_s32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
68089  int32_t __s0_263 = __p0_263; \
68090  int32_t __s1_263 = __p1_263; \
68091  int32x4_t __s2_263 = __p2_263; \
68092  int32x4_t __rev2_263; __rev2_263 = __builtin_shufflevector(__s2_263, __s2_263, 3, 2, 1, 0); \
68093  int32_t __ret_263; \
68094  __ret_263 = __noswap_vqsubs_s32(__s0_263, __noswap_vqrdmulhs_s32(__s1_263, __noswap_vgetq_lane_s32(__rev2_263, __p3_263))); \
68095  __ret_263; \
68096 })
68097 #endif
68098 
68099 #ifdef __LITTLE_ENDIAN__
68100 #define vqrdmlshh_laneq_s16(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
68101  int16_t __s0_264 = __p0_264; \
68102  int16_t __s1_264 = __p1_264; \
68103  int16x8_t __s2_264 = __p2_264; \
68104  int16_t __ret_264; \
68105  __ret_264 = vqsubh_s16(__s0_264, vqrdmulhh_s16(__s1_264, vgetq_lane_s16(__s2_264, __p3_264))); \
68106  __ret_264; \
68107 })
68108 #else
68109 #define vqrdmlshh_laneq_s16(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
68110  int16_t __s0_265 = __p0_265; \
68111  int16_t __s1_265 = __p1_265; \
68112  int16x8_t __s2_265 = __p2_265; \
68113  int16x8_t __rev2_265; __rev2_265 = __builtin_shufflevector(__s2_265, __s2_265, 7, 6, 5, 4, 3, 2, 1, 0); \
68114  int16_t __ret_265; \
68115  __ret_265 = __noswap_vqsubh_s16(__s0_265, __noswap_vqrdmulhh_s16(__s1_265, __noswap_vgetq_lane_s16(__rev2_265, __p3_265))); \
68116  __ret_265; \
68117 })
68118 #endif
68119 
68120 #endif
68121 #if defined(__aarch64__)
68122 #ifdef __LITTLE_ENDIAN__
68123 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68124  uint16x8_t __ret;
68125  __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
68126  return __ret;
68127 }
68128 #else
68129 __ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68130  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68131  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68132  uint16x8_t __ret;
68134  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68135  return __ret;
68136 }
68137 #endif
68138 
68139 #ifdef __LITTLE_ENDIAN__
68140 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68141  uint64x2_t __ret;
68142  __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
68143  return __ret;
68144 }
68145 #else
68146 __ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68147  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68148  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68149  uint64x2_t __ret;
68151  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68152  return __ret;
68153 }
68154 #endif
68155 
68156 #ifdef __LITTLE_ENDIAN__
68157 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68158  uint32x4_t __ret;
68159  __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
68160  return __ret;
68161 }
68162 #else
68163 __ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68164  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68165  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68166  uint32x4_t __ret;
68168  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68169  return __ret;
68170 }
68171 #endif
68172 
68173 #ifdef __LITTLE_ENDIAN__
68174 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68175  int16x8_t __ret;
68176  __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
68177  return __ret;
68178 }
68179 #else
68180 __ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68181  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68182  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68183  int16x8_t __ret;
68185  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68186  return __ret;
68187 }
68188 #endif
68189 
68190 #ifdef __LITTLE_ENDIAN__
68191 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68192  int64x2_t __ret;
68193  __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
68194  return __ret;
68195 }
68196 #else
68197 __ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68198  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68199  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68200  int64x2_t __ret;
68202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68203  return __ret;
68204 }
68205 #endif
68206 
68207 #ifdef __LITTLE_ENDIAN__
68208 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68209  int32x4_t __ret;
68210  __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
68211  return __ret;
68212 }
68213 #else
68214 __ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68215  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68216  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68217  int32x4_t __ret;
68219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68220  return __ret;
68221 }
68222 #endif
68223 
68224 #ifdef __LITTLE_ENDIAN__
68225 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68226  uint16x8_t __ret;
68227  __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
68228  return __ret;
68229 }
68230 #else
68231 __ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68232  uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68233  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68234  uint16x8_t __ret;
68235  __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
68236  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68237  return __ret;
68238 }
68239 #endif
68240 
68241 #ifdef __LITTLE_ENDIAN__
68242 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68243  uint64x2_t __ret;
68244  __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
68245  return __ret;
68246 }
68247 #else
68248 __ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68249  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68250  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68251  uint64x2_t __ret;
68252  __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
68253  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68254  return __ret;
68255 }
68256 #endif
68257 
68258 #ifdef __LITTLE_ENDIAN__
68259 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68260  uint32x4_t __ret;
68261  __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
68262  return __ret;
68263 }
68264 #else
68265 __ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68266  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68267  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68268  uint32x4_t __ret;
68269  __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
68270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68271  return __ret;
68272 }
68273 #endif
68274 
68275 #ifdef __LITTLE_ENDIAN__
68276 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68277  int16x8_t __ret;
68278  __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
68279  return __ret;
68280 }
68281 #else
68282 __ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68283  int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68284  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68285  int16x8_t __ret;
68286  __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
68287  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68288  return __ret;
68289 }
68290 #endif
68291 
68292 #ifdef __LITTLE_ENDIAN__
68293 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68294  int64x2_t __ret;
68295  __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
68296  return __ret;
68297 }
68298 #else
68299 __ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68300  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68301  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68302  int64x2_t __ret;
68303  __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
68304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68305  return __ret;
68306 }
68307 #endif
68308 
68309 #ifdef __LITTLE_ENDIAN__
68310 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68311  int32x4_t __ret;
68312  __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
68313  return __ret;
68314 }
68315 #else
68316 __ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68317  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68318  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68319  int32x4_t __ret;
68320  __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
68321  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68322  return __ret;
68323 }
68324 #endif
68325 
68326 #ifdef __LITTLE_ENDIAN__
68327 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68328  uint16x8_t __ret;
68329  __ret = __p0 + vmovl_high_u8(__p1);
68330  return __ret;
68331 }
68332 #else
68333 __ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68334  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68335  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68336  uint16x8_t __ret;
68337  __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
68338  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68339  return __ret;
68340 }
68341 #endif
68342 
68343 #ifdef __LITTLE_ENDIAN__
68344 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68345  uint64x2_t __ret;
68346  __ret = __p0 + vmovl_high_u32(__p1);
68347  return __ret;
68348 }
68349 #else
68350 __ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68351  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68352  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68353  uint64x2_t __ret;
68354  __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
68355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68356  return __ret;
68357 }
68358 #endif
68359 
68360 #ifdef __LITTLE_ENDIAN__
68361 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68362  uint32x4_t __ret;
68363  __ret = __p0 + vmovl_high_u16(__p1);
68364  return __ret;
68365 }
68366 #else
68367 __ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68368  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68369  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68370  uint32x4_t __ret;
68371  __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
68372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68373  return __ret;
68374 }
68375 #endif
68376 
68377 #ifdef __LITTLE_ENDIAN__
68378 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68379  int16x8_t __ret;
68380  __ret = __p0 + vmovl_high_s8(__p1);
68381  return __ret;
68382 }
68383 #else
68384 __ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68385  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68386  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68387  int16x8_t __ret;
68388  __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
68389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68390  return __ret;
68391 }
68392 #endif
68393 
68394 #ifdef __LITTLE_ENDIAN__
68395 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68396  int64x2_t __ret;
68397  __ret = __p0 + vmovl_high_s32(__p1);
68398  return __ret;
68399 }
68400 #else
68401 __ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68402  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68403  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68404  int64x2_t __ret;
68405  __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
68406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68407  return __ret;
68408 }
68409 #endif
68410 
68411 #ifdef __LITTLE_ENDIAN__
68412 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68413  int32x4_t __ret;
68414  __ret = __p0 + vmovl_high_s16(__p1);
68415  return __ret;
68416 }
68417 #else
68418 __ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68419  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68420  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68421  int32x4_t __ret;
68422  __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
68423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68424  return __ret;
68425 }
68426 #endif
68427 
68428 #ifdef __LITTLE_ENDIAN__
68429 #define vcopyq_lane_p64(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
68430  poly64x2_t __s0_266 = __p0_266; \
68431  poly64x1_t __s2_266 = __p2_266; \
68432  poly64x2_t __ret_266; \
68433  __ret_266 = vsetq_lane_p64(vget_lane_p64(__s2_266, __p3_266), __s0_266, __p1_266); \
68434  __ret_266; \
68435 })
68436 #else
68437 #define vcopyq_lane_p64(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
68438  poly64x2_t __s0_267 = __p0_267; \
68439  poly64x1_t __s2_267 = __p2_267; \
68440  poly64x2_t __rev0_267; __rev0_267 = __builtin_shufflevector(__s0_267, __s0_267, 1, 0); \
68441  poly64x2_t __ret_267; \
68442  __ret_267 = __noswap_vsetq_lane_p64(__noswap_vget_lane_p64(__s2_267, __p3_267), __rev0_267, __p1_267); \
68443  __ret_267 = __builtin_shufflevector(__ret_267, __ret_267, 1, 0); \
68444  __ret_267; \
68445 })
68446 #endif
68447 
68448 #ifdef __LITTLE_ENDIAN__
68449 #define vcopyq_lane_f64(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
68450  float64x2_t __s0_268 = __p0_268; \
68451  float64x1_t __s2_268 = __p2_268; \
68452  float64x2_t __ret_268; \
68453  __ret_268 = vsetq_lane_f64(vget_lane_f64(__s2_268, __p3_268), __s0_268, __p1_268); \
68454  __ret_268; \
68455 })
68456 #else
68457 #define vcopyq_lane_f64(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
68458  float64x2_t __s0_269 = __p0_269; \
68459  float64x1_t __s2_269 = __p2_269; \
68460  float64x2_t __rev0_269; __rev0_269 = __builtin_shufflevector(__s0_269, __s0_269, 1, 0); \
68461  float64x2_t __ret_269; \
68462  __ret_269 = __noswap_vsetq_lane_f64(__noswap_vget_lane_f64(__s2_269, __p3_269), __rev0_269, __p1_269); \
68463  __ret_269 = __builtin_shufflevector(__ret_269, __ret_269, 1, 0); \
68464  __ret_269; \
68465 })
68466 #endif
68467 
68468 #ifdef __LITTLE_ENDIAN__
68469 #define vcopy_lane_p64(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
68470  poly64x1_t __s0_270 = __p0_270; \
68471  poly64x1_t __s2_270 = __p2_270; \
68472  poly64x1_t __ret_270; \
68473  __ret_270 = vset_lane_p64(vget_lane_p64(__s2_270, __p3_270), __s0_270, __p1_270); \
68474  __ret_270; \
68475 })
68476 #else
68477 #define vcopy_lane_p64(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
68478  poly64x1_t __s0_271 = __p0_271; \
68479  poly64x1_t __s2_271 = __p2_271; \
68480  poly64x1_t __ret_271; \
68481  __ret_271 = __noswap_vset_lane_p64(__noswap_vget_lane_p64(__s2_271, __p3_271), __s0_271, __p1_271); \
68482  __ret_271; \
68483 })
68484 #endif
68485 
68486 #ifdef __LITTLE_ENDIAN__
68487 #define vcopy_lane_f64(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
68488  float64x1_t __s0_272 = __p0_272; \
68489  float64x1_t __s2_272 = __p2_272; \
68490  float64x1_t __ret_272; \
68491  __ret_272 = vset_lane_f64(vget_lane_f64(__s2_272, __p3_272), __s0_272, __p1_272); \
68492  __ret_272; \
68493 })
68494 #else
68495 #define vcopy_lane_f64(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
68496  float64x1_t __s0_273 = __p0_273; \
68497  float64x1_t __s2_273 = __p2_273; \
68498  float64x1_t __ret_273; \
68499  __ret_273 = __noswap_vset_lane_f64(__noswap_vget_lane_f64(__s2_273, __p3_273), __s0_273, __p1_273); \
68500  __ret_273; \
68501 })
68502 #endif
68503 
68504 #ifdef __LITTLE_ENDIAN__
68505 #define vcopyq_laneq_p64(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
68506  poly64x2_t __s0_274 = __p0_274; \
68507  poly64x2_t __s2_274 = __p2_274; \
68508  poly64x2_t __ret_274; \
68509  __ret_274 = vsetq_lane_p64(vgetq_lane_p64(__s2_274, __p3_274), __s0_274, __p1_274); \
68510  __ret_274; \
68511 })
68512 #else
68513 #define vcopyq_laneq_p64(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
68514  poly64x2_t __s0_275 = __p0_275; \
68515  poly64x2_t __s2_275 = __p2_275; \
68516  poly64x2_t __rev0_275; __rev0_275 = __builtin_shufflevector(__s0_275, __s0_275, 1, 0); \
68517  poly64x2_t __rev2_275; __rev2_275 = __builtin_shufflevector(__s2_275, __s2_275, 1, 0); \
68518  poly64x2_t __ret_275; \
68519  __ret_275 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_275, __p3_275), __rev0_275, __p1_275); \
68520  __ret_275 = __builtin_shufflevector(__ret_275, __ret_275, 1, 0); \
68521  __ret_275; \
68522 })
68523 #endif
68524 
68525 #ifdef __LITTLE_ENDIAN__
68526 #define vcopyq_laneq_f64(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
68527  float64x2_t __s0_276 = __p0_276; \
68528  float64x2_t __s2_276 = __p2_276; \
68529  float64x2_t __ret_276; \
68530  __ret_276 = vsetq_lane_f64(vgetq_lane_f64(__s2_276, __p3_276), __s0_276, __p1_276); \
68531  __ret_276; \
68532 })
68533 #else
68534 #define vcopyq_laneq_f64(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
68535  float64x2_t __s0_277 = __p0_277; \
68536  float64x2_t __s2_277 = __p2_277; \
68537  float64x2_t __rev0_277; __rev0_277 = __builtin_shufflevector(__s0_277, __s0_277, 1, 0); \
68538  float64x2_t __rev2_277; __rev2_277 = __builtin_shufflevector(__s2_277, __s2_277, 1, 0); \
68539  float64x2_t __ret_277; \
68540  __ret_277 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_277, __p3_277), __rev0_277, __p1_277); \
68541  __ret_277 = __builtin_shufflevector(__ret_277, __ret_277, 1, 0); \
68542  __ret_277; \
68543 })
68544 #endif
68545 
68546 #ifdef __LITTLE_ENDIAN__
68547 #define vcopy_laneq_p64(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
68548  poly64x1_t __s0_278 = __p0_278; \
68549  poly64x2_t __s2_278 = __p2_278; \
68550  poly64x1_t __ret_278; \
68551  __ret_278 = vset_lane_p64(vgetq_lane_p64(__s2_278, __p3_278), __s0_278, __p1_278); \
68552  __ret_278; \
68553 })
68554 #else
68555 #define vcopy_laneq_p64(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
68556  poly64x1_t __s0_279 = __p0_279; \
68557  poly64x2_t __s2_279 = __p2_279; \
68558  poly64x2_t __rev2_279; __rev2_279 = __builtin_shufflevector(__s2_279, __s2_279, 1, 0); \
68559  poly64x1_t __ret_279; \
68560  __ret_279 = __noswap_vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_279, __p3_279), __s0_279, __p1_279); \
68561  __ret_279; \
68562 })
68563 #endif
68564 
68565 #ifdef __LITTLE_ENDIAN__
68566 #define vcopy_laneq_f64(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
68567  float64x1_t __s0_280 = __p0_280; \
68568  float64x2_t __s2_280 = __p2_280; \
68569  float64x1_t __ret_280; \
68570  __ret_280 = vset_lane_f64(vgetq_lane_f64(__s2_280, __p3_280), __s0_280, __p1_280); \
68571  __ret_280; \
68572 })
68573 #else
68574 #define vcopy_laneq_f64(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
68575  float64x1_t __s0_281 = __p0_281; \
68576  float64x2_t __s2_281 = __p2_281; \
68577  float64x2_t __rev2_281; __rev2_281 = __builtin_shufflevector(__s2_281, __s2_281, 1, 0); \
68578  float64x1_t __ret_281; \
68579  __ret_281 = __noswap_vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_281, __p3_281), __s0_281, __p1_281); \
68580  __ret_281; \
68581 })
68582 #endif
68583 
68584 #ifdef __LITTLE_ENDIAN__
68585 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68586  uint16x8_t __ret;
68587  __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68588  return __ret;
68589 }
68590 #else
68591 __ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68592  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68593  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68594  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68595  uint16x8_t __ret;
68596  __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68597  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68598  return __ret;
68599 }
68600 #endif
68601 
68602 #ifdef __LITTLE_ENDIAN__
68603 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68604  uint64x2_t __ret;
68605  __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68606  return __ret;
68607 }
68608 #else
68609 __ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68610  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68611  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68612  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68613  uint64x2_t __ret;
68614  __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68616  return __ret;
68617 }
68618 #endif
68619 
68620 #ifdef __LITTLE_ENDIAN__
68621 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68622  uint32x4_t __ret;
68623  __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68624  return __ret;
68625 }
68626 #else
68627 __ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68628  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68629  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68630  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68631  uint32x4_t __ret;
68632  __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68633  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68634  return __ret;
68635 }
68636 #endif
68637 
68638 #ifdef __LITTLE_ENDIAN__
68639 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68640  int16x8_t __ret;
68641  __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68642  return __ret;
68643 }
68644 #else
68645 __ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68646  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68647  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68648  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68649  int16x8_t __ret;
68650  __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68651  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68652  return __ret;
68653 }
68654 #endif
68655 
68656 #ifdef __LITTLE_ENDIAN__
68657 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68658  int64x2_t __ret;
68659  __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68660  return __ret;
68661 }
68662 #else
68663 __ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68664  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68665  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68666  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68667  int64x2_t __ret;
68668  __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68669  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68670  return __ret;
68671 }
68672 #endif
68673 
68674 #ifdef __LITTLE_ENDIAN__
68675 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68676  int32x4_t __ret;
68677  __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68678  return __ret;
68679 }
68680 #else
68681 __ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68682  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68683  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68684  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68685  int32x4_t __ret;
68686  __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68687  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68688  return __ret;
68689 }
68690 #endif
68691 
68692 #ifdef __LITTLE_ENDIAN__
68693 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68694  uint64x2_t __ret;
68695  __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
68696  return __ret;
68697 }
68698 #else
68699 __ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68700  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68701  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68702  uint64x2_t __ret;
68703  __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68704  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68705  return __ret;
68706 }
68707 #endif
68708 
68709 #ifdef __LITTLE_ENDIAN__
68710 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68711  uint32x4_t __ret;
68712  __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
68713  return __ret;
68714 }
68715 #else
68716 __ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68717  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68718  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68719  uint32x4_t __ret;
68720  __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68721  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68722  return __ret;
68723 }
68724 #endif
68725 
68726 #ifdef __LITTLE_ENDIAN__
68727 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68728  int64x2_t __ret;
68729  __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
68730  return __ret;
68731 }
68732 #else
68733 __ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68734  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68735  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68736  int64x2_t __ret;
68737  __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68738  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68739  return __ret;
68740 }
68741 #endif
68742 
68743 #ifdef __LITTLE_ENDIAN__
68744 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68745  int32x4_t __ret;
68746  __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
68747  return __ret;
68748 }
68749 #else
68750 __ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68751  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68752  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68753  int32x4_t __ret;
68754  __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68755  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68756  return __ret;
68757 }
68758 #endif
68759 
68760 #ifdef __LITTLE_ENDIAN__
68761 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68762  uint16x8_t __ret;
68763  __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68764  return __ret;
68765 }
68766 #else
68767 __ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68768  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68769  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68770  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68771  uint16x8_t __ret;
68772  __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68774  return __ret;
68775 }
68776 #endif
68777 
68778 #ifdef __LITTLE_ENDIAN__
68779 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68780  uint64x2_t __ret;
68781  __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68782  return __ret;
68783 }
68784 #else
68785 __ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68786  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68787  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68788  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68789  uint64x2_t __ret;
68790  __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68791  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68792  return __ret;
68793 }
68794 #endif
68795 
68796 #ifdef __LITTLE_ENDIAN__
68797 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68798  uint32x4_t __ret;
68799  __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68800  return __ret;
68801 }
68802 #else
68803 __ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68804  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68805  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68806  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68807  uint32x4_t __ret;
68808  __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68809  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68810  return __ret;
68811 }
68812 #endif
68813 
68814 #ifdef __LITTLE_ENDIAN__
68815 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68816  int16x8_t __ret;
68817  __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68818  return __ret;
68819 }
68820 #else
68821 __ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68822  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68823  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68824  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68825  int16x8_t __ret;
68826  __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68827  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68828  return __ret;
68829 }
68830 #endif
68831 
68832 #ifdef __LITTLE_ENDIAN__
68833 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68834  int64x2_t __ret;
68835  __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68836  return __ret;
68837 }
68838 #else
68839 __ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68840  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68841  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68842  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68843  int64x2_t __ret;
68844  __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68846  return __ret;
68847 }
68848 #endif
68849 
68850 #ifdef __LITTLE_ENDIAN__
68851 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68852  int32x4_t __ret;
68853  __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68854  return __ret;
68855 }
68856 #else
68857 __ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68858  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68859  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68860  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68861  int32x4_t __ret;
68862  __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68863  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68864  return __ret;
68865 }
68866 #endif
68867 
68868 #ifdef __LITTLE_ENDIAN__
68869 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68870  uint64x2_t __ret;
68871  __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
68872  return __ret;
68873 }
68874 #else
68875 __ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68876  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68877  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68878  uint64x2_t __ret;
68879  __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68880  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68881  return __ret;
68882 }
68883 #endif
68884 
68885 #ifdef __LITTLE_ENDIAN__
68886 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68887  uint32x4_t __ret;
68888  __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
68889  return __ret;
68890 }
68891 #else
68892 __ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68893  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68894  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68895  uint32x4_t __ret;
68896  __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68897  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68898  return __ret;
68899 }
68900 #endif
68901 
68902 #ifdef __LITTLE_ENDIAN__
68903 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68904  int64x2_t __ret;
68905  __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
68906  return __ret;
68907 }
68908 #else
68909 __ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68910  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68911  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68912  int64x2_t __ret;
68913  __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68914  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68915  return __ret;
68916 }
68917 #endif
68918 
68919 #ifdef __LITTLE_ENDIAN__
68920 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68921  int32x4_t __ret;
68922  __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
68923  return __ret;
68924 }
68925 #else
68926 __ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68927  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68928  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68929  int32x4_t __ret;
68930  __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68931  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68932  return __ret;
68933 }
68934 #endif
68935 
68936 #ifdef __LITTLE_ENDIAN__
68937 #define vmulx_lane_f64(__p0_282, __p1_282, __p2_282) __extension__ ({ \
68938  float64x1_t __s0_282 = __p0_282; \
68939  float64x1_t __s1_282 = __p1_282; \
68940  float64x1_t __ret_282; \
68941  float64_t __x_282 = vget_lane_f64(__s0_282, 0); \
68942  float64_t __y_282 = vget_lane_f64(__s1_282, __p2_282); \
68943  float64_t __z_282 = vmulxd_f64(__x_282, __y_282); \
68944  __ret_282 = vset_lane_f64(__z_282, __s0_282, __p2_282); \
68945  __ret_282; \
68946 })
68947 #else
68948 #define vmulx_lane_f64(__p0_283, __p1_283, __p2_283) __extension__ ({ \
68949  float64x1_t __s0_283 = __p0_283; \
68950  float64x1_t __s1_283 = __p1_283; \
68951  float64x1_t __ret_283; \
68952  float64_t __x_283 = __noswap_vget_lane_f64(__s0_283, 0); \
68953  float64_t __y_283 = __noswap_vget_lane_f64(__s1_283, __p2_283); \
68954  float64_t __z_283 = __noswap_vmulxd_f64(__x_283, __y_283); \
68955  __ret_283 = __noswap_vset_lane_f64(__z_283, __s0_283, __p2_283); \
68956  __ret_283; \
68957 })
68958 #endif
68959 
68960 #ifdef __LITTLE_ENDIAN__
68961 #define vmulx_laneq_f64(__p0_284, __p1_284, __p2_284) __extension__ ({ \
68962  float64x1_t __s0_284 = __p0_284; \
68963  float64x2_t __s1_284 = __p1_284; \
68964  float64x1_t __ret_284; \
68965  float64_t __x_284 = vget_lane_f64(__s0_284, 0); \
68966  float64_t __y_284 = vgetq_lane_f64(__s1_284, __p2_284); \
68967  float64_t __z_284 = vmulxd_f64(__x_284, __y_284); \
68968  __ret_284 = vset_lane_f64(__z_284, __s0_284, 0); \
68969  __ret_284; \
68970 })
68971 #else
68972 #define vmulx_laneq_f64(__p0_285, __p1_285, __p2_285) __extension__ ({ \
68973  float64x1_t __s0_285 = __p0_285; \
68974  float64x2_t __s1_285 = __p1_285; \
68975  float64x2_t __rev1_285; __rev1_285 = __builtin_shufflevector(__s1_285, __s1_285, 1, 0); \
68976  float64x1_t __ret_285; \
68977  float64_t __x_285 = __noswap_vget_lane_f64(__s0_285, 0); \
68978  float64_t __y_285 = __noswap_vgetq_lane_f64(__rev1_285, __p2_285); \
68979  float64_t __z_285 = __noswap_vmulxd_f64(__x_285, __y_285); \
68980  __ret_285 = __noswap_vset_lane_f64(__z_285, __s0_285, 0); \
68981  __ret_285; \
68982 })
68983 #endif
68984 
68985 #endif
68986 #ifdef __LITTLE_ENDIAN__
68987 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68988  uint16x8_t __ret;
68989  __ret = __p0 + vabdl_u8(__p1, __p2);
68990  return __ret;
68991 }
68992 #else
68993 __ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68994  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68995  uint8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68996  uint8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68997  uint16x8_t __ret;
68998  __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
68999  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69000  return __ret;
69001 }
69002 __ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
69003  uint16x8_t __ret;
69004  __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
69005  return __ret;
69006 }
69007 #endif
69008 
69009 #ifdef __LITTLE_ENDIAN__
69010 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69011  uint64x2_t __ret;
69012  __ret = __p0 + vabdl_u32(__p1, __p2);
69013  return __ret;
69014 }
69015 #else
69016 __ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69017  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69018  uint32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69019  uint32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69020  uint64x2_t __ret;
69021  __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
69022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69023  return __ret;
69024 }
69025 __ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69026  uint64x2_t __ret;
69027  __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
69028  return __ret;
69029 }
69030 #endif
69031 
69032 #ifdef __LITTLE_ENDIAN__
69033 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69034  uint32x4_t __ret;
69035  __ret = __p0 + vabdl_u16(__p1, __p2);
69036  return __ret;
69037 }
69038 #else
69039 __ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69040  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69041  uint16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69042  uint16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69043  uint32x4_t __ret;
69044  __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
69045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69046  return __ret;
69047 }
69048 __ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69049  uint32x4_t __ret;
69050  __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
69051  return __ret;
69052 }
69053 #endif
69054 
69055 #ifdef __LITTLE_ENDIAN__
69056 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69057  int16x8_t __ret;
69058  __ret = __p0 + vabdl_s8(__p1, __p2);
69059  return __ret;
69060 }
69061 #else
69062 __ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69063  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69064  int8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69065  int8x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69066  int16x8_t __ret;
69067  __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
69068  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69069  return __ret;
69070 }
69071 __ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69072  int16x8_t __ret;
69073  __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
69074  return __ret;
69075 }
69076 #endif
69077 
69078 #ifdef __LITTLE_ENDIAN__
69079 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69080  int64x2_t __ret;
69081  __ret = __p0 + vabdl_s32(__p1, __p2);
69082  return __ret;
69083 }
69084 #else
69085 __ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69086  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69087  int32x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69088  int32x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69089  int64x2_t __ret;
69090  __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
69091  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69092  return __ret;
69093 }
69094 __ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69095  int64x2_t __ret;
69096  __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
69097  return __ret;
69098 }
69099 #endif
69100 
69101 #ifdef __LITTLE_ENDIAN__
69102 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69103  int32x4_t __ret;
69104  __ret = __p0 + vabdl_s16(__p1, __p2);
69105  return __ret;
69106 }
69107 #else
69108 __ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69109  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69110  int16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69111  int16x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69112  int32x4_t __ret;
69113  __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
69114  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69115  return __ret;
69116 }
69117 __ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69118  int32x4_t __ret;
69119  __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
69120  return __ret;
69121 }
69122 #endif
69123 
69124 #if defined(__aarch64__)
69125 #ifdef __LITTLE_ENDIAN__
69126 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69127  uint16x8_t __ret;
69128  __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
69129  return __ret;
69130 }
69131 #else
69132 __ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69133  uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69134  uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69135  uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69136  uint16x8_t __ret;
69137  __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
69138  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69139  return __ret;
69140 }
69141 #endif
69142 
69143 #ifdef __LITTLE_ENDIAN__
69144 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69145  uint64x2_t __ret;
69146  __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
69147  return __ret;
69148 }
69149 #else
69150 __ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69151  uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69152  uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69153  uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69154  uint64x2_t __ret;
69155  __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
69156  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69157  return __ret;
69158 }
69159 #endif
69160 
69161 #ifdef __LITTLE_ENDIAN__
69162 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69163  uint32x4_t __ret;
69164  __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
69165  return __ret;
69166 }
69167 #else
69168 __ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69169  uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69170  uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69171  uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69172  uint32x4_t __ret;
69173  __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
69174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69175  return __ret;
69176 }
69177 #endif
69178 
69179 #ifdef __LITTLE_ENDIAN__
69180 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69181  int16x8_t __ret;
69182  __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
69183  return __ret;
69184 }
69185 #else
69186 __ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69187  int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69188  int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69189  int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69190  int16x8_t __ret;
69191  __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
69192  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69193  return __ret;
69194 }
69195 #endif
69196 
69197 #ifdef __LITTLE_ENDIAN__
69198 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69199  int64x2_t __ret;
69200  __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
69201  return __ret;
69202 }
69203 #else
69204 __ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69205  int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69206  int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69207  int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69208  int64x2_t __ret;
69209  __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
69210  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69211  return __ret;
69212 }
69213 #endif
69214 
69215 #ifdef __LITTLE_ENDIAN__
69216 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69217  int32x4_t __ret;
69218  __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
69219  return __ret;
69220 }
69221 #else
69222 __ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69223  int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69224  int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69225  int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69226  int32x4_t __ret;
69227  __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
69228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69229  return __ret;
69230 }
69231 #endif
69232 
69233 #endif
69234 
69235 #undef __ai
69236 
69237 #endif /* __ARM_NEON_H */
__ai uint16x4_t vmov_n_u16(uint16_t __p0)
Definition: arm_neon.h:13752
__ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:552
__ai uint8x8_t vcnt_u8(uint8x8_t __p0)
Definition: arm_neon.h:3929
Definition: arm_neon.h:109
__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:22795
__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1)
Definition: arm_neon.h:14608
__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:27645
__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0)
Definition: arm_neon.h:32441
__ai poly16x4_t vrev32_p16(poly16x4_t __p0)
Definition: arm_neon.h:20266
__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0)
Definition: arm_neon.h:31839
__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:15675
__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:7404
__ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:67537
__ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:17713
__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17599
__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0)
Definition: arm_neon.h:30747
__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:5619
__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18078
__ai int16x8_t vqabsq_s16(int16x8_t __p0)
Definition: arm_neon.h:16737
__ai int16x4_t vneg_s16(int16x4_t __p0)
Definition: arm_neon.h:15410
__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:20985
__ai uint16x4_t vget_high_u16(uint16x8_t __p0)
Definition: arm_neon.h:6297
struct int16x8x4_t int16x8x4_t
int64x2_t val[2]
Definition: arm_neon.h:110
__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2)
Definition: arm_neon.h:13360
__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0)
Definition: arm_neon.h:31741
__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:19946
uint64x2_t val[2]
Definition: arm_neon.h:142
__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:16467
__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0)
Definition: arm_neon.h:30761
__ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0)
Definition: arm_neon.h:6346
__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2)
Definition: arm_neon.h:1865
#define vshll_n_s32(__p0, __p1)
Definition: arm_neon.h:23425
__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2)
Definition: arm_neon.h:12924
__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17621
__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0)
Definition: arm_neon.h:4667
__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:2421
poly16x8_t val[4]
Definition: arm_neon.h:416
struct int16x8x2_t int16x8x2_t
Definition: arm_neon.h:287
__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0)
Definition: arm_neon.h:31685
__ai uint8x16_t vrev16q_u8(uint8x16_t __p0)
Definition: arm_neon.h:20186
__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:67298
__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0)
Definition: arm_neon.h:31265
struct float32x2x4_t float32x2x4_t
__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:14720
__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:16925
__ai int16x8_t vmovq_n_s16(int16_t __p0)
Definition: arm_neon.h:13693
__ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0)
Definition: arm_neon.h:6367
__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28592
__ai uint32x4_t vrev64q_u32(uint32x4_t __p0)
Definition: arm_neon.h:20522
__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:27809
__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:11798
__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0)
Definition: arm_neon.h:31181
__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:14229
__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:29233
__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0)
Definition: arm_neon.h:31671
__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:16535
__ai uint64x1_t vcreate_u64(uint64_t __p0)
Definition: arm_neon.h:4257
Definition: arm_neon.h:415
__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:3067
__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0)
Definition: arm_neon.h:30215
Definition: arm_neon.h:377
__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:439
__ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:4187
__ai uint16x4_t vrev32_u16(uint16x4_t __p0)
Definition: arm_neon.h:20394
__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:15992
__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0)
Definition: arm_neon.h:30593
__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:1597
__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:11747
__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0)
Definition: arm_neon.h:31391
__ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:69025
__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:22846
__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:15460
__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2)
Definition: arm_neon.h:1919
__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:14263
__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2)
Definition: arm_neon.h:28549
__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28507
__ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17437
__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0)
Definition: arm_neon.h:31895
__ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0)
Definition: arm_neon.h:14075
__ai poly8x8_t vrev32_p8(poly8x8_t __p0)
Definition: arm_neon.h:20250
__ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:67836
__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:16886
float16x8_t val[3]
Definition: arm_neon.h:266
__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:15053
__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:18965
__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:11849
__ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:67629
__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:2863
__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:69085
__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:27594
__ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0)
Definition: arm_neon.h:17818
__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:27826
__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:3152
__ai uint32x2_t vqmovun_s64(int64x2_t __p0)
Definition: arm_neon.h:17874
__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:15477
__ai int32x4_t vrev64q_s32(int32x4_t __p0)
Definition: arm_neon.h:20586
__ai int32x2_t vdup_n_s32(int32_t __p0)
Definition: arm_neon.h:5391
__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0)
Definition: arm_neon.h:30607
int8x16_t val[3]
Definition: arm_neon.h:202
__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:7302
__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:29402
int8x8_t val[4]
Definition: arm_neon.h:314
__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0)
Definition: arm_neon.h:33029
__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:29114
__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:19638
__ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:17692
uint32x4_t val[2]
Definition: arm_neon.h:134
__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2)
Definition: arm_neon.h:2045
__ai int8x16_t vmvnq_s8(int8x16_t __p0)
Definition: arm_neon.h:15154
Definition: arm_neon.h:341
__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:2370
__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:2880
Definition: arm_neon.h:113
__ai poly8x16_t vrev16q_p8(poly8x16_t __p0)
Definition: arm_neon.h:20170
__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:1498
__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0)
Definition: arm_neon.h:32427
__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:7132
__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:2659
__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2217
__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:29288
__ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:67238
__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:21087
__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0)
Definition: arm_neon.h:32469
__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:27992
__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:16835
__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0)
Definition: arm_neon.h:31069
Definition: arm_neon.h:407
__ai int16x8_t vmovl_s8(int8x8_t __p0)
Definition: arm_neon.h:13921
__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:21169
struct int32x4x2_t int32x4x2_t
__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:28712
__ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:17234
__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2)
Definition: arm_neon.h:28614
__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2)
Definition: arm_neon.h:66612
struct uint32x2x2_t uint32x2x2_t
__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0)
Definition: arm_neon.h:33309
__ai int16x8_t vclsq_s16(int16x8_t __p0)
Definition: arm_neon.h:3371
__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0)
Definition: arm_neon.h:16277
__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28270
__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:22100
__ai poly8x8_t vget_low_p8(poly8x16_t __p0)
Definition: arm_neon.h:6942
__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2)
Definition: arm_neon.h:13394
__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:12817
__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:7353
__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18815
__ai uint16x4_t vqmovn_u32(uint32x4_t __p0)
Definition: arm_neon.h:17727
__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:27744
__ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0)
Definition: arm_neon.h:4461
__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:1365
__ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0)
Definition: arm_neon.h:13886
__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0)
Definition: arm_neon.h:30117
__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:11866
__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0)
Definition: arm_neon.h:29949
__ai int32x2_t vcreate_s32(uint64_t __p0)
Definition: arm_neon.h:4327
__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:1699
__ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:27900
__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0)
Definition: arm_neon.h:32567
__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0)
Definition: arm_neon.h:32259
__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0)
Definition: arm_neon.h:31251
#define vshll_n_s16(__p0, __p1)
Definition: arm_neon.h:23449
__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0)
Definition: arm_neon.h:31335
__ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:66902
__ai int8x8_t vneg_s8(int8x8_t __p0)
Definition: arm_neon.h:15362
__ai uint16x4_t vmvn_u16(uint16x4_t __p0)
Definition: arm_neon.h:15234
struct poly8x8x3_t poly8x8x3_t
__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:67130
__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2)
Definition: arm_neon.h:12698
__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:18384
__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1)
Definition: arm_neon.h:14990
__ai uint32x4_t vmovq_n_u32(uint32_t __p0)
Definition: arm_neon.h:13571
__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0)
Definition: arm_neon.h:32245
__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:19924
struct poly16x8x2_t poly16x8x2_t
__ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:1195
__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:15627
__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0)
Definition: arm_neon.h:31055
__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:1331
__ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:1239
__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2)
Definition: arm_neon.h:66630
__ai uint32x2_t vmov_n_u32(uint32_t __p0)
Definition: arm_neon.h:13723
__ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:14870
float16x4_t val[4]
Definition: arm_neon.h:378
__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0)
Definition: arm_neon.h:32301
__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:67462
__ai int32x2_t vcvt_s32_f32(float32x2_t __p0)
Definition: arm_neon.h:4635
__ai float32x2_t vcreate_f32(uint64_t __p0)
Definition: arm_neon.h:4299
Definition: arm_neon.h:141
__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:18866
__ai int16x4_t vpaddl_s8(int8x8_t __p0)
Definition: arm_neon.h:16420
__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:950
__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:29182
#define __noswap_vshll_n_u16(__p0, __p1)
Definition: arm_neon.h:23385
__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:29063
__ai uint32x2_t vrecpe_u32(uint32x2_t __p0)
Definition: arm_neon.h:20088
__ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0)
Definition: arm_neon.h:6388
__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0)
Definition: arm_neon.h:32805
poly8x16_t val[3]
Definition: arm_neon.h:292
__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:16995
__ai int32x2_t vqneg_s32(int32x2_t __p0)
Definition: arm_neon.h:17980
__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:7285
__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:17110
__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:11968
__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:16908
__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:7489
__ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:14782
__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:28826
__ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:67560
__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:29199
__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:17385
__ai uint64x1_t vdup_n_u64(uint64_t __p0)
Definition: arm_neon.h:5315
__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0)
Definition: arm_neon.h:33239
__ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:17371
__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0)
Definition: arm_neon.h:30943
__ai uint16x8_t vmovq_n_u16(uint16_t __p0)
Definition: arm_neon.h:13601
__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:29687
__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:3486
__ai uint8x16_t vclzq_u8(uint8x16_t __p0)
Definition: arm_neon.h:3673
__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0)
Definition: arm_neon.h:33099
__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0)
Definition: arm_neon.h:31125
__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1)
Definition: arm_neon.h:3978
__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:15610
__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:5653
__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0)
Definition: arm_neon.h:16309
__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:2676
__ai uint8x16_t vrev32q_u8(uint8x16_t __p0)
Definition: arm_neon.h:20314
__ai int64x2_t vdupq_n_s64(int64_t __p0)
Definition: arm_neon.h:5255
__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:28788
__ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:17393
__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:67528
__ai int8x16_t vdupq_n_s8(int8_t __p0)
Definition: arm_neon.h:5193
__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:899
__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0)
Definition: arm_neon.h:31097
__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:2727
__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18056
__ai int16x8_t vclzq_s16(int16x8_t __p0)
Definition: arm_neon.h:3753
__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28196
__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28009
__ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18042
__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:29592
__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0)
Definition: arm_neon.h:31979
__ai int16x4_t vget_high_s16(int16x8_t __p0)
Definition: arm_neon.h:6417
__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0)
Definition: arm_neon.h:33085
__ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:22086
__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0)
Definition: arm_neon.h:31503
__ai uint32x4_t vclzq_u32(uint32x4_t __p0)
Definition: arm_neon.h:3689
__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2268
__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:21002
__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:16586
__ai int32x4_t vclsq_s32(int32x4_t __p0)
Definition: arm_neon.h:3355
struct int16x4x4_t int16x4x4_t
__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:20012
__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:1515
__ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:15060
__ai int16x4_t vcls_s16(int16x4_t __p0)
Definition: arm_neon.h:3419
__ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18064
__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:18418
__ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17607
__ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0)
Definition: arm_neon.h:6247
__ai uint8x8_t vmov_n_u8(uint8_t __p0)
Definition: arm_neon.h:13708
__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0)
Definition: arm_neon.h:30789
__ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:14826
__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:967
__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1)
Definition: arm_neon.h:29611
__ai int32x2_t vmov_n_s32(int32_t __p0)
Definition: arm_neon.h:13814
__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0)
Definition: arm_neon.h:4406
__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:15760
__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1)
Definition: arm_neon.h:16141
struct uint64x2x4_t uint64x2x4_t
__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0)
Definition: arm_neon.h:33141
__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0)
Definition: arm_neon.h:31433
__ai uint64x1_t vpaddl_u32(uint32x2_t __p0)
Definition: arm_neon.h:16389
__ai int16x4_t vrev64_s16(int16x4_t __p0)
Definition: arm_neon.h:20714
__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0)
Definition: arm_neon.h:30173
__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:7336
__ai int32x2_t vneg_s32(int32x2_t __p0)
Definition: arm_neon.h:15394
__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2234
__ai uint32x2_t vcreate_u32(uint64_t __p0)
Definition: arm_neon.h:4243
struct uint16x8x2_t uint16x8x2_t
__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:22880
__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:22144
int64x2_t val[4]
Definition: arm_neon.h:342
__ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:69094
Definition: arm_neon.h:217
__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:11900
Definition: arm_neon.h:265
__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:1798
__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:29421
Definition: arm_neon.h:403
struct uint8x16x3_t uint8x16x3_t
__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:3452
__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:1231
float32x4_t val[4]
Definition: arm_neon.h:390
__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:14161
__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:28429
__ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:4025
__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0)
Definition: arm_neon.h:31447
__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0)
Definition: arm_neon.h:32833
__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0)
Definition: arm_neon.h:32021
__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:5684
__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0)
Definition: arm_neon.h:33043
__ai int32x2_t vrev64_s32(int32x2_t __p0)
Definition: arm_neon.h:20698
__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:15562
__ai uint32x4_t vmvnq_u32(uint32x4_t __p0)
Definition: arm_neon.h:15122
Definition: arm_neon.h:209
__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0)
Definition: arm_neon.h:31923
__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28528
__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0)
Definition: arm_neon.h:30873
__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0)
Definition: arm_neon.h:32861
__ai uint16x8_t vdupq_n_u16(uint16_t __p0)
Definition: arm_neon.h:5178
__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:3288
__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0)
Definition: arm_neon.h:33575
__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:14178
__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0)
Definition: arm_neon.h:32623
uint64x1_t val[2]
Definition: arm_neon.h:138
__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0)
Definition: arm_neon.h:30061
__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0)
Definition: arm_neon.h:30089
__ai uint8x16_t vcntq_u8(uint8x16_t __p0)
Definition: arm_neon.h:3897
__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:18798
__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:5435
__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0)
Definition: arm_neon.h:31769
__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:1035
__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0)
Definition: arm_neon.h:33589
__ai int8x8_t vclz_s8(int8x8_t __p0)
Definition: arm_neon.h:3817
__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:11917
__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:7370
__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:16569
__ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:19910
__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0)
Definition: arm_neon.h:33449
__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:3118
__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:19818
int32x4_t val[3]
Definition: arm_neon.h:218
struct poly8x8x4_t poly8x8x4_t
__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:69062
__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28845
__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0)
Definition: arm_neon.h:31755
__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:23027
__ai uint64x1_t vget_high_u64(uint64x2_t __p0)
Definition: arm_neon.h:6282
__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:1100
__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:29554
__ai float32x2_t vmov_n_f32(float32_t __p0)
Definition: arm_neon.h:13782
Definition: arm_neon.h:183
__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0)
Definition: arm_neon.h:31545
__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:3305
__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:3503
__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:1682
poly8x8_t val[4]
Definition: arm_neon.h:404
__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:1416
__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:21053
__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1)
Definition: arm_neon.h:29307
__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:19689
int8x16_t val[4]
Definition: arm_neon.h:318
__ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1)
Definition: arm_neon.h:4128
__ai uint8x8_t vrev64_u8(uint8x8_t __p0)
Definition: arm_neon.h:20618
Definition: arm_neon.h:197
__ai poly16x4_t vdup_n_p16(poly16_t __p0)
Definition: arm_neon.h:5088
__ai poly16x8_t vrev64q_p16(poly16x8_t __p0)
Definition: arm_neon.h:20490
__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1)
Definition: arm_neon.h:18200
__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:2795
__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2)
Definition: arm_neon.h:2099
__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0)
Definition: arm_neon.h:30355
__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2)
Definition: arm_neon.h:13032
Definition: arm_neon.h:129
__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:29046
__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18367
__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0)
Definition: arm_neon.h:33001
__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2)
Definition: arm_neon.h:1829
__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:27679
__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2931
struct int32x4x3_t int32x4x3_t
__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:20934
__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:16654
poly8x16_t val[4]
Definition: arm_neon.h:408
__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:16260
__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:17181
__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0)
Definition: arm_neon.h:30985
__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0)
Definition: arm_neon.h:32105
__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:17051
__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:16688
__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:15876
__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0)
Definition: arm_neon.h:32707
__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:23013
__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1)
Definition: arm_neon.h:14624
__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1)
Definition: arm_neon.h:4120
__ai int16x8_t vrev64q_s16(int16x8_t __p0)
Definition: arm_neon.h:20602
__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:12996
Definition: arm_neon.h:345
__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:865
__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:67620
struct int32x4x4_t int32x4x4_t
__ai int16x4_t vmvn_s16(int16x4_t __p0)
Definition: arm_neon.h:15282
__ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:67814
__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:544
__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:28769
__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:16226
__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2455
__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:28026
__ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:16933
int8x16_t val[2]
Definition: arm_neon.h:86
__ai float32x4_t vmovq_n_f32(float32_t __p0)
Definition: arm_neon.h:13631
Definition: arm_neon.h:175
__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:19672
__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:11815
__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:22166
struct uint16x4x4_t uint16x4x4_t
__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0)
Definition: arm_neon.h:32413
__ai uint8x8_t vrev32_u8(uint8x8_t __p0)
Definition: arm_neon.h:20378
__ai uint8x8_t vmovn_u16(uint16x8_t __p0)
Definition: arm_neon.h:14026
#define vshll_n_s8(__p0, __p1)
Definition: arm_neon.h:23401
__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2319
__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:67147
__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:21709
__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:12749
__ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:67652
__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:2761
__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:2625
__ai uint64x2_t vmovq_n_u64(uint64_t __p0)
Definition: arm_neon.h:13586
Definition: arm_neon.h:381
__ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:69117
__ai int32x4_t vmovl_s16(int16x4_t __p0)
Definition: arm_neon.h:13963
int64x1_t val[3]
Definition: arm_neon.h:222
__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0)
Definition: arm_neon.h:31699
__ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:66924
__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:28807
__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0)
Definition: arm_neon.h:32987
uint32x2_t val[3]
Definition: arm_neon.h:246
__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:3588
#define __noswap_vshll_n_u32(__p0, __p1)
Definition: arm_neon.h:23361
__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:29459
__ai int8x16_t vmovq_n_s8(int8_t __p0)
Definition: arm_neon.h:13616
__ai int16x4_t vmovn_s32(int32x4_t __p0)
Definition: arm_neon.h:14047
__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:7149
__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:17248
__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:22078
__ai int8x8_t vqabs_s8(int8x8_t __p0)
Definition: arm_neon.h:16753
__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28351
__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2)
Definition: arm_neon.h:1937
__ai poly16x8_t vmovq_n_p16(poly16_t __p0)
Definition: arm_neon.h:13541
__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0)
Definition: arm_neon.h:30929
__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:19902
__ai uint16x4_t vqmovun_s32(int32x4_t __p0)
Definition: arm_neon.h:17853
__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2)
Definition: arm_neon.h:1883
__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:18832
__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:15579
__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:11730
__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:14348
__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2)
Definition: arm_neon.h:1901
__ai int32x4_t vnegq_s32(int32x4_t __p0)
Definition: arm_neon.h:15330
__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0)
Definition: arm_neon.h:31307
__ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:574
Definition: arm_neon.h:213
float16x8_t val[4]
Definition: arm_neon.h:382
__ai float32x2_t vneg_f32(float32x2_t __p0)
Definition: arm_neon.h:15378
__ai int16x4_t vabs_s16(int16x4_t __p0)
Definition: arm_neon.h:849
__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2)
Definition: arm_neon.h:1973
__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:5588
__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:527
struct poly16x4x4_t poly16x4x4_t
Definition: arm_neon.h:373
__ai poly16x8_t vrev32q_p16(poly16x8_t __p0)
Definition: arm_neon.h:20298
__ai int16x4_t vrev32_s16(int16x4_t __p0)
Definition: arm_neon.h:20426
__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:3622
__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:67252
__ai uint32x2_t vrev64_u32(uint32x2_t __p0)
Definition: arm_neon.h:20634
__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28213
__ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0)
Definition: arm_neon.h:14012
__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:2081
__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0)
Definition: arm_neon.h:31657
__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:27848
__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:5636
__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:715
poly16x4_t val[4]
Definition: arm_neon.h:412
typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t
int8x8_t val[2]
Definition: arm_neon.h:82
__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1)
Definition: arm_neon.h:16109
__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:1467
__ai poly16x8_t vdupq_n_p16(poly16_t __p0)
Definition: arm_neon.h:5118
struct uint64x1x4_t uint64x1x4_t
__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:5571
__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:5452
int64x1_t val[4]
Definition: arm_neon.h:338
__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0)
Definition: arm_neon.h:30341
__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:16964
__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:66916
__ai poly8x8_t vcreate_p8(uint64_t __p0)
Definition: arm_neon.h:4201
__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0)
Definition: arm_neon.h:30439
__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2)
Definition: arm_neon.h:66684
__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:28636
__ai uint8x16_t vmovq_n_u8(uint8_t __p0)
Definition: arm_neon.h:13556
__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0)
Definition: arm_neon.h:31587
struct float16x8x3_t float16x8x3_t
__ai int32x4_t vdupq_n_s32(int32_t __p0)
Definition: arm_neon.h:5240
__ai uint64x2_t vmovl_u32(uint32x2_t __p0)
Definition: arm_neon.h:13879
__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2)
Definition: arm_neon.h:12942
__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0)
Definition: arm_neon.h:33603
__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0)
Definition: arm_neon.h:30201
struct int64x2x2_t int64x2x2_t
__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:29668
__ai float32x4_t vnegq_f32(float32x4_t __p0)
Definition: arm_neon.h:15314
__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0)
Definition: arm_neon.h:30551
int8x8_t val[3]
Definition: arm_neon.h:198
__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0)
Definition: arm_neon.h:30677
__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:3016
__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:29801
__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18781
__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:29497
__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:3033
__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:13411
__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:29763
__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18497
Definition: arm_neon.h:229
int32x2_t val[2]
Definition: arm_neon.h:98
struct int8x8x3_t int8x8x3_t
__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28179
__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:66774
__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0)
Definition: arm_neon.h:30369
__ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17415
__ai uint16x8_t vrev32q_u16(uint16x8_t __p0)
Definition: arm_neon.h:20330
__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:27560
__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:14818
__fp16 float16_t
Definition: arm_neon.h:34
__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:3995
__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0)
Definition: arm_neon.h:30313
__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18883
__ai uint32x2_t vqmovn_u64(uint64x2_t __p0)
Definition: arm_neon.h:17748
__ai int64x2_t vpaddlq_s32(int32x4_t __p0)
Definition: arm_neon.h:16341
__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:22778
Definition: arm_neon.h:133
__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:7183
__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:20730
__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0)
Definition: arm_neon.h:33225
uint32x4_t val[4]
Definition: arm_neon.h:366
__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0)
Definition: arm_neon.h:33253
__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:2965
__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0)
Definition: arm_neon.h:32161
__ai uint32x2_t vclz_u32(uint32x2_t __p0)
Definition: arm_neon.h:3785
__ai uint8x16_t vrev64q_u8(uint8x16_t __p0)
Definition: arm_neon.h:20506
__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:15726
__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:22979
__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:28883
__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:18350
__ai poly16x4_t vrev64_p16(poly16x4_t __p0)
Definition: arm_neon.h:20458
int32_t val
Definition: Sensors.cpp:524
__ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:69048
__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:2199
__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28162
struct uint8x8x3_t uint8x8x3_t
__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:14144
__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:3435
__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:29012
__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0)
Definition: arm_neon.h:33519
__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0)
Definition: arm_neon.h:33351
__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2)
Definition: arm_neon.h:12370
__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0)
Definition: arm_neon.h:32329
__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:14840
__ai int8x16_t vrev64q_s8(int8x16_t __p0)
Definition: arm_neon.h:20554
__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1)
Definition: arm_neon.h:28995
__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:5503
__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0)
Definition: arm_neon.h:32791
__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:22948
__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:66938
__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:22996
__ai float32x4_t vabsq_f32(float32x4_t __p0)
Definition: arm_neon.h:753
__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:22914
__ai int16x4_t vqmovn_s32(int32x4_t __p0)
Definition: arm_neon.h:17790
__ai int16x4_t vget_low_s16(int16x8_t __p0)
Definition: arm_neon.h:7116
__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0)
Definition: arm_neon.h:33547
__ai uint16x8_t vmvnq_u16(uint16x8_t __p0)
Definition: arm_neon.h:15138
__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0)
Definition: arm_neon.h:32539
__ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:723
__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:21692
__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:2353
__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:27870
__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2)
Definition: arm_neon.h:13326
__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28447
__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:68993
__ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:14804
__ai uint32x4_t vdupq_n_u32(uint32_t __p0)
Definition: arm_neon.h:5148
Definition: arm_neon.h:257
__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:66850
__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:67506
__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:16060
__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:13462
__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0)
Definition: arm_neon.h:30495
__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:15596
__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:2897
__ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:22152
__ai uint8x8_t vget_low_u8(uint8x16_t __p0)
Definition: arm_neon.h:6974
__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:17685
Definition: arm_neon.h:225
__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:15975
__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:2167
__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:676
__ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:4003
__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:693
Definition: arm_neon.h:117
__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:12800
__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:16603
__ai uint16x4_t vcreate_u16(uint64_t __p0)
Definition: arm_neon.h:4271
int16x8_t val[2]
Definition: arm_neon.h:94
__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:1001
__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0)
Definition: arm_neon.h:33533
__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:1532
__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:29440
Definition: arm_neon.h:361
uint8x8_t val[4]
Definition: arm_neon.h:346
__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:17341
__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:20764
__ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0)
Definition: arm_neon.h:14096
__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0)
Definition: arm_neon.h:31475
__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:17583
__ai int32x2_t vqmovn_s64(int64x2_t __p0)
Definition: arm_neon.h:17811
__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0)
Definition: arm_neon.h:31881
uint64x2_t val[4]
Definition: arm_neon.h:374
__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:1781
__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:984
#define __noswap_vshll_n_s8(__p0, __p1)
Definition: arm_neon.h:23409
__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0)
Definition: arm_neon.h:30691
__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0)
Definition: arm_neon.h:32931
__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0)
Definition: arm_neon.h:32637
__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0)
Definition: arm_neon.h:30831
__ai uint32x2_t vget_high_u32(uint32x4_t __p0)
Definition: arm_neon.h:6261
float16x4_t val[2]
Definition: arm_neon.h:146
__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:19706
__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:27628
__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:7319
__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0)
Definition: arm_neon.h:32091
__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:2829
struct float16x4x4_t float16x4x4_t
__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:20815
__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0)
Definition: arm_neon.h:30621
__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:3639
__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:14246
__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:20747
__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0)
Definition: arm_neon.h:33267
__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:610
__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0)
Definition: arm_neon.h:33127
__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:27892
struct int32x2x2_t int32x2x2_t
__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0)
Definition: arm_neon.h:30579
__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0)
Definition: arm_neon.h:32553
__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0)
Definition: arm_neon.h:32721
__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:13068
struct uint32x2x4_t uint32x2x4_t
__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:1716
__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0)
Definition: arm_neon.h:30845
uint16x8_t val[3]
Definition: arm_neon.h:242
struct poly8x16x3_t poly8x16x3_t
__ai int32x2_t vqabs_s32(int32x2_t __p0)
Definition: arm_neon.h:16769
__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:3322
__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0)
Definition: arm_neon.h:32175
__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18747
__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18432
#define vshll_n_u16(__p0, __p1)
Definition: arm_neon.h:23377
__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2)
Definition: arm_neon.h:12226
int32x4_t val[4]
Definition: arm_neon.h:334
__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0)
Definition: arm_neon.h:33463
__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0)
Definition: arm_neon.h:32357
__ai float32x2_t vget_low_f32(float32x4_t __p0)
Definition: arm_neon.h:7053
__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:15777
struct float32x4x2_t float32x4x2_t
__ai int64x1_t vmov_n_s64(int64_t __p0)
Definition: arm_neon.h:13829
__ai int8x8_t vget_low_s8(int8x16_t __p0)
Definition: arm_neon.h:7037
struct float16x8x4_t float16x8x4_t
__ai uint32x2_t vpaddl_u16(uint16x4_t __p0)
Definition: arm_neon.h:16404
struct uint64x2x2_t uint64x2x2_t
__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1)
Definition: arm_neon.h:17535
__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0)
Definition: arm_neon.h:32511
__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0)
Definition: arm_neon.h:31153
__ai float16x4_t vcreate_f16(uint64_t __p0)
Definition: arm_neon.h:4313
__ai float32x2_t vdup_n_f32(float32_t __p0)
Definition: arm_neon.h:5359
__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:461
__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:20849
__ai int32x4_t vmovq_n_s32(int32_t __p0)
Definition: arm_neon.h:13663
__ai int8x8_t vrev32_s8(int8x8_t __p0)
Definition: arm_neon.h:20410
__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2)
Definition: arm_neon.h:12870
__ai int16x8_t vqnegq_s16(int16x8_t __p0)
Definition: arm_neon.h:17948
__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1)
Definition: arm_neon.h:14656
__ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:19770
__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:12766
__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:15893
__ai uint16x8_t vrev64q_u16(uint16x8_t __p0)
Definition: arm_neon.h:20538
__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:69108
struct poly16x8x3_t poly16x8x3_t
__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:66792
__ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:1261
__ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1)
Definition: arm_neon.h:14997
__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:2642
__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28310
__ai int64x2_t vmovq_n_s64(int64_t __p0)
Definition: arm_neon.h:13678
__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:2557
__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0)
Definition: arm_neon.h:33281
__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0)
Definition: arm_neon.h:32483
__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0)
Definition: arm_neon.h:33505
__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0)
Definition: arm_neon.h:31643
__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2)
Definition: arm_neon.h:12172
__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0)
Definition: arm_neon.h:32007
__ai int32x4_t vpaddlq_s16(int16x8_t __p0)
Definition: arm_neon.h:16357
__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:1314
__ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:17257
__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:2914
__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:16852
__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:21036
__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0)
Definition: arm_neon.h:32217
__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0)
Definition: arm_neon.h:30523
__ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:67307
struct poly8x16x2_t poly8x16x2_t
__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2)
Definition: arm_neon.h:1955
struct int16x4x3_t int16x4x3_t
__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:1481
__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0)
Definition: arm_neon.h:30425
Definition: arm_neon.h:245
uint32x2_t val[2]
Definition: arm_neon.h:130
__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:67762
struct uint32x4x3_t uint32x4x3_t
__ai int32x2_t vmvn_s32(int32x2_t __p0)
Definition: arm_neon.h:15266
__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:28674
Definition: arm_neon.h:85
__ai int8x16_t vrev16q_s8(int8x16_t __p0)
Definition: arm_neon.h:20202
__ai int64x1_t vdup_n_s64(int64_t __p0)
Definition: arm_neon.h:5406
__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1)
Definition: arm_neon.h:14592
__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0)
Definition: arm_neon.h:32903
__ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:67284
__ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:27878
__ai int8x16_t vqnegq_s8(int8x16_t __p0)
Definition: arm_neon.h:17916
__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:3537
__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:20798
__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0)
Definition: arm_neon.h:31811
__ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:20042
__ai int16x4_t vclz_s16(int16x4_t __p0)
Definition: arm_neon.h:3849
__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:916
struct float32x4x3_t float32x4x3_t
__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:2063
__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:20120
__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:14331
__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:2506
__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0)
Definition: arm_neon.h:30075
__ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0)
Definition: arm_neon.h:17776
__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:29782
Definition: arm_neon.h:365
__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:15941
__ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:4084
__ai uint32x4_t vmovl_u16(uint16x4_t __p0)
Definition: arm_neon.h:13900
__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0)
Definition: arm_neon.h:31573
__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0)
Definition: arm_neon.h:33155
__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:1209
__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0)
Definition: arm_neon.h:33295
__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0)
Definition: arm_neon.h:31139
__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:4054
__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:16671
__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18034
__ai int8x16_t vrev32q_s8(int8x16_t __p0)
Definition: arm_neon.h:20346
uint8x16_t val[2]
Definition: arm_neon.h:118
struct poly8x16x4_t poly8x16x4_t
__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:16077
__ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0)
Definition: arm_neon.h:17797
__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:14195
__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:29725
__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:67551
__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0)
Definition: arm_neon.h:33197
__ai int8x8_t vcreate_s8(uint64_t __p0)
Definition: arm_neon.h:4285
__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0)
Definition: arm_neon.h:30131
__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:2778
__ai float32x4_t vrecpeq_f32(float32x4_t __p0)
Definition: arm_neon.h:20072
__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28251
__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0)
Definition: arm_neon.h:33113
__ai int8x8_t vget_high_s8(int8x16_t __p0)
Definition: arm_neon.h:6318
struct int8x16x2_t int8x16x2_t
__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:7234
__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:27543
__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0)
Definition: arm_neon.h:30971
__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0)
Definition: arm_neon.h:32819
__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:20968
__ai uint16x4_t vrev64_u16(uint16x4_t __p0)
Definition: arm_neon.h:20650
__ai uint64x2_t vdupq_n_u64(uint64_t __p0)
Definition: arm_neon.h:5163
__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:4164
__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:20900
__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0)
Definition: arm_neon.h:31727
__ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:17349
__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0)
Definition: arm_neon.h:32497
__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:29383
__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:29250
__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0)
Definition: arm_neon.h:30229
__ai float32x2_t vrev64_f32(float32x2_t __p0)
Definition: arm_neon.h:20682
Definition: arm_neon.h:121
__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:7472
Definition: arm_neon.h:317
__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0)
Definition: arm_neon.h:33435
__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0)
Definition: arm_neon.h:33169
__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2)
Definition: arm_neon.h:2185
__ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0)
Definition: arm_neon.h:4376
__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:14365
__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0)
Definition: arm_neon.h:33491
__ai int64x2_t vmovl_s32(int32x2_t __p0)
Definition: arm_neon.h:13942
__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2)
Definition: arm_neon.h:13445
__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0)
Definition: arm_neon.h:32595
__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:11934
poly16x8_t val[3]
Definition: arm_neon.h:300
__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2)
Definition: arm_neon.h:66648
struct int8x8x2_t int8x8x2_t
Definition: arm_neon.h:349
__ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:662
__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17429
float16x8_t val[2]
Definition: arm_neon.h:150
__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:3961
__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0)
Definition: arm_neon.h:4422
__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:29896
__ai uint16x4_t vmovn_u32(uint32x4_t __p0)
Definition: arm_neon.h:13984
Definition: arm_neon.h:153
struct float32x2x2_t float32x2x2_t
#define vshll_n_u32(__p0, __p1)
Definition: arm_neon.h:23353
__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:3237
__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:12121
__ai int8x8_t vcnt_s8(int8x8_t __p0)
Definition: arm_neon.h:3945
__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:3203
__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:66994
__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0)
Definition: arm_neon.h:30481
__ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18020
__ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17037
__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:12388
__ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0)
Definition: arm_neon.h:14054
__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:4098
__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:1134
__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1)
Definition: arm_neon.h:15011
__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0)
Definition: arm_neon.h:33379
__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:14212
__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18012
struct poly16x8x4_t poly16x8x4_t
__ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:67215
__ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0)
Definition: arm_neon.h:13907
__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:12334
__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0)
Definition: arm_neon.h:32035
__ai uint32x2_t vmvn_u32(uint32x2_t __p0)
Definition: arm_neon.h:15218
uint64x1_t val[3]
Definition: arm_neon.h:254
__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:22863
struct int16x8x3_t int16x8x3_t
__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1)
Definition: arm_neon.h:16043
__ai uint8x8_t vcreate_u8(uint64_t __p0)
Definition: arm_neon.h:4229
__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0)
Definition: arm_neon.h:31167
__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:16818
__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1)
Definition: arm_neon.h:14672
__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:2540
__ai uint8x8_t vclz_u8(uint8x8_t __p0)
Definition: arm_neon.h:3769
__ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:66836
__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:20917
__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0)
Definition: arm_neon.h:31321
__ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:67448
__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:1165
__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:67062
__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:1764
__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0)
Definition: arm_neon.h:33365
__ai int64x1_t vget_high_s64(int64x2_t __p0)
Definition: arm_neon.h:6402
__ai int32x2_t vabs_s32(int32x2_t __p0)
Definition: arm_neon.h:833
Definition: arm_neon.h:249
int64x1_t val[2]
Definition: arm_neon.h:106
struct int8x16x4_t int8x16x4_t
__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:4017
__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:1187
__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0)
Definition: arm_neon.h:32889
__ai poly16x4_t vget_low_p16(poly16x8_t __p0)
Definition: arm_neon.h:6958
struct int32x2x3_t int32x2x3_t
float32x2_t val[4]
Definition: arm_neon.h:386
struct int32x2x4_t int32x2x4_t
__ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0)
Definition: arm_neon.h:17860
__ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:17327
uint64x2_t val[3]
Definition: arm_neon.h:258
__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:19866
__ai int16x4_t vcreate_s16(uint64_t __p0)
Definition: arm_neon.h:4355
uint64x1_t val[4]
Definition: arm_neon.h:370
__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:29148
__ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:27856
__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:1580
__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:67275
__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18466
#define __noswap_vshll_n_u8(__p0, __p1)
Definition: arm_neon.h:23337
__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0)
Definition: arm_neon.h:32315
__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:15743
__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0)
Definition: arm_neon.h:31965
__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:16801
__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:28978
__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0)
Definition: arm_neon.h:33071
__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:14884
__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0)
Definition: arm_neon.h:32581
__ai int8x8_t vmov_n_s8(int8_t __p0)
Definition: arm_neon.h:13767
struct int64x1x3_t int64x1x3_t
__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:16620
__ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0)
Definition: arm_neon.h:6210
__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:20137
__ai poly8x8_t vmov_n_p8(poly8_t __p0)
Definition: arm_neon.h:13496
__ai int8x8_t vqmovn_s16(int16x8_t __p0)
Definition: arm_neon.h:17832
uint16x8_t val[4]
Definition: arm_neon.h:358
struct int16x4x2_t int16x4x2_t
__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:15511
__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:29744
struct uint32x2x3_t uint32x2x3_t
Definition: arm_neon.h:137
Definition: arm_neon.h:295
Definition: arm_neon.h:105
__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0)
Definition: arm_neon.h:31027
__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28290
__ai float32x2_t vrsqrte_f32(float32x2_t __p0)
Definition: arm_neon.h:21676
__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:15545
__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0)
Definition: arm_neon.h:31951
__ai poly8x8_t vcnt_p8(poly8x8_t __p0)
Definition: arm_neon.h:3865
__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0)
Definition: arm_neon.h:32777
__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0)
Definition: arm_neon.h:30243
__ai int16x4_t vqabs_s16(int16x4_t __p0)
Definition: arm_neon.h:16785
__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:21152
__ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:618
__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0)
Definition: arm_neon.h:32399
__ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:19976
__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:16484
__ai uint16x4_t vclz_u16(uint16x4_t __p0)
Definition: arm_neon.h:3801
__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:3186
__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:15958
__ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17629
__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0)
Definition: arm_neon.h:33617
struct uint8x8x2_t uint8x8x2_t
__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:1433
float32x4_t val[3]
Definition: arm_neon.h:274
__ai poly8x8_t vmvn_p8(poly8x8_t __p0)
Definition: arm_neon.h:15074
int16x8_t val[3]
Definition: arm_neon.h:210
__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0)
Definition: arm_neon.h:31993
__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:66828
__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0)
Definition: arm_neon.h:30411
__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:21104
__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0)
Definition: arm_neon.h:32693
Definition: arm_neon.h:233
__ai float32x2_t vrecpe_f32(float32x2_t __p0)
Definition: arm_neon.h:20104
__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2)
Definition: arm_neon.h:12888
__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:29706
struct int8x8x4_t int8x8x4_t
__ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0)
Definition: arm_neon.h:6424
__ai uint8x8_t vget_high_u8(uint8x16_t __p0)
Definition: arm_neon.h:6240
__ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:67261
__ai int8x16_t vnegq_s8(int8x16_t __p0)
Definition: arm_neon.h:15298
__ai uint16x8_t vmovl_u8(uint8x8_t __p0)
Definition: arm_neon.h:13858
__ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:67792
__ai int8x8_t vmovn_s16(int16x8_t __p0)
Definition: arm_neon.h:14089
__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:15494
__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:67011
__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:66720
__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0)
Definition: arm_neon.h:31559
__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:15644
__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1)
Definition: arm_neon.h:29649
__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2336
__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:67028
__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:29839
__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:1614
struct int64x1x2_t int64x1x2_t
uint16x4_t val[2]
Definition: arm_neon.h:122
__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:15426
__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2)
Definition: arm_neon.h:1847
__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0)
Definition: arm_neon.h:32749
__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0)
Definition: arm_neon.h:33211
__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:2131
__ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:14760
__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0)
Definition: arm_neon.h:32651
__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:27526
__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0)
Definition: arm_neon.h:30537
__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:1066
Definition: arm_neon.h:157
__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0)
Definition: arm_neon.h:33337
__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0)
Definition: arm_neon.h:31013
__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:2489
__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0)
Definition: arm_neon.h:31377
__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0)
Definition: arm_neon.h:30467
__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:27713
__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:29326
__ai int16x4_t vqneg_s16(int16x4_t __p0)
Definition: arm_neon.h:17996
#define __noswap_vshll_n_s16(__p0, __p1)
Definition: arm_neon.h:23457
__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:66872
__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:19723
__ai int32x2_t vmovn_s64(int64x2_t __p0)
Definition: arm_neon.h:14068
__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:3520
__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0)
Definition: arm_neon.h:32203
Definition: arm_neon.h:337
__ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:22108
__ai uint64x1_t vget_low_u64(uint64x2_t __p0)
Definition: arm_neon.h:7006
__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2251
struct uint8x16x4_t uint8x16x4_t
__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2302
uint16x4_t val[3]
Definition: arm_neon.h:238
__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:2999
__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:66738
#define __noswap_vshll_n_s32(__p0, __p1)
Definition: arm_neon.h:23433
__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:1348
__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:17363
__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0)
Definition: arm_neon.h:32525
__ai poly8x16_t vrev64q_p8(poly8x16_t __p0)
Definition: arm_neon.h:20474
__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:1450
__ai int8x16_t vclzq_s8(int8x16_t __p0)
Definition: arm_neon.h:3721
__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1)
Definition: arm_neon.h:28655
uint16x8_t val[2]
Definition: arm_neon.h:126
__ai int8x8_t vrev64_s8(int8x8_t __p0)
Definition: arm_neon.h:20666
__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:12002
__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0)
Definition: arm_neon.h:30285
uint8x16_t val[4]
Definition: arm_neon.h:350
__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:7455
__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28902
__ai uint16x4_t vdup_n_u16(uint16_t __p0)
Definition: arm_neon.h:5329
__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:4039
__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:17225
__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:21118
poly16x8_t val[2]
Definition: arm_neon.h:184
__ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:447
__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:5537
__ai int8x8_t vdup_n_s8(int8_t __p0)
Definition: arm_neon.h:5344
__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0)
Definition: arm_neon.h:31937
__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0)
Definition: arm_neon.h:30299
__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0)
Definition: arm_neon.h:32945
__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:7200
__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:5520
struct uint32x4x2_t uint32x4x2_t
__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:483
__ai float32x4_t vcvt_f32_f16(float16x4_t __p0)
Definition: arm_neon.h:4454
__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:11696
__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:29097
__ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0)
Definition: arm_neon.h:6325
__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2)
Definition: arm_neon.h:12190
__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:27914
struct poly16x4x2_t poly16x4x2_t
__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0)
Definition: arm_neon.h:31489
__ai poly8x16_t vcntq_p8(poly8x16_t __p0)
Definition: arm_neon.h:3881
Definition: arm_neon.h:89
Definition: arm_neon.h:353
__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0)
Definition: arm_neon.h:33057
__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:1665
__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:27958
__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0)
Definition: arm_neon.h:33421
__ai int32x2_t vget_high_s32(int32x4_t __p0)
Definition: arm_neon.h:6381
__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:16192
__ai int8x8_t vcls_s8(int8x8_t __p0)
Definition: arm_neon.h:3387
__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:14862
__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:27696
__ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:596
__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:28145
__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:15692
__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0)
Definition: arm_neon.h:31279
__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0)
Definition: arm_neon.h:32343
__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2)
Definition: arm_neon.h:6157
__ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:19731
__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1)
Definition: arm_neon.h:17551
__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:4142
__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0)
Definition: arm_neon.h:31111
__ai int8x8_t vabs_s8(int8x8_t __p0)
Definition: arm_neon.h:801
__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:3135
struct poly16x4x3_t poly16x4x3_t
__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:20866
__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0)
Definition: arm_neon.h:31405
__ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:22174
__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:3469
Definition: arm_neon.h:385
__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:3220
int32x2_t val[3]
Definition: arm_neon.h:214
__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0)
Definition: arm_neon.h:20056
__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18265
__ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:14892
__ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:1217
__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:12053
__ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:19954
struct int64x2x4_t int64x2x4_t
__ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:67770
__ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1)
Definition: arm_neon.h:15018
float float32_t
Definition: arm_neon.h:33
__ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:67470
__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0)
Definition: arm_neon.h:30047
__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:2608
__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:12087
__ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:22130
__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:12406
__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:28077
__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:27577
__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:1117
__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:5605
__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:11985
poly16x4_t val[2]
Definition: arm_neon.h:180
__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0)
Definition: arm_neon.h:31713
__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:505
__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:20951
poly8x8_t val[3]
Definition: arm_neon.h:288
__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2285
__ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:491
__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:27611
__ai int16x8_t vabsq_s16(int16x8_t __p0)
Definition: arm_neon.h:785
__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:7438
__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:4076
struct uint16x8x3_t uint16x8x3_t
__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:21019
__ai int16x4_t vdup_n_s16(int16_t __p0)
Definition: arm_neon.h:5420
__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:16981
__ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:17096
__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:28959
Definition: arm_neon.h:145
__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:1083
__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0)
Definition: arm_neon.h:30257
Definition: arm_neon.h:269
__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:18900
__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:12036
__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2)
Definition: arm_neon.h:12208
__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:5554
__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:29858
Definition: arm_neon.h:125
__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:67045
__ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:27922
__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0)
Definition: arm_neon.h:32371
__ai poly8x16_t vmovq_n_p8(poly8_t __p0)
Definition: arm_neon.h:13526
__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:17567
__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:11832
__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0)
Definition: arm_neon.h:32147
__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:28128
__ai int16x8_t vmvnq_s16(int16x8_t __p0)
Definition: arm_neon.h:15186
__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:5670
__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0)
Definition: arm_neon.h:31909
Definition: arm_neon.h:321
__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0)
Definition: arm_neon.h:30663
__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:11951
struct float32x2x3_t float32x2x3_t
__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:16501
__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:19832
int16x4_t val[2]
Definition: arm_neon.h:90
__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:28094
struct uint64x1x3_t uint64x1x3_t
__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:1297
__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18948
__ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:17189
__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0)
Definition: arm_neon.h:32875
__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:1399
__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:588
Definition: arm_neon.h:369
__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:16947
__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0)
Definition: arm_neon.h:30145
#define __ai
Definition: arm_neon.h:430
__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2)
Definition: arm_neon.h:12262
__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:69039
__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2)
Definition: arm_neon.h:12783
float16x4_t val[3]
Definition: arm_neon.h:262
__ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:17211
__ai int8x8_t vmvn_s8(int8x8_t __p0)
Definition: arm_neon.h:15250
__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:15528
__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:1812
__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:29877
__ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0)
Definition: arm_neon.h:6304
Definition: arm_neon.h:237
__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2)
Definition: arm_neon.h:1991
__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2)
Definition: arm_neon.h:12732
__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0)
Definition: arm_neon.h:31041
__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:19801
__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:2846
__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28372
__ai float32x2_t vabs_f32(float32x2_t __p0)
Definition: arm_neon.h:817
__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:12070
__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28467
__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0)
Definition: arm_neon.h:30957
__ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:469
__ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2)
Definition: arm_neon.h:6166
__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:69016
__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1)
Definition: arm_neon.h:16009
__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28393
uint8x16_t val[3]
Definition: arm_neon.h:234
__ai poly8x8_t vget_high_p8(poly8x16_t __p0)
Definition: arm_neon.h:6203
__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:28111
__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0)
Definition: arm_neon.h:30159
__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17407
__ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:67606
__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0)
Definition: arm_neon.h:32847
__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:11764
__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:29535
__ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:67492
__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0)
Definition: arm_neon.h:31223
__ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0)
Definition: arm_neon.h:17755
__ai poly8x16_t vdupq_n_p8(poly8_t __p0)
Definition: arm_neon.h:5103
__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:11713
return(x >> y)|(x<< (32-y))
__ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:19874
struct uint16x4x2_t uint16x4x2_t
__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:15032
int32x2_t val[4]
Definition: arm_neon.h:330
__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:66894
__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:18401
__ai uint8x8_t vqmovun_s16(int16x8_t __p0)
Definition: arm_neon.h:17895
__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:7523
__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:13479
__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:12316
__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:14736
__ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:4062
__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:12155
__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:67079
Definition: arm_neon.h:389
__ai uint8x8_t vmvn_u8(uint8x8_t __p0)
Definition: arm_neon.h:15202
__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:18931
__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2)
Definition: arm_neon.h:12906
Definition: arm_neon.h:299
__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0)
Definition: arm_neon.h:30565
__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0)
Definition: arm_neon.h:32735
__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1)
Definition: arm_neon.h:18184
__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0)
Definition: arm_neon.h:30019
__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:28043
__ai float32x4_t vrev64q_f32(float32x4_t __p0)
Definition: arm_neon.h:20570
__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:1275
__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0)
Definition: arm_neon.h:31853
__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0)
Definition: arm_neon.h:31531
__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:2744
__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:1148
__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:2710
__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:1052
__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0)
Definition: arm_neon.h:30803
__ai int16x8_t vdupq_n_s16(int16_t __p0)
Definition: arm_neon.h:5270
__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:15443
__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:7251
__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:13050
__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0)
Definition: arm_neon.h:31629
__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2)
Definition: arm_neon.h:2009
__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:21135
__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:21183
__ai float32x2_t vcvt_f32_s32(int32x2_t __p0)
Definition: arm_neon.h:4438
__ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:20020
__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1)
Definition: arm_neon.h:16026
__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0)
Definition: arm_neon.h:32231
__ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18086
__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:66960
poly8x8_t val[2]
Definition: arm_neon.h:172
__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:18316
__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:17012
Definition: arm_neon.h:201
__ai uint16x4_t vpaddl_u8(uint8x8_t __p0)
Definition: arm_neon.h:16373
__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:17319
int16x4_t val[4]
Definition: arm_neon.h:322
__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0)
Definition: arm_neon.h:32917
__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:22962
__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:2113
__ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0)
Definition: arm_neon.h:13991
__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0)
Definition: arm_neon.h:32189
__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:27778
__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0)
Definition: arm_neon.h:4651
struct uint32x4x4_t uint32x4x4_t
__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1)
Definition: arm_neon.h:14640
__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0)
Definition: arm_neon.h:31419
__ai float16x4_t vget_low_f16(float16x8_t __p0)
Definition: arm_neon.h:7069
__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:14280
__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:16175
Definition: arm_neon.h:273
__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0)
Definition: arm_neon.h:31083
__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:67113
__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:29165
__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:7421
__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18914
__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:15845
__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:16243
__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:7217
__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:15661
__ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:67330
__ai float16x4_t vcvt_f16_f32(float32x4_t __p0)
Definition: arm_neon.h:4369
__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:17706
__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:1546
struct int64x2x3_t int64x2x3_t
__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:28731
Definition: arm_neon.h:253
__ai float32x2_t vget_high_f32(float32x4_t __p0)
Definition: arm_neon.h:6339
Definition: arm_neon.h:291
__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1)
Definition: arm_neon.h:29630
__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:3050
__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0)
Definition: arm_neon.h:30901
__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0)
Definition: arm_neon.h:32763
__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:7506
__ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2)
Definition: arm_neon.h:6189
__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1)
Definition: arm_neon.h:5469
__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:3084
__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0)
Definition: arm_neon.h:32455
__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0)
Definition: arm_neon.h:32063
__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:66810
__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:12138
__ai uint8x8_t vqmovn_u16(uint16x8_t __p0)
Definition: arm_neon.h:17769
__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2)
Definition: arm_neon.h:17087
__ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:16894
__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0)
Definition: arm_neon.h:30775
__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0)
Definition: arm_neon.h:30635
__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:2404
__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0)
Definition: arm_neon.h:31601
__ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0)
Definition: arm_neon.h:6268
__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:3254
__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:18483
__ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0)
Definition: arm_neon.h:13949
__ai uint16x4_t vget_low_u16(uint16x8_t __p0)
Definition: arm_neon.h:7021
__ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:1173
__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0)
Definition: arm_neon.h:32119
__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:3101
__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2)
Definition: arm_neon.h:66666
__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0)
Definition: arm_neon.h:21660
__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2)
Definition: arm_neon.h:13377
__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0)
Definition: arm_neon.h:32679
__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:2591
__ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0)
Definition: arm_neon.h:14033
__ai int32x4_t vmvnq_s32(int32x4_t __p0)
Definition: arm_neon.h:15170
int32x4_t val[2]
Definition: arm_neon.h:102
__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:16869
__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:15910
__ai int32x4_t vabsq_s32(int32x4_t __p0)
Definition: arm_neon.h:769
__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2)
Definition: arm_neon.h:66702
__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:2693
__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:29820
__ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0)
Definition: arm_neon.h:13970
__ai int8x16_t vclsq_s8(int8x16_t __p0)
Definition: arm_neon.h:3339
__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:67229
__ai uint32x2_t vdup_n_u32(uint32_t __p0)
Definition: arm_neon.h:5300
__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2)
Definition: arm_neon.h:12715
__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:29573
__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0)
Definition: arm_neon.h:31293
__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0)
Definition: arm_neon.h:32665
__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:27662
__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:7387
__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1)
Definition: arm_neon.h:3169
__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0)
Definition: arm_neon.h:33561
__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0)
Definition: arm_neon.h:30397
__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:67321
__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:28864
__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2)
Definition: arm_neon.h:12664
__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:1563
Definition: arm_neon.h:101
__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:28940
__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:67643
__ai int8x16_t vcntq_s8(int8x16_t __p0)
Definition: arm_neon.h:3913
__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2)
Definition: arm_neon.h:2149
__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:27936
__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0)
Definition: arm_neon.h:30383
__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:67574
__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2)
Definition: arm_neon.h:12978
__ai int64x1_t vcreate_s64(uint64_t __p0)
Definition: arm_neon.h:4341
struct float16x4x3_t float16x4x3_t
__ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:66858
__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0)
Definition: arm_neon.h:31349
__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:29216
__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:1253
__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:632
int16x8_t val[4]
Definition: arm_neon.h:326
__ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0)
Definition: arm_neon.h:13865
__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:19849
__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1)
Definition: arm_neon.h:29029
__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0)
Definition: arm_neon.h:31363
__ai int8x8_t vqneg_s8(int8x8_t __p0)
Definition: arm_neon.h:17964
__ai poly16x4_t vget_high_p16(poly16x8_t __p0)
Definition: arm_neon.h:6224
__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0)
Definition: arm_neon.h:4390
uint8x8_t val[2]
Definition: arm_neon.h:114
__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:15862
Definition: arm_neon.h:93
__ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:27944
__ai int8x16_t vabsq_s8(int8x16_t __p0)
Definition: arm_neon.h:737
__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28570
Definition: arm_neon.h:357
__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:1382
__ai uint32x2_t vmovn_u64(uint64x2_t __p0)
Definition: arm_neon.h:14005
Definition: arm_neon.h:411
__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:2438
__ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:22064
__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:19745
struct poly8x8x2_t poly8x8x2_t
__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:14774
__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:29080
__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:17203
__ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:27834
__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:18979
__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0)
Definition: arm_neon.h:29991
__ai int8x8_t vrev16_s8(int8x8_t __p0)
Definition: arm_neon.h:20234
uint32x4_t val[3]
Definition: arm_neon.h:250
float32x4_t val[2]
Definition: arm_neon.h:158
__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0)
Definition: arm_neon.h:29977
__ai uint8x16_t vdupq_n_u8(uint8_t __p0)
Definition: arm_neon.h:5133
__ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17073
__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:18333
struct int64x1x4_t int64x1x4_t
__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:16209
__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:18449
__ai int32x4_t vqnegq_s32(int32x4_t __p0)
Definition: arm_neon.h:17932
__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28330
__ai poly8x8_t vdup_n_p8(poly8_t __p0)
Definition: arm_neon.h:5073
__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:29516
__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2)
Definition: arm_neon.h:12681
__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:13428
__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0)
Definition: arm_neon.h:32959
__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0)
Definition: arm_neon.h:32609
__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2)
Definition: arm_neon.h:12244
__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0)
Definition: arm_neon.h:33477
__ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:67514
__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:3554
__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:29131
__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:27727
__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:27975
__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:19968
int16_t poly16_t
Definition: arm_neon.h:46
__ai float16x4_t vget_high_f16(float16x8_t __p0)
Definition: arm_neon.h:6360
__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:2472
__ai int32x2_t vcls_s32(int32x2_t __p0)
Definition: arm_neon.h:3403
__ai int8x16_t vqabsq_s8(int8x16_t __p0)
Definition: arm_neon.h:16705
__ai int32x2_t vget_low_s32(int32x4_t __p0)
Definition: arm_neon.h:7085
__ai poly8x16_t vmvnq_p8(poly8x16_t __p0)
Definition: arm_neon.h:15090
Definition: arm_neon.h:81
int64x2_t val[3]
Definition: arm_neon.h:226
__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1)
Definition: arm_neon.h:14704
__ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:19932
__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0)
Definition: arm_neon.h:30859
__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:18216
__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1)
Definition: arm_neon.h:29269
__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:566
__ai int16x8_t vnegq_s16(int16x8_t __p0)
Definition: arm_neon.h:15346
__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:2387
__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0)
Definition: arm_neon.h:31195
struct uint64x2x3_t uint64x2x3_t
__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0)
Definition: arm_neon.h:4619
__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:67206
__ai int32x4_t vclzq_s32(int32x4_t __p0)
Definition: arm_neon.h:3737
__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:2523
__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0)
Definition: arm_neon.h:21628
Definition: arm_neon.h:333
__ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:19998
struct int8x16x3_t int8x16x3_t
Definition: arm_neon.h:149
__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1)
Definition: arm_neon.h:16126
__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:29364
int16x4_t val[3]
Definition: arm_neon.h:206
__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:29345
__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0)
Definition: arm_neon.h:31867
__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:19990
__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1)
Definition: arm_neon.h:14752
uint32x2_t val[4]
Definition: arm_neon.h:362
__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2)
Definition: arm_neon.h:12852
__ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1)
Definition: arm_neon.h:15039
struct float16x8x2_t float16x8x2_t
__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:16518
Definition: arm_neon.h:325
__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:3571
__ai int16x4_t vmov_n_s16(int16_t __p0)
Definition: arm_neon.h:13843
__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0)
Definition: arm_neon.h:33323
__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:16092
__ai uint16x8_t vclzq_u16(uint16x8_t __p0)
Definition: arm_neon.h:3705
__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0)
Definition: arm_neon.h:31797
__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0)
Definition: arm_neon.h:32385
Definition: arm_neon.h:221
__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:15709
__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:19655
__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0)
Definition: arm_neon.h:30817
__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:15927
__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0)
Definition: arm_neon.h:16293
__ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2)
Definition: arm_neon.h:17119
__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:22931
__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:12960
__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0)
Definition: arm_neon.h:31461
__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:66977
__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:2948
__ai uint8x8_t vrev16_u8(uint8x8_t __p0)
Definition: arm_neon.h:20218
__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:14314
__ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:67583
__ai poly8x8_t vrev16_p8(poly8x8_t __p0)
Definition: arm_neon.h:20154
Definition: arm_neon.h:171
__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:28750
__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:14297
__ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:4106
struct uint64x1x2_t uint64x1x2_t
__ai int16x8_t vpaddlq_s8(int8x16_t __p0)
Definition: arm_neon.h:16325
__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0)
Definition: arm_neon.h:30187
__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:1018
__ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:69071
__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0)
Definition: arm_neon.h:30915
__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:29915
__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:11781
__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0)
Definition: arm_neon.h:31615
__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0)
Definition: arm_neon.h:30649
__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:20034
__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:22122
__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:16637
__ai poly16x4_t vmov_n_p16(poly16_t __p0)
Definition: arm_neon.h:13511
__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0)
Definition: arm_neon.h:33393
__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:7166
Definition: arm_neon.h:241
poly8x16_t val[2]
Definition: arm_neon.h:176
Definition: arm_neon.h:313
__ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:1283
__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0)
Definition: arm_neon.h:30103
__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:19784
__ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0)
Definition: arm_neon.h:13928
__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1)
Definition: arm_neon.h:5486
__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:20832
__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:67484
__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1)
Definition: arm_neon.h:18232
Definition: arm_neon.h:97
__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0)
Definition: arm_neon.h:31237
__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0)
Definition: arm_neon.h:32287
__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0)
Definition: arm_neon.h:30005
__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2812
__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:18248
__ai int32x4_t vqabsq_s32(int32x4_t __p0)
Definition: arm_neon.h:16721
__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2)
Definition: arm_neon.h:12834
__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0)
Definition: arm_neon.h:30733
__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:933
__ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:14848
__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:21070
poly16x4_t val[3]
Definition: arm_neon.h:296
__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:7268
__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:15811
__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28232
__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0)
Definition: arm_neon.h:30271
__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2)
Definition: arm_neon.h:66756
__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0)
Definition: arm_neon.h:30509
__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:16158
__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0)
Definition: arm_neon.h:30705
__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0)
Definition: arm_neon.h:30033
__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2)
Definition: arm_neon.h:12280
uint16x4_t val[4]
Definition: arm_neon.h:354
__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0)
Definition: arm_neon.h:30887
__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18849
__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:17029
struct float32x4x4_t float32x4x4_t
uint8x8_t val[3]
Definition: arm_neon.h:230
__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:12104
#define vshll_n_u8(__p0, __p1)
Definition: arm_neon.h:23329
__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:2574
__ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:513
__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:3605
__ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:69002
__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:22829
__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:19762
struct uint8x16x2_t uint8x16x2_t
__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0)
Definition: arm_neon.h:32077
struct float16x4x2_t float16x4x2_t
__ai poly16x4_t vcreate_p16(uint64_t __p0)
Definition: arm_neon.h:4215
__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:4179
__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:22812
__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:17065
__ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0)
Definition: arm_neon.h:17839
__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1)
Definition: arm_neon.h:28693
__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:654
__ai poly8x16_t vrev32q_p8(poly8x16_t __p0)
Definition: arm_neon.h:20282
__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:18282
struct uint16x8x4_t uint16x8x4_t
__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0)
Definition: arm_neon.h:30327
__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:15794
__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2)
Definition: arm_neon.h:67806
__ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:4150
__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1)
Definition: arm_neon.h:14688
__ai uint8x16_t vmvnq_u8(uint8x16_t __p0)
Definition: arm_neon.h:15106
__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0)
Definition: arm_neon.h:32273
__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:1747
__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:22897
__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:1648
__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1)
Definition: arm_neon.h:18764
__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:28921
float32x2_t val[3]
Definition: arm_neon.h:270
__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:12352
float32x2_t val[2]
Definition: arm_neon.h:154
__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:12298
__ai poly8x8_t vrev64_p8(poly8x8_t __p0)
Definition: arm_neon.h:20442
__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:3271
__ai uint8x8_t vdup_n_u8(uint8_t __p0)
Definition: arm_neon.h:5285
__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:13014
Definition: arm_neon.h:261
struct uint16x4x3_t uint16x4x3_t
__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:27761
__ai uint64x1_t vmov_n_u64(uint64_t __p0)
Definition: arm_neon.h:13738
__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2)
Definition: arm_neon.h:13343
__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0)
Definition: arm_neon.h:32049
__ai float32x4_t vdupq_n_f32(float32_t __p0)
Definition: arm_neon.h:5208
__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2)
Definition: arm_neon.h:67784
__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0)
Definition: arm_neon.h:33407
__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1)
Definition: arm_neon.h:11883
__ai int32x2_t vpaddl_s16(int16x4_t __p0)
Definition: arm_neon.h:16451
__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0)
Definition: arm_neon.h:31209
__ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:66880
__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2)
Definition: arm_neon.h:67828
__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0)
Definition: arm_neon.h:30719
__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0)
Definition: arm_neon.h:30453
__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0)
Definition: arm_neon.h:31517
__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:19888
__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2)
Definition: arm_neon.h:67440
__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:1631
Definition: arm_neon.h:329
Definition: arm_neon.h:205
__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:14127
__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:15828
__ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0)
Definition: arm_neon.h:17734
struct uint8x8x4_t uint8x8x4_t
__ai int16x8_t vrev32q_s16(int16x8_t __p0)
Definition: arm_neon.h:20362
__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1)
Definition: arm_neon.h:20883
__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1)
Definition: arm_neon.h:12019
__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2)
Definition: arm_neon.h:67597
__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2)
Definition: arm_neon.h:28411
__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2)
Definition: arm_neon.h:2027
__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0)
Definition: arm_neon.h:32973
__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0)
Definition: arm_neon.h:29963
__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0)
Definition: arm_neon.h:31783
__ai int64x1_t vpaddl_s32(int32x2_t __p0)
Definition: arm_neon.h:16436
__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1)
Definition: arm_neon.h:1733
__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0)
Definition: arm_neon.h:31825
__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0)
Definition: arm_neon.h:32133
__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:18730
__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0)
Definition: arm_neon.h:33015
__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:22056
__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:67096
__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:14796
__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1)
Definition: arm_neon.h:14110
__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1)
Definition: arm_neon.h:20781
__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1)
Definition: arm_neon.h:882
__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0)
Definition: arm_neon.h:33183
__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:28060
__ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0)
Definition: arm_neon.h:17902
Definition: arm_neon.h:179
__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1)
Definition: arm_neon.h:27795
__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1)
Definition: arm_neon.h:29478
__ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1)
Definition: arm_neon.h:640
__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1)
Definition: arm_neon.h:2982
__ai int32x2_t vclz_s32(int32x2_t __p0)
Definition: arm_neon.h:3833
__ai int64x1_t vget_low_s64(int64x2_t __p0)
Definition: arm_neon.h:7101
__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2)
Definition: arm_neon.h:6180
__ai uint32x2_t vget_low_u32(uint32x4_t __p0)
Definition: arm_neon.h:6990
__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1)
Definition: arm_neon.h:18299
__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:16552
__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:3656
int8_t poly8_t
Definition: arm_neon.h:45
__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0)
Definition: arm_neon.h:29935
__ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1)
Definition: arm_neon.h:66946
__ai float32x4_t vrsqrteq_f32(float32x4_t __p0)
Definition: arm_neon.h:21644
__ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1)
Definition: arm_neon.h:701
__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0)
Definition: arm_neon.h:30999
__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2)
Definition: arm_neon.h:28487
__ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0)
Definition: arm_neon.h:17881