gccxml_gem 0.9.2-x86-linux → 0.9.3-x86-linux

Sign up to get free protection for your applications and to get access to all the features.
Files changed (49) hide show
  1. data/Rakefile +15 -6
  2. data/bin/gccxml +0 -0
  3. data/bin/gccxml_cc1plus +0 -0
  4. data/gccxml.rb +5 -5
  5. data/share/gccxml-0.9/GCC/3.2/bits/gthr-default.h +4 -0
  6. data/share/gccxml-0.9/GCC/3.4/bits/gthr-default.h +5 -0
  7. data/share/gccxml-0.9/GCC/4.0/emmintrin.h +5 -0
  8. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_emmintrin.h +1037 -0
  9. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_mmintrin.h +669 -0
  10. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_xmmintrin.h +870 -0
  11. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_emmintrin.h +977 -0
  12. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_mmintrin.h +636 -0
  13. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_xmmintrin.h +833 -0
  14. data/share/gccxml-0.9/GCC/4.0/mmintrin.h +5 -0
  15. data/share/gccxml-0.9/GCC/4.0/xmmintrin.h +5 -0
  16. data/share/gccxml-0.9/GCC/4.1/bits/gthr-default.h +4 -0
  17. data/share/gccxml-0.9/GCC/4.1/emmintrin.h +5 -0
  18. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_emmintrin.h +1509 -0
  19. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_mmintrin.h +942 -0
  20. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_xmmintrin.h +1192 -0
  21. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_emmintrin.h +1004 -0
  22. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_mmintrin.h +637 -0
  23. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_xmmintrin.h +834 -0
  24. data/share/gccxml-0.9/GCC/4.1/mmintrin.h +5 -0
  25. data/share/gccxml-0.9/GCC/4.1/xmmintrin.h +5 -0
  26. data/share/gccxml-0.9/GCC/4.2/emmintrin.h +5 -0
  27. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_emmintrin.h +1509 -0
  28. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_mmintrin.h +942 -0
  29. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_xmmintrin.h +1192 -0
  30. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_emmintrin.h +1013 -0
  31. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_mmintrin.h +663 -0
  32. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_xmmintrin.h +860 -0
  33. data/share/gccxml-0.9/GCC/4.2/mmintrin.h +5 -0
  34. data/share/gccxml-0.9/GCC/4.2/xmmintrin.h +5 -0
  35. data/share/gccxml-0.9/GCC/4.3/emmintrin.h +1043 -0
  36. data/share/gccxml-0.9/GCC/4.3/gccxml_builtins.h +1 -0
  37. data/share/gccxml-0.9/GCC/4.3/mmintrin.h +663 -0
  38. data/share/gccxml-0.9/GCC/4.3/xmmintrin.h +867 -0
  39. data/share/gccxml-0.9/GCC/4.4/bits/c++config.h +1431 -0
  40. data/share/gccxml-0.9/GCC/4.4/emmintrin.h +1041 -0
  41. data/share/gccxml-0.9/GCC/4.4/gccxml_builtins.h +153 -0
  42. data/share/gccxml-0.9/GCC/4.4/mmintrin.h +662 -0
  43. data/share/gccxml-0.9/GCC/4.4/xmmintrin.h +864 -0
  44. data/share/gccxml-0.9/GCC/4.5/gccxml_builtins.h +154 -0
  45. data/share/gccxml-0.9/GCC/4.5/iomanip +349 -0
  46. data/share/gccxml-0.9/GCC/COPYING.RUNTIME +73 -0
  47. data/share/gccxml-0.9/GCC/COPYING3 +674 -0
  48. data/share/man/man1/gccxml.1 +1 -1
  49. metadata +165 -114
@@ -0,0 +1,977 @@
1
+ /* Copyright (C) 2003, 2004, 2005 Free Software Foundation, Inc.
2
+
3
+ This file is part of GCC.
4
+
5
+ GCC is free software; you can redistribute it and/or modify
6
+ it under the terms of the GNU General Public License as published by
7
+ the Free Software Foundation; either version 2, or (at your option)
8
+ any later version.
9
+
10
+ GCC is distributed in the hope that it will be useful,
11
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
12
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13
+ GNU General Public License for more details.
14
+
15
+ You should have received a copy of the GNU General Public License
16
+ along with GCC; see the file COPYING. If not, write to
17
+ the Free Software Foundation, 59 Temple Place - Suite 330,
18
+ Boston, MA 02111-1307, USA. */
19
+
20
+ /* As a special exception, if you include this header file into source
21
+ files compiled by GCC, this header file does not by itself cause
22
+ the resulting executable to be covered by the GNU General Public
23
+ License. This exception does not however invalidate any other
24
+ reasons why the executable file might be covered by the GNU General
25
+ Public License. */
26
+
27
+ /* Implemented from the specification included in the Intel C++ Compiler
28
+ User Guide and Reference, version 8.0. */
29
+
30
+ #ifndef _EMMINTRIN_H_INCLUDED
31
+ #define _EMMINTRIN_H_INCLUDED
32
+
33
+ #ifdef __SSE2__
34
+ #include <xmmintrin.h>
35
+
36
+ /* SSE2 */
37
+ typedef double __v2df __attribute__ ((__vector_size__ (16)));
38
+ typedef long long __v2di __attribute__ ((__vector_size__ (16)));
39
+ typedef int __v4si __attribute__ ((__vector_size__ (16)));
40
+ typedef short __v8hi __attribute__ ((__vector_size__ (16)));
41
+ typedef char __v16qi __attribute__ ((__vector_size__ (16)));
42
+
43
+ typedef __v2di __m128i;
44
+ typedef __v2df __m128d;
45
+
46
+ /* Create a selector for use with the SHUFPD instruction. */
47
+ #define _MM_SHUFFLE2(fp1,fp0) \
48
+ (((fp1) << 1) | (fp0))
49
+
50
+ /* Create a vector with element 0 as F and the rest zero. */
51
+ static __inline __m128d
52
+ _mm_set_sd (double __F)
53
+ ;
54
+
55
+ /* Create a vector with both elements equal to F. */
56
+ static __inline __m128d
57
+ _mm_set1_pd (double __F)
58
+ ;
59
+
60
+ static __inline __m128d
61
+ _mm_set_pd1 (double __F)
62
+ ;
63
+
64
+ /* Create a vector with the lower value X and upper value W. */
65
+ static __inline __m128d
66
+ _mm_set_pd (double __W, double __X)
67
+ ;
68
+
69
+ /* Create a vector with the lower value W and upper value X. */
70
+ static __inline __m128d
71
+ _mm_setr_pd (double __W, double __X)
72
+ ;
73
+
74
+ /* Create a vector of zeros. */
75
+ static __inline __m128d
76
+ _mm_setzero_pd (void)
77
+ ;
78
+
79
+ /* Sets the low DPFP value of A from the low value of B. */
80
+ static __inline __m128d
81
+ _mm_move_sd (__m128d __A, __m128d __B)
82
+ ;
83
+
84
+ /* Load two DPFP values from P. The address must be 16-byte aligned. */
85
+ static __inline __m128d
86
+ _mm_load_pd (double const *__P)
87
+ ;
88
+
89
+ /* Load two DPFP values from P. The address need not be 16-byte aligned. */
90
+ static __inline __m128d
91
+ _mm_loadu_pd (double const *__P)
92
+ ;
93
+
94
+ /* Create a vector with all two elements equal to *P. */
95
+ static __inline __m128d
96
+ _mm_load1_pd (double const *__P)
97
+ ;
98
+
99
+ /* Create a vector with element 0 as *P and the rest zero. */
100
+ static __inline __m128d
101
+ _mm_load_sd (double const *__P)
102
+ ;
103
+
104
+ static __inline __m128d
105
+ _mm_load_pd1 (double const *__P)
106
+ ;
107
+
108
+ /* Load two DPFP values in reverse order. The address must be aligned. */
109
+ static __inline __m128d
110
+ _mm_loadr_pd (double const *__P)
111
+ ;
112
+
113
+ /* Store two DPFP values. The address must be 16-byte aligned. */
114
+ static __inline void
115
+ _mm_store_pd (double *__P, __m128d __A)
116
+ ;
117
+
118
+ /* Store two DPFP values. The address need not be 16-byte aligned. */
119
+ static __inline void
120
+ _mm_storeu_pd (double *__P, __m128d __A)
121
+ ;
122
+
123
+ /* Stores the lower DPFP value. */
124
+ static __inline void
125
+ _mm_store_sd (double *__P, __m128d __A)
126
+ ;
127
+
128
+ static __inline void
129
+ _mm_storel_pd (double *__P, __m128d __A)
130
+ ;
131
+
132
+ /* Stores the upper DPFP value. */
133
+ static __inline void
134
+ _mm_storeh_pd (double *__P, __m128d __A)
135
+ ;
136
+
137
+ /* Store the lower DPFP value across two words.
138
+ The address must be 16-byte aligned. */
139
+ static __inline void
140
+ _mm_store1_pd (double *__P, __m128d __A)
141
+ ;
142
+
143
+ static __inline void
144
+ _mm_store_pd1 (double *__P, __m128d __A)
145
+ ;
146
+
147
+ /* Store two DPFP values in reverse order. The address must be aligned. */
148
+ static __inline void
149
+ _mm_storer_pd (double *__P, __m128d __A)
150
+ ;
151
+
152
+ static __inline int
153
+ _mm_cvtsi128_si32 (__m128i __A)
154
+ ;
155
+
156
+ #ifdef __x86_64__
157
+ static __inline long long
158
+ _mm_cvtsi128_si64x (__m128i __A)
159
+ ;
160
+ #endif
161
+
162
+ static __inline __m128d
163
+ _mm_add_pd (__m128d __A, __m128d __B)
164
+ ;
165
+
166
+ static __inline __m128d
167
+ _mm_add_sd (__m128d __A, __m128d __B)
168
+ ;
169
+
170
+ static __inline __m128d
171
+ _mm_sub_pd (__m128d __A, __m128d __B)
172
+ ;
173
+
174
+ static __inline __m128d
175
+ _mm_sub_sd (__m128d __A, __m128d __B)
176
+ ;
177
+
178
+ static __inline __m128d
179
+ _mm_mul_pd (__m128d __A, __m128d __B)
180
+ ;
181
+
182
+ static __inline __m128d
183
+ _mm_mul_sd (__m128d __A, __m128d __B)
184
+ ;
185
+
186
+ static __inline __m128d
187
+ _mm_div_pd (__m128d __A, __m128d __B)
188
+ ;
189
+
190
+ static __inline __m128d
191
+ _mm_div_sd (__m128d __A, __m128d __B)
192
+ ;
193
+
194
+ static __inline __m128d
195
+ _mm_sqrt_pd (__m128d __A)
196
+ ;
197
+
198
+ /* Return pair {sqrt (A[0), B[1]}. */
199
+ static __inline __m128d
200
+ _mm_sqrt_sd (__m128d __A, __m128d __B)
201
+ ;
202
+
203
+ static __inline __m128d
204
+ _mm_min_pd (__m128d __A, __m128d __B)
205
+ ;
206
+
207
+ static __inline __m128d
208
+ _mm_min_sd (__m128d __A, __m128d __B)
209
+ ;
210
+
211
+ static __inline __m128d
212
+ _mm_max_pd (__m128d __A, __m128d __B)
213
+ ;
214
+
215
+ static __inline __m128d
216
+ _mm_max_sd (__m128d __A, __m128d __B)
217
+ ;
218
+
219
+ static __inline __m128d
220
+ _mm_and_pd (__m128d __A, __m128d __B)
221
+ ;
222
+
223
+ static __inline __m128d
224
+ _mm_andnot_pd (__m128d __A, __m128d __B)
225
+ ;
226
+
227
+ static __inline __m128d
228
+ _mm_or_pd (__m128d __A, __m128d __B)
229
+ ;
230
+
231
+ static __inline __m128d
232
+ _mm_xor_pd (__m128d __A, __m128d __B)
233
+ ;
234
+
235
+ static __inline __m128d
236
+ _mm_cmpeq_pd (__m128d __A, __m128d __B)
237
+ ;
238
+
239
+ static __inline __m128d
240
+ _mm_cmplt_pd (__m128d __A, __m128d __B)
241
+ ;
242
+
243
+ static __inline __m128d
244
+ _mm_cmple_pd (__m128d __A, __m128d __B)
245
+ ;
246
+
247
+ static __inline __m128d
248
+ _mm_cmpgt_pd (__m128d __A, __m128d __B)
249
+ ;
250
+
251
+ static __inline __m128d
252
+ _mm_cmpge_pd (__m128d __A, __m128d __B)
253
+ ;
254
+
255
+ static __inline __m128d
256
+ _mm_cmpneq_pd (__m128d __A, __m128d __B)
257
+ ;
258
+
259
+ static __inline __m128d
260
+ _mm_cmpnlt_pd (__m128d __A, __m128d __B)
261
+ ;
262
+
263
+ static __inline __m128d
264
+ _mm_cmpnle_pd (__m128d __A, __m128d __B)
265
+ ;
266
+
267
+ static __inline __m128d
268
+ _mm_cmpngt_pd (__m128d __A, __m128d __B)
269
+ ;
270
+
271
+ static __inline __m128d
272
+ _mm_cmpnge_pd (__m128d __A, __m128d __B)
273
+ ;
274
+
275
+ static __inline __m128d
276
+ _mm_cmpord_pd (__m128d __A, __m128d __B)
277
+ ;
278
+
279
+ static __inline __m128d
280
+ _mm_cmpunord_pd (__m128d __A, __m128d __B)
281
+ ;
282
+
283
+ static __inline __m128d
284
+ _mm_cmpeq_sd (__m128d __A, __m128d __B)
285
+ ;
286
+
287
+ static __inline __m128d
288
+ _mm_cmplt_sd (__m128d __A, __m128d __B)
289
+ ;
290
+
291
+ static __inline __m128d
292
+ _mm_cmple_sd (__m128d __A, __m128d __B)
293
+ ;
294
+
295
+ static __inline __m128d
296
+ _mm_cmpgt_sd (__m128d __A, __m128d __B)
297
+ ;
298
+
299
+ static __inline __m128d
300
+ _mm_cmpge_sd (__m128d __A, __m128d __B)
301
+ ;
302
+
303
+ static __inline __m128d
304
+ _mm_cmpneq_sd (__m128d __A, __m128d __B)
305
+ ;
306
+
307
+ static __inline __m128d
308
+ _mm_cmpnlt_sd (__m128d __A, __m128d __B)
309
+ ;
310
+
311
+ static __inline __m128d
312
+ _mm_cmpnle_sd (__m128d __A, __m128d __B)
313
+ ;
314
+
315
+ static __inline __m128d
316
+ _mm_cmpngt_sd (__m128d __A, __m128d __B)
317
+ ;
318
+
319
+ static __inline __m128d
320
+ _mm_cmpnge_sd (__m128d __A, __m128d __B)
321
+ ;
322
+
323
+ static __inline __m128d
324
+ _mm_cmpord_sd (__m128d __A, __m128d __B)
325
+ ;
326
+
327
+ static __inline __m128d
328
+ _mm_cmpunord_sd (__m128d __A, __m128d __B)
329
+ ;
330
+
331
+ static __inline int
332
+ _mm_comieq_sd (__m128d __A, __m128d __B)
333
+ ;
334
+
335
+ static __inline int
336
+ _mm_comilt_sd (__m128d __A, __m128d __B)
337
+ ;
338
+
339
+ static __inline int
340
+ _mm_comile_sd (__m128d __A, __m128d __B)
341
+ ;
342
+
343
+ static __inline int
344
+ _mm_comigt_sd (__m128d __A, __m128d __B)
345
+ ;
346
+
347
+ static __inline int
348
+ _mm_comige_sd (__m128d __A, __m128d __B)
349
+ ;
350
+
351
+ static __inline int
352
+ _mm_comineq_sd (__m128d __A, __m128d __B)
353
+ ;
354
+
355
+ static __inline int
356
+ _mm_ucomieq_sd (__m128d __A, __m128d __B)
357
+ ;
358
+
359
+ static __inline int
360
+ _mm_ucomilt_sd (__m128d __A, __m128d __B)
361
+ ;
362
+
363
+ static __inline int
364
+ _mm_ucomile_sd (__m128d __A, __m128d __B)
365
+ ;
366
+
367
+ static __inline int
368
+ _mm_ucomigt_sd (__m128d __A, __m128d __B)
369
+ ;
370
+
371
+ static __inline int
372
+ _mm_ucomige_sd (__m128d __A, __m128d __B)
373
+ ;
374
+
375
+ static __inline int
376
+ _mm_ucomineq_sd (__m128d __A, __m128d __B)
377
+ ;
378
+
379
+ /* Create a vector of Qi, where i is the element number. */
380
+
381
+ static __inline __m128i
382
+ _mm_set_epi64x (long long __q1, long long __q0)
383
+ ;
384
+
385
+ static __inline __m128i
386
+ _mm_set_epi64 (__m64 __q1, __m64 __q0)
387
+ ;
388
+
389
+ static __inline __m128i
390
+ _mm_set_epi32 (int __q3, int __q2, int __q1, int __q0)
391
+ ;
392
+
393
+ static __inline __m128i
394
+ _mm_set_epi16 (short __q7, short __q6, short __q5, short __q4,
395
+ short __q3, short __q2, short __q1, short __q0)
396
+ ;
397
+
398
+ static __inline __m128i
399
+ _mm_set_epi8 (char __q15, char __q14, char __q13, char __q12,
400
+ char __q11, char __q10, char __q09, char __q08,
401
+ char __q07, char __q06, char __q05, char __q04,
402
+ char __q03, char __q02, char __q01, char __q00)
403
+ ;
404
+
405
+ /* Set all of the elements of the vector to A. */
406
+
407
+ static __inline __m128i
408
+ _mm_set1_epi64x (long long __A)
409
+ ;
410
+
411
+ static __inline __m128i
412
+ _mm_set1_epi64 (__m64 __A)
413
+ ;
414
+
415
+ static __inline __m128i
416
+ _mm_set1_epi32 (int __A)
417
+ ;
418
+
419
+ static __inline __m128i
420
+ _mm_set1_epi16 (short __A)
421
+ ;
422
+
423
+ static __inline __m128i
424
+ _mm_set1_epi8 (char __A)
425
+ ;
426
+
427
+ /* Create a vector of Qi, where i is the element number.
428
+ The parameter order is reversed from the _mm_set_epi* functions. */
429
+
430
+ static __inline __m128i
431
+ _mm_setr_epi64 (__m64 __q0, __m64 __q1)
432
+ ;
433
+
434
+ static __inline __m128i
435
+ _mm_setr_epi32 (int __q0, int __q1, int __q2, int __q3)
436
+ ;
437
+
438
+ static __inline __m128i
439
+ _mm_setr_epi16 (short __q0, short __q1, short __q2, short __q3,
440
+ short __q4, short __q5, short __q6, short __q7)
441
+ ;
442
+
443
+ static __inline __m128i
444
+ _mm_setr_epi8 (char __q00, char __q01, char __q02, char __q03,
445
+ char __q04, char __q05, char __q06, char __q07,
446
+ char __q08, char __q09, char __q10, char __q11,
447
+ char __q12, char __q13, char __q14, char __q15)
448
+ ;
449
+
450
+ /* Create a vector with element 0 as *P and the rest zero. */
451
+
452
+ static __inline __m128i
453
+ _mm_load_si128 (__m128i const *__P)
454
+ ;
455
+
456
+ static __inline __m128i
457
+ _mm_loadu_si128 (__m128i const *__P)
458
+ ;
459
+
460
+ static __inline __m128i
461
+ _mm_loadl_epi64 (__m128i const *__P)
462
+ ;
463
+
464
+ static __inline void
465
+ _mm_store_si128 (__m128i *__P, __m128i __B)
466
+ ;
467
+
468
+ static __inline void
469
+ _mm_storeu_si128 (__m128i *__P, __m128i __B)
470
+ ;
471
+
472
+ static __inline void
473
+ _mm_storel_epi64 (__m128i *__P, __m128i __B)
474
+ ;
475
+
476
+ static __inline __m64
477
+ _mm_movepi64_pi64 (__m128i __B)
478
+ ;
479
+
480
+ static __inline __m128i
481
+ _mm_movpi64_epi64 (__m64 __A)
482
+ ;
483
+
484
+ static __inline __m128i
485
+ _mm_move_epi64 (__m128i __A)
486
+ ;
487
+
488
+ /* Create a vector of zeros. */
489
+ static __inline __m128i
490
+ _mm_setzero_si128 (void)
491
+ ;
492
+
493
+ static __inline __m128d
494
+ _mm_cvtepi32_pd (__m128i __A)
495
+ ;
496
+
497
+ static __inline __m128
498
+ _mm_cvtepi32_ps (__m128i __A)
499
+ ;
500
+
501
+ static __inline __m128i
502
+ _mm_cvtpd_epi32 (__m128d __A)
503
+ ;
504
+
505
+ static __inline __m64
506
+ _mm_cvtpd_pi32 (__m128d __A)
507
+ ;
508
+
509
+ static __inline __m128
510
+ _mm_cvtpd_ps (__m128d __A)
511
+ ;
512
+
513
+ static __inline __m128i
514
+ _mm_cvttpd_epi32 (__m128d __A)
515
+ ;
516
+
517
+ static __inline __m64
518
+ _mm_cvttpd_pi32 (__m128d __A)
519
+ ;
520
+
521
+ static __inline __m128d
522
+ _mm_cvtpi32_pd (__m64 __A)
523
+ ;
524
+
525
+ static __inline __m128i
526
+ _mm_cvtps_epi32 (__m128 __A)
527
+ ;
528
+
529
+ static __inline __m128i
530
+ _mm_cvttps_epi32 (__m128 __A)
531
+ ;
532
+
533
+ static __inline __m128d
534
+ _mm_cvtps_pd (__m128 __A)
535
+ ;
536
+
537
+ static __inline int
538
+ _mm_cvtsd_si32 (__m128d __A)
539
+ ;
540
+
541
+ #ifdef __x86_64__
542
+ static __inline long long
543
+ _mm_cvtsd_si64x (__m128d __A)
544
+ ;
545
+ #endif
546
+
547
+ static __inline int
548
+ _mm_cvttsd_si32 (__m128d __A)
549
+ ;
550
+
551
+ #ifdef __x86_64__
552
+ static __inline long long
553
+ _mm_cvttsd_si64x (__m128d __A)
554
+ ;
555
+ #endif
556
+
557
+ static __inline __m128
558
+ _mm_cvtsd_ss (__m128 __A, __m128d __B)
559
+ ;
560
+
561
+ static __inline __m128d
562
+ _mm_cvtsi32_sd (__m128d __A, int __B)
563
+ ;
564
+
565
+ #ifdef __x86_64__
566
+ static __inline __m128d
567
+ _mm_cvtsi64x_sd (__m128d __A, long long __B)
568
+ ;
569
+ #endif
570
+
571
+ static __inline __m128d
572
+ _mm_cvtss_sd (__m128d __A, __m128 __B)
573
+ ;
574
+
575
+ #define _mm_shuffle_pd(__A, __B, __C) ((__m128d)__builtin_ia32_shufpd ((__v2df)__A, (__v2df)__B, (__C)))
576
+
577
+ static __inline __m128d
578
+ _mm_unpackhi_pd (__m128d __A, __m128d __B)
579
+ ;
580
+
581
+ static __inline __m128d
582
+ _mm_unpacklo_pd (__m128d __A, __m128d __B)
583
+ ;
584
+
585
+ static __inline __m128d
586
+ _mm_loadh_pd (__m128d __A, double const *__B)
587
+ ;
588
+
589
+ static __inline __m128d
590
+ _mm_loadl_pd (__m128d __A, double const *__B)
591
+ ;
592
+
593
+ static __inline int
594
+ _mm_movemask_pd (__m128d __A)
595
+ ;
596
+
597
+ static __inline __m128i
598
+ _mm_packs_epi16 (__m128i __A, __m128i __B)
599
+ ;
600
+
601
+ static __inline __m128i
602
+ _mm_packs_epi32 (__m128i __A, __m128i __B)
603
+ ;
604
+
605
+ static __inline __m128i
606
+ _mm_packus_epi16 (__m128i __A, __m128i __B)
607
+ ;
608
+
609
+ static __inline __m128i
610
+ _mm_unpackhi_epi8 (__m128i __A, __m128i __B)
611
+ ;
612
+
613
+ static __inline __m128i
614
+ _mm_unpackhi_epi16 (__m128i __A, __m128i __B)
615
+ ;
616
+
617
+ static __inline __m128i
618
+ _mm_unpackhi_epi32 (__m128i __A, __m128i __B)
619
+ ;
620
+
621
+ static __inline __m128i
622
+ _mm_unpackhi_epi64 (__m128i __A, __m128i __B)
623
+ ;
624
+
625
+ static __inline __m128i
626
+ _mm_unpacklo_epi8 (__m128i __A, __m128i __B)
627
+ ;
628
+
629
+ static __inline __m128i
630
+ _mm_unpacklo_epi16 (__m128i __A, __m128i __B)
631
+ ;
632
+
633
+ static __inline __m128i
634
+ _mm_unpacklo_epi32 (__m128i __A, __m128i __B)
635
+ ;
636
+
637
+ static __inline __m128i
638
+ _mm_unpacklo_epi64 (__m128i __A, __m128i __B)
639
+ ;
640
+
641
+ static __inline __m128i
642
+ _mm_add_epi8 (__m128i __A, __m128i __B)
643
+ ;
644
+
645
+ static __inline __m128i
646
+ _mm_add_epi16 (__m128i __A, __m128i __B)
647
+ ;
648
+
649
+ static __inline __m128i
650
+ _mm_add_epi32 (__m128i __A, __m128i __B)
651
+ ;
652
+
653
+ static __inline __m128i
654
+ _mm_add_epi64 (__m128i __A, __m128i __B)
655
+ ;
656
+
657
+ static __inline __m128i
658
+ _mm_adds_epi8 (__m128i __A, __m128i __B)
659
+ ;
660
+
661
+ static __inline __m128i
662
+ _mm_adds_epi16 (__m128i __A, __m128i __B)
663
+ ;
664
+
665
+ static __inline __m128i
666
+ _mm_adds_epu8 (__m128i __A, __m128i __B)
667
+ ;
668
+
669
+ static __inline __m128i
670
+ _mm_adds_epu16 (__m128i __A, __m128i __B)
671
+ ;
672
+
673
+ static __inline __m128i
674
+ _mm_sub_epi8 (__m128i __A, __m128i __B)
675
+ ;
676
+
677
+ static __inline __m128i
678
+ _mm_sub_epi16 (__m128i __A, __m128i __B)
679
+ ;
680
+
681
+ static __inline __m128i
682
+ _mm_sub_epi32 (__m128i __A, __m128i __B)
683
+ ;
684
+
685
+ static __inline __m128i
686
+ _mm_sub_epi64 (__m128i __A, __m128i __B)
687
+ ;
688
+
689
+ static __inline __m128i
690
+ _mm_subs_epi8 (__m128i __A, __m128i __B)
691
+ ;
692
+
693
+ static __inline __m128i
694
+ _mm_subs_epi16 (__m128i __A, __m128i __B)
695
+ ;
696
+
697
+ static __inline __m128i
698
+ _mm_subs_epu8 (__m128i __A, __m128i __B)
699
+ ;
700
+
701
+ static __inline __m128i
702
+ _mm_subs_epu16 (__m128i __A, __m128i __B)
703
+ ;
704
+
705
+ static __inline __m128i
706
+ _mm_madd_epi16 (__m128i __A, __m128i __B)
707
+ ;
708
+
709
+ static __inline __m128i
710
+ _mm_mulhi_epi16 (__m128i __A, __m128i __B)
711
+ ;
712
+
713
+ static __inline __m128i
714
+ _mm_mullo_epi16 (__m128i __A, __m128i __B)
715
+ ;
716
+
717
+ static __inline __m64
718
+ _mm_mul_su32 (__m64 __A, __m64 __B)
719
+ ;
720
+
721
+ static __inline __m128i
722
+ _mm_mul_epu32 (__m128i __A, __m128i __B)
723
+ ;
724
+
725
+ static __inline __m128i
726
+ _mm_slli_epi16 (__m128i __A, int __B)
727
+ ;
728
+
729
+ static __inline __m128i
730
+ _mm_slli_epi32 (__m128i __A, int __B)
731
+ ;
732
+
733
+ static __inline __m128i
734
+ _mm_slli_epi64 (__m128i __A, int __B)
735
+ ;
736
+
737
+ static __inline __m128i
738
+ _mm_srai_epi16 (__m128i __A, int __B)
739
+ ;
740
+
741
+ static __inline __m128i
742
+ _mm_srai_epi32 (__m128i __A, int __B)
743
+ ;
744
+
745
+ #if 0
746
+ static __m128i __attribute__((__always_inline__))
747
+ _mm_srli_si128 (__m128i __A, const int __B)
748
+ ;
749
+
750
+ static __m128i __attribute__((__always_inline__))
751
+ _mm_srli_si128 (__m128i __A, const int __B)
752
+ ;
753
+ #else
754
+ #define _mm_srli_si128(__A, __B) \
755
+ ((__m128i)__builtin_ia32_psrldqi128 (__A, (__B) * 8))
756
+ #define _mm_slli_si128(__A, __B) \
757
+ ((__m128i)__builtin_ia32_pslldqi128 (__A, (__B) * 8))
758
+ #endif
759
+
760
+ static __inline __m128i
761
+ _mm_srli_epi16 (__m128i __A, int __B)
762
+ ;
763
+
764
+ static __inline __m128i
765
+ _mm_srli_epi32 (__m128i __A, int __B)
766
+ ;
767
+
768
+ static __inline __m128i
769
+ _mm_srli_epi64 (__m128i __A, int __B)
770
+ ;
771
+
772
+ static __inline __m128i
773
+ _mm_sll_epi16 (__m128i __A, __m128i __B)
774
+ ;
775
+
776
+ static __inline __m128i
777
+ _mm_sll_epi32 (__m128i __A, __m128i __B)
778
+ ;
779
+
780
+ static __inline __m128i
781
+ _mm_sll_epi64 (__m128i __A, __m128i __B)
782
+ ;
783
+
784
+ static __inline __m128i
785
+ _mm_sra_epi16 (__m128i __A, __m128i __B)
786
+ ;
787
+
788
+ static __inline __m128i
789
+ _mm_sra_epi32 (__m128i __A, __m128i __B)
790
+ ;
791
+
792
+ static __inline __m128i
793
+ _mm_srl_epi16 (__m128i __A, __m128i __B)
794
+ ;
795
+
796
+ static __inline __m128i
797
+ _mm_srl_epi32 (__m128i __A, __m128i __B)
798
+ ;
799
+
800
+ static __inline __m128i
801
+ _mm_srl_epi64 (__m128i __A, __m128i __B)
802
+ ;
803
+
804
+ static __inline __m128i
805
+ _mm_and_si128 (__m128i __A, __m128i __B)
806
+ ;
807
+
808
+ static __inline __m128i
809
+ _mm_andnot_si128 (__m128i __A, __m128i __B)
810
+ ;
811
+
812
+ static __inline __m128i
813
+ _mm_or_si128 (__m128i __A, __m128i __B)
814
+ ;
815
+
816
+ static __inline __m128i
817
+ _mm_xor_si128 (__m128i __A, __m128i __B)
818
+ ;
819
+
820
+ static __inline __m128i
821
+ _mm_cmpeq_epi8 (__m128i __A, __m128i __B)
822
+ ;
823
+
824
+ static __inline __m128i
825
+ _mm_cmpeq_epi16 (__m128i __A, __m128i __B)
826
+ ;
827
+
828
+ static __inline __m128i
829
+ _mm_cmpeq_epi32 (__m128i __A, __m128i __B)
830
+ ;
831
+
832
+ static __inline __m128i
833
+ _mm_cmplt_epi8 (__m128i __A, __m128i __B)
834
+ ;
835
+
836
+ static __inline __m128i
837
+ _mm_cmplt_epi16 (__m128i __A, __m128i __B)
838
+ ;
839
+
840
+ static __inline __m128i
841
+ _mm_cmplt_epi32 (__m128i __A, __m128i __B)
842
+ ;
843
+
844
+ static __inline __m128i
845
+ _mm_cmpgt_epi8 (__m128i __A, __m128i __B)
846
+ ;
847
+
848
+ static __inline __m128i
849
+ _mm_cmpgt_epi16 (__m128i __A, __m128i __B)
850
+ ;
851
+
852
+ static __inline __m128i
853
+ _mm_cmpgt_epi32 (__m128i __A, __m128i __B)
854
+ ;
855
+
856
+ #if 0
857
+ static __inline int __attribute__((__always_inline__))
858
+ _mm_extract_epi16 (__m128i const __A, int const __N)
859
+ ;
860
+
861
+ static __inline __m128i __attribute__((__always_inline__))
862
+ _mm_insert_epi16 (__m128i const __A, int const __D, int const __N)
863
+ ;
864
+ #else
865
+ #define _mm_extract_epi16(A, N) \
866
+ ((int) __builtin_ia32_vec_ext_v8hi ((__v8hi)(A), (N)))
867
+ #define _mm_insert_epi16(A, D, N) \
868
+ ((__m128i) __builtin_ia32_vec_set_v8hi ((__v8hi)(A), (D), (N)))
869
+ #endif
870
+
871
+ static __inline __m128i
872
+ _mm_max_epi16 (__m128i __A, __m128i __B)
873
+ ;
874
+
875
+ static __inline __m128i
876
+ _mm_max_epu8 (__m128i __A, __m128i __B)
877
+ ;
878
+
879
+ static __inline __m128i
880
+ _mm_min_epi16 (__m128i __A, __m128i __B)
881
+ ;
882
+
883
+ static __inline __m128i
884
+ _mm_min_epu8 (__m128i __A, __m128i __B)
885
+ ;
886
+
887
+ static __inline int
888
+ _mm_movemask_epi8 (__m128i __A)
889
+ ;
890
+
891
+ static __inline __m128i
892
+ _mm_mulhi_epu16 (__m128i __A, __m128i __B)
893
+ ;
894
+
895
+ #define _mm_shufflehi_epi16(__A, __B) ((__m128i)__builtin_ia32_pshufhw ((__v8hi)__A, __B))
896
+ #define _mm_shufflelo_epi16(__A, __B) ((__m128i)__builtin_ia32_pshuflw ((__v8hi)__A, __B))
897
+ #define _mm_shuffle_epi32(__A, __B) ((__m128i)__builtin_ia32_pshufd ((__v4si)__A, __B))
898
+
899
+ static __inline void
900
+ _mm_maskmoveu_si128 (__m128i __A, __m128i __B, char *__C)
901
+ ;
902
+
903
+ static __inline __m128i
904
+ _mm_avg_epu8 (__m128i __A, __m128i __B)
905
+ ;
906
+
907
+ static __inline __m128i
908
+ _mm_avg_epu16 (__m128i __A, __m128i __B)
909
+ ;
910
+
911
+ static __inline __m128i
912
+ _mm_sad_epu8 (__m128i __A, __m128i __B)
913
+ ;
914
+
915
+ static __inline void
916
+ _mm_stream_si32 (int *__A, int __B)
917
+ ;
918
+
919
+ static __inline void
920
+ _mm_stream_si128 (__m128i *__A, __m128i __B)
921
+ ;
922
+
923
+ static __inline void
924
+ _mm_stream_pd (double *__A, __m128d __B)
925
+ ;
926
+
927
+ static __inline void
928
+ _mm_clflush (void const *__A)
929
+ ;
930
+
931
+ static __inline void
932
+ _mm_lfence (void)
933
+ ;
934
+
935
+ static __inline void
936
+ _mm_mfence (void)
937
+ ;
938
+
939
+ static __inline __m128i
940
+ _mm_cvtsi32_si128 (int __A)
941
+ ;
942
+
943
+ #ifdef __x86_64__
944
+ static __inline __m128i
945
+ _mm_cvtsi64x_si128 (long long __A)
946
+ ;
947
+ #endif
948
+
949
+ /* Casts between various SP, DP, INT vector types. Note that these do no
950
+ conversion of values, they just change the type. */
951
+ static __inline __m128
952
+ _mm_castpd_ps(__m128d __A)
953
+ ;
954
+
955
+ static __inline __m128i
956
+ _mm_castpd_si128(__m128d __A)
957
+ ;
958
+
959
+ static __inline __m128d
960
+ _mm_castps_pd(__m128 __A)
961
+ ;
962
+
963
+ static __inline __m128i
964
+ _mm_castps_si128(__m128 __A)
965
+ ;
966
+
967
+ static __inline __m128
968
+ _mm_castsi128_ps(__m128i __A)
969
+ ;
970
+
971
+ static __inline __m128d
972
+ _mm_castsi128_pd(__m128i __A)
973
+ ;
974
+
975
+ #endif /* __SSE2__ */
976
+
977
+ #endif /* _EMMINTRIN_H_INCLUDED */