gccxml_gem 0.9.2-x86-linux → 0.9.3-x86-linux

Sign up to get free protection for your applications and to get access to all the features.
Files changed (49) hide show
  1. data/Rakefile +15 -6
  2. data/bin/gccxml +0 -0
  3. data/bin/gccxml_cc1plus +0 -0
  4. data/gccxml.rb +5 -5
  5. data/share/gccxml-0.9/GCC/3.2/bits/gthr-default.h +4 -0
  6. data/share/gccxml-0.9/GCC/3.4/bits/gthr-default.h +5 -0
  7. data/share/gccxml-0.9/GCC/4.0/emmintrin.h +5 -0
  8. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_emmintrin.h +1037 -0
  9. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_mmintrin.h +669 -0
  10. data/share/gccxml-0.9/GCC/4.0/gccxml_apple_xmmintrin.h +870 -0
  11. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_emmintrin.h +977 -0
  12. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_mmintrin.h +636 -0
  13. data/share/gccxml-0.9/GCC/4.0/gccxml_gnu_xmmintrin.h +833 -0
  14. data/share/gccxml-0.9/GCC/4.0/mmintrin.h +5 -0
  15. data/share/gccxml-0.9/GCC/4.0/xmmintrin.h +5 -0
  16. data/share/gccxml-0.9/GCC/4.1/bits/gthr-default.h +4 -0
  17. data/share/gccxml-0.9/GCC/4.1/emmintrin.h +5 -0
  18. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_emmintrin.h +1509 -0
  19. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_mmintrin.h +942 -0
  20. data/share/gccxml-0.9/GCC/4.1/gccxml_apple_xmmintrin.h +1192 -0
  21. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_emmintrin.h +1004 -0
  22. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_mmintrin.h +637 -0
  23. data/share/gccxml-0.9/GCC/4.1/gccxml_gnu_xmmintrin.h +834 -0
  24. data/share/gccxml-0.9/GCC/4.1/mmintrin.h +5 -0
  25. data/share/gccxml-0.9/GCC/4.1/xmmintrin.h +5 -0
  26. data/share/gccxml-0.9/GCC/4.2/emmintrin.h +5 -0
  27. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_emmintrin.h +1509 -0
  28. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_mmintrin.h +942 -0
  29. data/share/gccxml-0.9/GCC/4.2/gccxml_apple_xmmintrin.h +1192 -0
  30. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_emmintrin.h +1013 -0
  31. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_mmintrin.h +663 -0
  32. data/share/gccxml-0.9/GCC/4.2/gccxml_gnu_xmmintrin.h +860 -0
  33. data/share/gccxml-0.9/GCC/4.2/mmintrin.h +5 -0
  34. data/share/gccxml-0.9/GCC/4.2/xmmintrin.h +5 -0
  35. data/share/gccxml-0.9/GCC/4.3/emmintrin.h +1043 -0
  36. data/share/gccxml-0.9/GCC/4.3/gccxml_builtins.h +1 -0
  37. data/share/gccxml-0.9/GCC/4.3/mmintrin.h +663 -0
  38. data/share/gccxml-0.9/GCC/4.3/xmmintrin.h +867 -0
  39. data/share/gccxml-0.9/GCC/4.4/bits/c++config.h +1431 -0
  40. data/share/gccxml-0.9/GCC/4.4/emmintrin.h +1041 -0
  41. data/share/gccxml-0.9/GCC/4.4/gccxml_builtins.h +153 -0
  42. data/share/gccxml-0.9/GCC/4.4/mmintrin.h +662 -0
  43. data/share/gccxml-0.9/GCC/4.4/xmmintrin.h +864 -0
  44. data/share/gccxml-0.9/GCC/4.5/gccxml_builtins.h +154 -0
  45. data/share/gccxml-0.9/GCC/4.5/iomanip +349 -0
  46. data/share/gccxml-0.9/GCC/COPYING.RUNTIME +73 -0
  47. data/share/gccxml-0.9/GCC/COPYING3 +674 -0
  48. data/share/man/man1/gccxml.1 +1 -1
  49. metadata +165 -114
data/Rakefile CHANGED
@@ -3,12 +3,16 @@ require 'rake/rdoctask'
3
3
  require 'rake/gempackagetask'
4
4
 
5
5
  PROJECT_NAME = "gccxml_gem"
6
- GCCXML_VERSION = "0.9.2"
6
+ GCCXML_VERSION = "0.9.3"
7
7
  RUBYFORGE_USERNAME = "jameskilton"
8
8
 
9
9
  desc "Build gccxml for this system"
10
10
  task :build_gccxml => [:clean, :unpack, :build, :install]
11
11
 
12
+ def make_cmd
13
+ PLATFORM =~ /mswin/ ? "mingw32-make" : "make"
14
+ end
15
+
12
16
  task :unpack do
13
17
  cd "ext" do
14
18
  sh "tar xzvf gccxml.tar.gz"
@@ -18,19 +22,26 @@ end
18
22
 
19
23
  task :build do
20
24
  install_path = File.expand_path(File.dirname(__FILE__))
25
+
26
+ platform = PLATFORM =~ /mswin/ ? "-G \"MinGW Makefiles\"" : ""
27
+
21
28
  cd "ext/gccxml-build" do
22
- sh "cmake -DCMAKE_INSTALL_PREFIX:PATH=#{install_path} ../gccxml"
23
- sh "make"
29
+ sh "cmake -DCMAKE_INSTALL_PREFIX:PATH=#{install_path} #{platform} ../gccxml"
30
+ sh make_cmd
24
31
  end
25
32
  end
26
33
 
27
34
  task :install do
28
35
  cd "ext/gccxml-build" do
29
- sh "make install"
36
+ sh "#{make_cmd} install"
30
37
  end
31
38
 
32
39
  sh "chmod a+x bin/*"
33
40
  sh "chmod -R a+rx share/"
41
+
42
+ cd "bin" do
43
+ `cp gccxml_cc1plus.exe gccxml_cc1plus` if File.exists?("gccxml_cc1plus.exe")
44
+ end
34
45
  end
35
46
 
36
47
  desc "Clean up everything"
@@ -70,6 +81,4 @@ and using RbGCCXML.
70
81
  end
71
82
 
72
83
  Rake::GemPackageTask.new(spec) do |pkg|
73
- pkg.need_zip = true
74
- pkg.need_tar = true
75
84
  end
data/bin/gccxml CHANGED
Binary file
Binary file
data/gccxml.rb CHANGED
@@ -31,12 +31,12 @@ class GCCXML
31
31
 
32
32
  private
33
33
 
34
+ def windows?
35
+ RUBY_PLATFORM =~ /(mswin|cygwin)/
36
+ end
37
+
34
38
  def find_exe
35
- ext = if RUBY_PLATFORM =~ /(mswin|cygwin)/
36
- ".exe"
37
- else
38
- ""
39
- end
39
+ ext = windows? ? ".exe" : ""
40
40
 
41
41
  path = File.expand_path(File.join(File.dirname(__FILE__), "bin", "gccxml#{ext}"))
42
42
  path.chomp!
@@ -1,3 +1,6 @@
1
+ #if defined(__MINGW32__)
2
+ # include_next <bits/gthr-default.h>
3
+ #else
1
4
  /* Threads compatibility routines for libgcc2 and libobjc. */
2
5
  /* Compile this one with gcc. */
3
6
  /* Copyright (C) 1997, 1999, 2000, 2001 Free Software Foundation, Inc.
@@ -579,3 +582,4 @@ __gthread_mutex_unlock (__gthread_mutex_t *mutex)
579
582
  #endif /* _LIBOBJC */
580
583
 
581
584
  #endif /* ! _GLIBCPP_GCC_GTHR_POSIX_H */
585
+ #endif
@@ -1,3 +1,6 @@
1
+ #if defined(__MINGW32__)
2
+ # include_next <bits/gthr-default.h>
3
+ #else
1
4
  /* Threads compatibility routines for libgcc2 and libobjc. */
2
5
  /* Compile this one with gcc. */
3
6
  /* Copyright (C) 1997, 1999, 2000, 2001, 2002, 2003
@@ -662,3 +665,5 @@ __gthread_recursive_mutex_unlock (__gthread_recursive_mutex_t *mutex)
662
665
  #endif /* _LIBOBJC */
663
666
 
664
667
  #endif /* ! _GLIBCXX_GCC_GTHR_POSIX_H */
668
+ #endif
669
+
@@ -0,0 +1,5 @@
1
+ #ifdef __APPLE__
2
+ # include "gccxml_apple_emmintrin.h"
3
+ #else
4
+ # include "gccxml_gnu_emmintrin.h"
5
+ #endif
@@ -0,0 +1,1037 @@
1
+ /* APPLE LOCAL file mainline 2005-06-30 Radar 4131077 */
2
+ /* Copyright (C) 2003, 2004, 2005, 2006 Free Software Foundation, Inc.
3
+
4
+ This file is part of GCC.
5
+
6
+ GCC is free software; you can redistribute it and/or modify
7
+ it under the terms of the GNU General Public License as published by
8
+ the Free Software Foundation; either version 2, or (at your option)
9
+ any later version.
10
+
11
+ GCC is distributed in the hope that it will be useful,
12
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
13
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14
+ GNU General Public License for more details.
15
+
16
+ You should have received a copy of the GNU General Public License
17
+ along with GCC; see the file COPYING. If not, write to
18
+ the Free Software Foundation, 59 Temple Place - Suite 330,
19
+ Boston, MA 02111-1307, USA. */
20
+
21
+ /* As a special exception, if you include this header file into source
22
+ files compiled by GCC, this header file does not by itself cause
23
+ the resulting executable to be covered by the GNU General Public
24
+ License. This exception does not however invalidate any other
25
+ reasons why the executable file might be covered by the GNU General
26
+ Public License. */
27
+
28
+ /* Implemented from the specification included in the Intel C++ Compiler
29
+ User Guide and Reference, version 9.0. */
30
+
31
+ #ifndef _EMMINTRIN_H_INCLUDED
32
+ #define _EMMINTRIN_H_INCLUDED
33
+
34
+ #ifdef __SSE2__
35
+ #include <xmmintrin.h>
36
+
37
+ /* SSE2 */
38
+ typedef double __v2df __attribute__ ((__vector_size__ (16)));
39
+ typedef long long __v2di __attribute__ ((__vector_size__ (16)));
40
+ typedef int __v4si __attribute__ ((__vector_size__ (16)));
41
+ typedef short __v8hi __attribute__ ((__vector_size__ (16)));
42
+ typedef char __v16qi __attribute__ ((__vector_size__ (16)));
43
+
44
+ typedef __v2di __m128i;
45
+ typedef __v2df __m128d;
46
+
47
+ /* Create a selector for use with the SHUFPD instruction. */
48
+ #define _MM_SHUFFLE2(fp1,fp0) \
49
+ (((fp1) << 1) | (fp0))
50
+
51
+ /* APPLE LOCAL begin nodebug inline 4152603 */
52
+ #define __always_inline__ __always_inline__, __nodebug__
53
+ /* APPLE LOCAL end nodebug inline 4152603 */
54
+
55
+ /* APPLE LOCAL begin radar 4152603 */
56
+ /* Create a vector with element 0 as F and the rest zero. */
57
+ static __inline __m128d __attribute__((__always_inline__))
58
+ _mm_set_sd (double __F)
59
+ ;
60
+
61
+ /* Create a vector with both elements equal to F. */
62
+ static __inline __m128d __attribute__((__always_inline__))
63
+ _mm_set1_pd (double __F)
64
+ ;
65
+
66
+ static __inline __m128d __attribute__((__always_inline__))
67
+ _mm_set_pd1 (double __F)
68
+ ;
69
+
70
+ /* Create a vector with the lower value X and upper value W. */
71
+ static __inline __m128d __attribute__((__always_inline__))
72
+ _mm_set_pd (double __W, double __X)
73
+ ;
74
+
75
+ /* Create a vector with the lower value W and upper value X. */
76
+ static __inline __m128d __attribute__((__always_inline__))
77
+ _mm_setr_pd (double __W, double __X)
78
+ ;
79
+
80
+ /* Create a vector of zeros. */
81
+ static __inline __m128d __attribute__((__always_inline__))
82
+ _mm_setzero_pd (void)
83
+ ;
84
+
85
+ /* Sets the low DPFP value of A from the low value of B. */
86
+ static __inline __m128d __attribute__((__always_inline__))
87
+ _mm_move_sd (__m128d __A, __m128d __B)
88
+ ;
89
+
90
+ /* Load two DPFP values from P. The address must be 16-byte aligned. */
91
+ static __inline __m128d __attribute__((__always_inline__))
92
+ _mm_load_pd (double const *__P)
93
+ ;
94
+
95
+ /* Load two DPFP values from P. The address need not be 16-byte aligned. */
96
+ static __inline __m128d __attribute__((__always_inline__))
97
+ _mm_loadu_pd (double const *__P)
98
+ ;
99
+
100
+ /* Create a vector with all two elements equal to *P. */
101
+ static __inline __m128d __attribute__((__always_inline__))
102
+ _mm_load1_pd (double const *__P)
103
+ ;
104
+
105
+ /* Create a vector with element 0 as *P and the rest zero. */
106
+ static __inline __m128d __attribute__((__always_inline__))
107
+ _mm_load_sd (double const *__P)
108
+ ;
109
+
110
+ static __inline __m128d __attribute__((__always_inline__))
111
+ _mm_load_pd1 (double const *__P)
112
+ ;
113
+
114
+ /* Load two DPFP values in reverse order. The address must be aligned. */
115
+ static __inline __m128d __attribute__((__always_inline__))
116
+ _mm_loadr_pd (double const *__P)
117
+ ;
118
+
119
+ /* Store two DPFP values. The address must be 16-byte aligned. */
120
+ static __inline void __attribute__((__always_inline__))
121
+ _mm_store_pd (double *__P, __m128d __A)
122
+ ;
123
+
124
+ /* Store two DPFP values. The address need not be 16-byte aligned. */
125
+ static __inline void __attribute__((__always_inline__))
126
+ _mm_storeu_pd (double *__P, __m128d __A)
127
+ ;
128
+
129
+ /* Stores the lower DPFP value. */
130
+ static __inline void __attribute__((__always_inline__))
131
+ _mm_store_sd (double *__P, __m128d __A)
132
+ ;
133
+
134
+ static __inline double __attribute__((__always_inline__))
135
+ _mm_cvtsd_f64 (__m128d __A)
136
+ ;
137
+
138
+ static __inline void __attribute__((__always_inline__))
139
+ _mm_storel_pd (double *__P, __m128d __A)
140
+ ;
141
+
142
+ /* Stores the upper DPFP value. */
143
+ static __inline void __attribute__((__always_inline__))
144
+ _mm_storeh_pd (double *__P, __m128d __A)
145
+ ;
146
+
147
+ /* Store the lower DPFP value across two words.
148
+ The address must be 16-byte aligned. */
149
+ static __inline void __attribute__((__always_inline__))
150
+ _mm_store1_pd (double *__P, __m128d __A)
151
+ ;
152
+
153
+ static __inline void __attribute__((__always_inline__))
154
+ _mm_store_pd1 (double *__P, __m128d __A)
155
+ ;
156
+
157
+ /* Store two DPFP values in reverse order. The address must be aligned. */
158
+ static __inline void __attribute__((__always_inline__))
159
+ _mm_storer_pd (double *__P, __m128d __A)
160
+ ;
161
+
162
+ static __inline int __attribute__((__always_inline__))
163
+ _mm_cvtsi128_si32 (__m128i __A)
164
+ ;
165
+
166
+ #ifdef __x86_64__
167
+ /* Intel intrinsic. */
168
+ static __inline long long __attribute__((__always_inline__))
169
+ _mm_cvtsi128_si64 (__m128i __A)
170
+ ;
171
+
172
+ /* Microsoft intrinsic. */
173
+ static __inline long long __attribute__((__always_inline__))
174
+ _mm_cvtsi128_si64x (__m128i __A)
175
+ ;
176
+ #endif
177
+
178
+ static __inline __m128d __attribute__((__always_inline__))
179
+ _mm_add_pd (__m128d __A, __m128d __B)
180
+ ;
181
+
182
+ static __inline __m128d __attribute__((__always_inline__))
183
+ _mm_add_sd (__m128d __A, __m128d __B)
184
+ ;
185
+
186
+ static __inline __m128d __attribute__((__always_inline__))
187
+ _mm_sub_pd (__m128d __A, __m128d __B)
188
+ ;
189
+
190
+ static __inline __m128d __attribute__((__always_inline__))
191
+ _mm_sub_sd (__m128d __A, __m128d __B)
192
+ ;
193
+
194
+ static __inline __m128d __attribute__((__always_inline__))
195
+ _mm_mul_pd (__m128d __A, __m128d __B)
196
+ ;
197
+
198
+ static __inline __m128d __attribute__((__always_inline__))
199
+ _mm_mul_sd (__m128d __A, __m128d __B)
200
+ ;
201
+
202
+ static __inline __m128d __attribute__((__always_inline__))
203
+ _mm_div_pd (__m128d __A, __m128d __B)
204
+ ;
205
+
206
+ static __inline __m128d __attribute__((__always_inline__))
207
+ _mm_div_sd (__m128d __A, __m128d __B)
208
+ ;
209
+
210
+ static __inline __m128d __attribute__((__always_inline__))
211
+ _mm_sqrt_pd (__m128d __A)
212
+ ;
213
+
214
+ /* Return pair {sqrt (A[0), B[1]}. */
215
+ static __inline __m128d __attribute__((__always_inline__))
216
+ _mm_sqrt_sd (__m128d __A, __m128d __B)
217
+ ;
218
+
219
+ static __inline __m128d __attribute__((__always_inline__))
220
+ _mm_min_pd (__m128d __A, __m128d __B)
221
+ ;
222
+
223
+ static __inline __m128d __attribute__((__always_inline__))
224
+ _mm_min_sd (__m128d __A, __m128d __B)
225
+ ;
226
+
227
+ static __inline __m128d __attribute__((__always_inline__))
228
+ _mm_max_pd (__m128d __A, __m128d __B)
229
+ ;
230
+
231
+ static __inline __m128d __attribute__((__always_inline__))
232
+ _mm_max_sd (__m128d __A, __m128d __B)
233
+ ;
234
+
235
+ static __inline __m128d __attribute__((__always_inline__))
236
+ _mm_and_pd (__m128d __A, __m128d __B)
237
+ ;
238
+
239
+ static __inline __m128d __attribute__((__always_inline__))
240
+ _mm_andnot_pd (__m128d __A, __m128d __B)
241
+ ;
242
+
243
+ static __inline __m128d __attribute__((__always_inline__))
244
+ _mm_or_pd (__m128d __A, __m128d __B)
245
+ ;
246
+
247
+ static __inline __m128d __attribute__((__always_inline__))
248
+ _mm_xor_pd (__m128d __A, __m128d __B)
249
+ ;
250
+
251
+ static __inline __m128d __attribute__((__always_inline__))
252
+ _mm_cmpeq_pd (__m128d __A, __m128d __B)
253
+ ;
254
+
255
+ static __inline __m128d __attribute__((__always_inline__))
256
+ _mm_cmplt_pd (__m128d __A, __m128d __B)
257
+ ;
258
+
259
+ static __inline __m128d __attribute__((__always_inline__))
260
+ _mm_cmple_pd (__m128d __A, __m128d __B)
261
+ ;
262
+
263
+ static __inline __m128d __attribute__((__always_inline__))
264
+ _mm_cmpgt_pd (__m128d __A, __m128d __B)
265
+ ;
266
+
267
+ static __inline __m128d __attribute__((__always_inline__))
268
+ _mm_cmpge_pd (__m128d __A, __m128d __B)
269
+ ;
270
+
271
+ static __inline __m128d __attribute__((__always_inline__))
272
+ _mm_cmpneq_pd (__m128d __A, __m128d __B)
273
+ ;
274
+
275
+ static __inline __m128d __attribute__((__always_inline__))
276
+ _mm_cmpnlt_pd (__m128d __A, __m128d __B)
277
+ ;
278
+
279
+ static __inline __m128d __attribute__((__always_inline__))
280
+ _mm_cmpnle_pd (__m128d __A, __m128d __B)
281
+ ;
282
+
283
+ static __inline __m128d __attribute__((__always_inline__))
284
+ _mm_cmpngt_pd (__m128d __A, __m128d __B)
285
+ ;
286
+
287
+ static __inline __m128d __attribute__((__always_inline__))
288
+ _mm_cmpnge_pd (__m128d __A, __m128d __B)
289
+ ;
290
+
291
+ static __inline __m128d __attribute__((__always_inline__))
292
+ _mm_cmpord_pd (__m128d __A, __m128d __B)
293
+ ;
294
+
295
+ static __inline __m128d __attribute__((__always_inline__))
296
+ _mm_cmpunord_pd (__m128d __A, __m128d __B)
297
+ ;
298
+
299
+ static __inline __m128d __attribute__((__always_inline__))
300
+ _mm_cmpeq_sd (__m128d __A, __m128d __B)
301
+ ;
302
+
303
+ static __inline __m128d __attribute__((__always_inline__))
304
+ _mm_cmplt_sd (__m128d __A, __m128d __B)
305
+ ;
306
+
307
+ static __inline __m128d __attribute__((__always_inline__))
308
+ _mm_cmple_sd (__m128d __A, __m128d __B)
309
+ ;
310
+
311
+ static __inline __m128d __attribute__((__always_inline__))
312
+ _mm_cmpgt_sd (__m128d __A, __m128d __B)
313
+ ;
314
+
315
+ static __inline __m128d __attribute__((__always_inline__))
316
+ _mm_cmpge_sd (__m128d __A, __m128d __B)
317
+ ;
318
+
319
+ static __inline __m128d __attribute__((__always_inline__))
320
+ _mm_cmpneq_sd (__m128d __A, __m128d __B)
321
+ ;
322
+
323
+ static __inline __m128d __attribute__((__always_inline__))
324
+ _mm_cmpnlt_sd (__m128d __A, __m128d __B)
325
+ ;
326
+
327
+ static __inline __m128d __attribute__((__always_inline__))
328
+ _mm_cmpnle_sd (__m128d __A, __m128d __B)
329
+ ;
330
+
331
+ static __inline __m128d __attribute__((__always_inline__))
332
+ _mm_cmpngt_sd (__m128d __A, __m128d __B)
333
+ ;
334
+
335
+ static __inline __m128d __attribute__((__always_inline__))
336
+ _mm_cmpnge_sd (__m128d __A, __m128d __B)
337
+ ;
338
+
339
+ static __inline __m128d __attribute__((__always_inline__))
340
+ _mm_cmpord_sd (__m128d __A, __m128d __B)
341
+ ;
342
+
343
+ static __inline __m128d __attribute__((__always_inline__))
344
+ _mm_cmpunord_sd (__m128d __A, __m128d __B)
345
+ ;
346
+
347
+ static __inline int __attribute__((__always_inline__))
348
+ _mm_comieq_sd (__m128d __A, __m128d __B)
349
+ ;
350
+
351
+ static __inline int __attribute__((__always_inline__))
352
+ _mm_comilt_sd (__m128d __A, __m128d __B)
353
+ ;
354
+
355
+ static __inline int __attribute__((__always_inline__))
356
+ _mm_comile_sd (__m128d __A, __m128d __B)
357
+ ;
358
+
359
+ static __inline int __attribute__((__always_inline__))
360
+ _mm_comigt_sd (__m128d __A, __m128d __B)
361
+ ;
362
+
363
+ static __inline int __attribute__((__always_inline__))
364
+ _mm_comige_sd (__m128d __A, __m128d __B)
365
+ ;
366
+
367
+ static __inline int __attribute__((__always_inline__))
368
+ _mm_comineq_sd (__m128d __A, __m128d __B)
369
+ ;
370
+
371
+ static __inline int __attribute__((__always_inline__))
372
+ _mm_ucomieq_sd (__m128d __A, __m128d __B)
373
+ ;
374
+
375
+ static __inline int __attribute__((__always_inline__))
376
+ _mm_ucomilt_sd (__m128d __A, __m128d __B)
377
+ ;
378
+
379
+ static __inline int __attribute__((__always_inline__))
380
+ _mm_ucomile_sd (__m128d __A, __m128d __B)
381
+ ;
382
+
383
+ static __inline int __attribute__((__always_inline__))
384
+ _mm_ucomigt_sd (__m128d __A, __m128d __B)
385
+ ;
386
+
387
+ static __inline int __attribute__((__always_inline__))
388
+ _mm_ucomige_sd (__m128d __A, __m128d __B)
389
+ ;
390
+
391
+ static __inline int __attribute__((__always_inline__))
392
+ _mm_ucomineq_sd (__m128d __A, __m128d __B)
393
+ ;
394
+
395
+ /* Create a vector of Qi, where i is the element number. */
396
+
397
+ static __inline __m128i __attribute__((__always_inline__))
398
+ _mm_set_epi64x (long long __q1, long long __q0)
399
+ ;
400
+
401
+ static __inline __m128i __attribute__((__always_inline__))
402
+ _mm_set_epi64 (__m64 __q1, __m64 __q0)
403
+ ;
404
+
405
+ static __inline __m128i __attribute__((__always_inline__))
406
+ _mm_set_epi32 (int __q3, int __q2, int __q1, int __q0)
407
+ ;
408
+
409
+ static __inline __m128i __attribute__((__always_inline__))
410
+ _mm_set_epi16 (short __q7, short __q6, short __q5, short __q4,
411
+ short __q3, short __q2, short __q1, short __q0)
412
+ ;
413
+
414
+ static __inline __m128i __attribute__((__always_inline__))
415
+ _mm_set_epi8 (char __q15, char __q14, char __q13, char __q12,
416
+ char __q11, char __q10, char __q09, char __q08,
417
+ char __q07, char __q06, char __q05, char __q04,
418
+ char __q03, char __q02, char __q01, char __q00)
419
+ ;
420
+
421
+ /* APPLE LOCAL begin 4220129 */
422
+ /* functions moved to end of file */
423
+ /* APPLE LOCAL end 4220129 */
424
+
425
+ /* Create a vector of Qi, where i is the element number.
426
+ The parameter order is reversed from the _mm_set_epi* functions. */
427
+
428
+ static __inline __m128i __attribute__((__always_inline__))
429
+ _mm_setr_epi64 (__m64 __q0, __m64 __q1)
430
+ ;
431
+
432
+ static __inline __m128i __attribute__((__always_inline__))
433
+ _mm_setr_epi32 (int __q0, int __q1, int __q2, int __q3)
434
+ ;
435
+
436
+ static __inline __m128i __attribute__((__always_inline__))
437
+ _mm_setr_epi16 (short __q0, short __q1, short __q2, short __q3,
438
+ short __q4, short __q5, short __q6, short __q7)
439
+ ;
440
+
441
+ static __inline __m128i __attribute__((__always_inline__))
442
+ _mm_setr_epi8 (char __q00, char __q01, char __q02, char __q03,
443
+ char __q04, char __q05, char __q06, char __q07,
444
+ char __q08, char __q09, char __q10, char __q11,
445
+ char __q12, char __q13, char __q14, char __q15)
446
+ ;
447
+
448
+ /* Create a vector with element 0 as *P and the rest zero. */
449
+
450
+ static __inline __m128i __attribute__((__always_inline__))
451
+ _mm_load_si128 (__m128i const *__P)
452
+ ;
453
+
454
+ static __inline __m128i __attribute__((__always_inline__))
455
+ _mm_loadu_si128 (__m128i const *__P)
456
+ ;
457
+
458
+ /* APPLE LOCAL begin 4099020 */
459
+ static __inline __m128i __attribute__((__always_inline__))
460
+ _mm_loadl_epi64 (__m128i const *__P)
461
+ ;
462
+ /* APPLE LOCAL end 4099020 */
463
+
464
+ static __inline void __attribute__((__always_inline__))
465
+ _mm_store_si128 (__m128i *__P, __m128i __B)
466
+ ;
467
+
468
+ static __inline void __attribute__((__always_inline__))
469
+ _mm_storeu_si128 (__m128i *__P, __m128i __B)
470
+ ;
471
+
472
+ /* APPLE LOCAL begin 4099020 */
473
+ static __inline void __attribute__((__always_inline__))
474
+ _mm_storel_epi64 (__m128i *__P, __m128i __B)
475
+ ;
476
+ /* APPLE LOCAL end 4099020 */
477
+
478
+ static __inline __m64 __attribute__((__always_inline__))
479
+ _mm_movepi64_pi64 (__m128i __B)
480
+ ;
481
+
482
+ static __inline __m128i __attribute__((__always_inline__))
483
+ _mm_movpi64_epi64 (__m64 __A)
484
+ ;
485
+
486
+ /* APPLE LOCAL begin 4099020 */
487
+ static __inline __m128i __attribute__((__always_inline__))
488
+ _mm_move_epi64 (__m128i __A)
489
+ ;
490
+ /* APPLE LOCAL end 4099020 */
491
+
492
+ /* Create a vector of zeros. */
493
+ static __inline __m128i __attribute__((__always_inline__))
494
+ _mm_setzero_si128 (void)
495
+ ;
496
+
497
+ static __inline __m128d __attribute__((__always_inline__))
498
+ _mm_cvtepi32_pd (__m128i __A)
499
+ ;
500
+
501
+ static __inline __m128 __attribute__((__always_inline__))
502
+ _mm_cvtepi32_ps (__m128i __A)
503
+ ;
504
+
505
+ static __inline __m128i __attribute__((__always_inline__))
506
+ _mm_cvtpd_epi32 (__m128d __A)
507
+ ;
508
+
509
+ static __inline __m64 __attribute__((__always_inline__))
510
+ _mm_cvtpd_pi32 (__m128d __A)
511
+ ;
512
+
513
+ static __inline __m128 __attribute__((__always_inline__))
514
+ _mm_cvtpd_ps (__m128d __A)
515
+ ;
516
+
517
+ static __inline __m128i __attribute__((__always_inline__))
518
+ _mm_cvttpd_epi32 (__m128d __A)
519
+ ;
520
+
521
+ static __inline __m64 __attribute__((__always_inline__))
522
+ _mm_cvttpd_pi32 (__m128d __A)
523
+ ;
524
+
525
+ static __inline __m128d __attribute__((__always_inline__))
526
+ _mm_cvtpi32_pd (__m64 __A)
527
+ ;
528
+
529
+ static __inline __m128i __attribute__((__always_inline__))
530
+ _mm_cvtps_epi32 (__m128 __A)
531
+ ;
532
+
533
+ static __inline __m128i __attribute__((__always_inline__))
534
+ _mm_cvttps_epi32 (__m128 __A)
535
+ ;
536
+
537
+ static __inline __m128d __attribute__((__always_inline__))
538
+ _mm_cvtps_pd (__m128 __A)
539
+ ;
540
+
541
+ static __inline int __attribute__((__always_inline__))
542
+ _mm_cvtsd_si32 (__m128d __A)
543
+ ;
544
+
545
+ #ifdef __x86_64__
546
+ /* Intel intrinsic. */
547
+ static __inline long long __attribute__((__always_inline__))
548
+ _mm_cvtsd_si64 (__m128d __A)
549
+ ;
550
+
551
+ /* Microsoft intrinsic. */
552
+ static __inline long long __attribute__((__always_inline__))
553
+ _mm_cvtsd_si64x (__m128d __A)
554
+ ;
555
+ #endif
556
+
557
+ static __inline int __attribute__((__always_inline__))
558
+ _mm_cvttsd_si32 (__m128d __A)
559
+ ;
560
+
561
+ #ifdef __x86_64__
562
+ /* Intel intrinsic. */
563
+ static __inline long long __attribute__((__always_inline__))
564
+ _mm_cvttsd_si64 (__m128d __A)
565
+ ;
566
+
567
+ /* Microsoft intrinsic. */
568
+ static __inline long long __attribute__((__always_inline__))
569
+ _mm_cvttsd_si64x (__m128d __A)
570
+ ;
571
+ #endif
572
+
573
+ static __inline __m128 __attribute__((__always_inline__))
574
+ _mm_cvtsd_ss (__m128 __A, __m128d __B)
575
+ ;
576
+
577
+ static __inline __m128d __attribute__((__always_inline__))
578
+ _mm_cvtsi32_sd (__m128d __A, int __B)
579
+ ;
580
+
581
+ #ifdef __x86_64__
582
+ /* Intel intrinsic. */
583
+ static __inline __m128d __attribute__((__always_inline__))
584
+ _mm_cvtsi64_sd (__m128d __A, long long __B)
585
+ ;
586
+
587
+ /* Microsoft intrinsic. */
588
+ static __inline __m128d __attribute__((__always_inline__))
589
+ _mm_cvtsi64x_sd (__m128d __A, long long __B)
590
+ ;
591
+ #endif
592
+
593
+ static __inline __m128d __attribute__((__always_inline__))
594
+ _mm_cvtss_sd (__m128d __A, __m128 __B)
595
+ ;
596
+
597
+ /* APPLE LOCAL 5814283 */
598
+ #define _mm_shuffle_pd(__A, __B, __C) ((__m128d)__builtin_ia32_shufpd ((__v2df)(__A), (__v2df)(__B), (__C)))
599
+
600
+ static __inline __m128d __attribute__((__always_inline__))
601
+ _mm_unpackhi_pd (__m128d __A, __m128d __B)
602
+ ;
603
+
604
+ static __inline __m128d __attribute__((__always_inline__))
605
+ _mm_unpacklo_pd (__m128d __A, __m128d __B)
606
+ ;
607
+
608
+ static __inline __m128d __attribute__((__always_inline__))
609
+ _mm_loadh_pd (__m128d __A, double const *__B)
610
+ ;
611
+
612
+ static __inline __m128d __attribute__((__always_inline__))
613
+ _mm_loadl_pd (__m128d __A, double const *__B)
614
+ ;
615
+
616
+ static __inline int __attribute__((__always_inline__))
617
+ _mm_movemask_pd (__m128d __A)
618
+ ;
619
+
620
+ static __inline __m128i __attribute__((__always_inline__))
621
+ _mm_packs_epi16 (__m128i __A, __m128i __B)
622
+ ;
623
+
624
+ static __inline __m128i __attribute__((__always_inline__))
625
+ _mm_packs_epi32 (__m128i __A, __m128i __B)
626
+ ;
627
+
628
+ static __inline __m128i __attribute__((__always_inline__))
629
+ _mm_packus_epi16 (__m128i __A, __m128i __B)
630
+ ;
631
+
632
+ static __inline __m128i __attribute__((__always_inline__))
633
+ _mm_unpackhi_epi8 (__m128i __A, __m128i __B)
634
+ ;
635
+
636
+ static __inline __m128i __attribute__((__always_inline__))
637
+ _mm_unpackhi_epi16 (__m128i __A, __m128i __B)
638
+ ;
639
+
640
+ static __inline __m128i __attribute__((__always_inline__))
641
+ _mm_unpackhi_epi32 (__m128i __A, __m128i __B)
642
+ ;
643
+
644
+ static __inline __m128i __attribute__((__always_inline__))
645
+ _mm_unpackhi_epi64 (__m128i __A, __m128i __B)
646
+ ;
647
+
648
+ static __inline __m128i __attribute__((__always_inline__))
649
+ _mm_unpacklo_epi8 (__m128i __A, __m128i __B)
650
+ ;
651
+
652
+ static __inline __m128i __attribute__((__always_inline__))
653
+ _mm_unpacklo_epi16 (__m128i __A, __m128i __B)
654
+ ;
655
+
656
+ static __inline __m128i __attribute__((__always_inline__))
657
+ _mm_unpacklo_epi32 (__m128i __A, __m128i __B)
658
+ ;
659
+
660
+ static __inline __m128i __attribute__((__always_inline__))
661
+ _mm_unpacklo_epi64 (__m128i __A, __m128i __B)
662
+ ;
663
+
664
+ static __inline __m128i __attribute__((__always_inline__))
665
+ _mm_add_epi8 (__m128i __A, __m128i __B)
666
+ ;
667
+
668
+ static __inline __m128i __attribute__((__always_inline__))
669
+ _mm_add_epi16 (__m128i __A, __m128i __B)
670
+ ;
671
+
672
+ static __inline __m128i __attribute__((__always_inline__))
673
+ _mm_add_epi32 (__m128i __A, __m128i __B)
674
+ ;
675
+
676
+ static __inline __m128i __attribute__((__always_inline__))
677
+ _mm_add_epi64 (__m128i __A, __m128i __B)
678
+ ;
679
+
680
+ static __inline __m128i __attribute__((__always_inline__))
681
+ _mm_adds_epi8 (__m128i __A, __m128i __B)
682
+ ;
683
+
684
+ static __inline __m128i __attribute__((__always_inline__))
685
+ _mm_adds_epi16 (__m128i __A, __m128i __B)
686
+ ;
687
+
688
+ static __inline __m128i __attribute__((__always_inline__))
689
+ _mm_adds_epu8 (__m128i __A, __m128i __B)
690
+ ;
691
+
692
+ static __inline __m128i __attribute__((__always_inline__))
693
+ _mm_adds_epu16 (__m128i __A, __m128i __B)
694
+ ;
695
+
696
+ static __inline __m128i __attribute__((__always_inline__))
697
+ _mm_sub_epi8 (__m128i __A, __m128i __B)
698
+ ;
699
+
700
+ static __inline __m128i __attribute__((__always_inline__))
701
+ _mm_sub_epi16 (__m128i __A, __m128i __B)
702
+ ;
703
+
704
+ static __inline __m128i __attribute__((__always_inline__))
705
+ _mm_sub_epi32 (__m128i __A, __m128i __B)
706
+ ;
707
+
708
+ static __inline __m128i __attribute__((__always_inline__))
709
+ _mm_sub_epi64 (__m128i __A, __m128i __B)
710
+ ;
711
+
712
+ static __inline __m128i __attribute__((__always_inline__))
713
+ _mm_subs_epi8 (__m128i __A, __m128i __B)
714
+ ;
715
+
716
+ static __inline __m128i __attribute__((__always_inline__))
717
+ _mm_subs_epi16 (__m128i __A, __m128i __B)
718
+ ;
719
+
720
+ static __inline __m128i __attribute__((__always_inline__))
721
+ _mm_subs_epu8 (__m128i __A, __m128i __B)
722
+ ;
723
+
724
+ static __inline __m128i __attribute__((__always_inline__))
725
+ _mm_subs_epu16 (__m128i __A, __m128i __B)
726
+ ;
727
+
728
+ static __inline __m128i __attribute__((__always_inline__))
729
+ _mm_madd_epi16 (__m128i __A, __m128i __B)
730
+ ;
731
+
732
+ static __inline __m128i __attribute__((__always_inline__))
733
+ _mm_mulhi_epi16 (__m128i __A, __m128i __B)
734
+ ;
735
+
736
+ static __inline __m128i __attribute__((__always_inline__))
737
+ _mm_mullo_epi16 (__m128i __A, __m128i __B)
738
+ ;
739
+
740
+ static __inline __m64 __attribute__((__always_inline__))
741
+ _mm_mul_su32 (__m64 __A, __m64 __B)
742
+ ;
743
+
744
+ static __inline __m128i __attribute__((__always_inline__))
745
+ _mm_mul_epu32 (__m128i __A, __m128i __B)
746
+ ;
747
+
748
+ static __inline __m128i __attribute__((__always_inline__))
749
+ _mm_slli_epi16 (__m128i __A, int __B)
750
+ ;
751
+
752
+ static __inline __m128i __attribute__((__always_inline__))
753
+ _mm_slli_epi32 (__m128i __A, int __B)
754
+ ;
755
+
756
+ static __inline __m128i __attribute__((__always_inline__))
757
+ _mm_slli_epi64 (__m128i __A, int __B)
758
+ ;
759
+
760
+ static __inline __m128i __attribute__((__always_inline__))
761
+ _mm_srai_epi16 (__m128i __A, int __B)
762
+ ;
763
+
764
+ static __inline __m128i __attribute__((__always_inline__))
765
+ _mm_srai_epi32 (__m128i __A, int __B)
766
+ ;
767
+
768
+ #if 0
769
+ static __m128i __attribute__((__always_inline__))
770
+ _mm_srli_si128 (__m128i __A, const int __B)
771
+ ;
772
+
773
+ static __m128i __attribute__((__always_inline__))
774
+ _mm_srli_si128 (__m128i __A, const int __B)
775
+ ;
776
+ #else
777
+ #define _mm_srli_si128(__A, __B) \
778
+ ((__m128i)__builtin_ia32_psrldqi128 (__A, (__B) * 8))
779
+ #define _mm_slli_si128(__A, __B) \
780
+ ((__m128i)__builtin_ia32_pslldqi128 (__A, (__B) * 8))
781
+ #endif
782
+
783
+ static __inline __m128i __attribute__((__always_inline__))
784
+ _mm_srli_epi16 (__m128i __A, int __B)
785
+ ;
786
+
787
+ static __inline __m128i __attribute__((__always_inline__))
788
+ _mm_srli_epi32 (__m128i __A, int __B)
789
+ ;
790
+
791
+ static __inline __m128i __attribute__((__always_inline__))
792
+ _mm_srli_epi64 (__m128i __A, int __B)
793
+ ;
794
+
795
+ static __inline __m128i __attribute__((__always_inline__))
796
+ _mm_sll_epi16 (__m128i __A, __m128i __B)
797
+ ;
798
+
799
+ static __inline __m128i __attribute__((__always_inline__))
800
+ _mm_sll_epi32 (__m128i __A, __m128i __B)
801
+ ;
802
+
803
+ static __inline __m128i __attribute__((__always_inline__))
804
+ _mm_sll_epi64 (__m128i __A, __m128i __B)
805
+ ;
806
+
807
+ static __inline __m128i __attribute__((__always_inline__))
808
+ _mm_sra_epi16 (__m128i __A, __m128i __B)
809
+ ;
810
+
811
+ static __inline __m128i __attribute__((__always_inline__))
812
+ _mm_sra_epi32 (__m128i __A, __m128i __B)
813
+ ;
814
+
815
+ static __inline __m128i __attribute__((__always_inline__))
816
+ _mm_srl_epi16 (__m128i __A, __m128i __B)
817
+ ;
818
+
819
+ static __inline __m128i __attribute__((__always_inline__))
820
+ _mm_srl_epi32 (__m128i __A, __m128i __B)
821
+ ;
822
+
823
+ static __inline __m128i __attribute__((__always_inline__))
824
+ _mm_srl_epi64 (__m128i __A, __m128i __B)
825
+ ;
826
+
827
+ static __inline __m128i __attribute__((__always_inline__))
828
+ _mm_and_si128 (__m128i __A, __m128i __B)
829
+ ;
830
+
831
+ static __inline __m128i __attribute__((__always_inline__))
832
+ _mm_andnot_si128 (__m128i __A, __m128i __B)
833
+ ;
834
+
835
+ static __inline __m128i __attribute__((__always_inline__))
836
+ _mm_or_si128 (__m128i __A, __m128i __B)
837
+ ;
838
+
839
+ static __inline __m128i __attribute__((__always_inline__))
840
+ _mm_xor_si128 (__m128i __A, __m128i __B)
841
+ ;
842
+
843
+ static __inline __m128i __attribute__((__always_inline__))
844
+ _mm_cmpeq_epi8 (__m128i __A, __m128i __B)
845
+ ;
846
+
847
+ static __inline __m128i __attribute__((__always_inline__))
848
+ _mm_cmpeq_epi16 (__m128i __A, __m128i __B)
849
+ ;
850
+
851
+ static __inline __m128i __attribute__((__always_inline__))
852
+ _mm_cmpeq_epi32 (__m128i __A, __m128i __B)
853
+ ;
854
+
855
+ static __inline __m128i __attribute__((__always_inline__))
856
+ _mm_cmplt_epi8 (__m128i __A, __m128i __B)
857
+ ;
858
+
859
+ static __inline __m128i __attribute__((__always_inline__))
860
+ _mm_cmplt_epi16 (__m128i __A, __m128i __B)
861
+ ;
862
+
863
+ static __inline __m128i __attribute__((__always_inline__))
864
+ _mm_cmplt_epi32 (__m128i __A, __m128i __B)
865
+ ;
866
+
867
+ static __inline __m128i __attribute__((__always_inline__))
868
+ _mm_cmpgt_epi8 (__m128i __A, __m128i __B)
869
+ ;
870
+
871
+ static __inline __m128i __attribute__((__always_inline__))
872
+ _mm_cmpgt_epi16 (__m128i __A, __m128i __B)
873
+ ;
874
+
875
+ static __inline __m128i __attribute__((__always_inline__))
876
+ _mm_cmpgt_epi32 (__m128i __A, __m128i __B)
877
+ ;
878
+
879
+ #if 0
880
+ static __inline int __attribute__((__always_inline__))
881
+ _mm_extract_epi16 (__m128i const __A, int const __N)
882
+ ;
883
+
884
+ static __inline __m128i __attribute__((__always_inline__))
885
+ _mm_insert_epi16 (__m128i const __A, int const __D, int const __N)
886
+ ;
887
+ #else
888
+ #define _mm_extract_epi16(A, N) \
889
+ ((int) __builtin_ia32_vec_ext_v8hi ((__v8hi)(A), (N)))
890
+ #define _mm_insert_epi16(A, D, N) \
891
+ ((__m128i) __builtin_ia32_vec_set_v8hi ((__v8hi)(A), (D), (N)))
892
+ #endif
893
+
894
+ static __inline __m128i __attribute__((__always_inline__))
895
+ _mm_max_epi16 (__m128i __A, __m128i __B)
896
+ ;
897
+
898
+ static __inline __m128i __attribute__((__always_inline__))
899
+ _mm_max_epu8 (__m128i __A, __m128i __B)
900
+ ;
901
+
902
+ static __inline __m128i __attribute__((__always_inline__))
903
+ _mm_min_epi16 (__m128i __A, __m128i __B)
904
+ ;
905
+
906
+ static __inline __m128i __attribute__((__always_inline__))
907
+ _mm_min_epu8 (__m128i __A, __m128i __B)
908
+ ;
909
+
910
+ static __inline int __attribute__((__always_inline__))
911
+ _mm_movemask_epi8 (__m128i __A)
912
+ ;
913
+
914
+ static __inline __m128i __attribute__((__always_inline__))
915
+ _mm_mulhi_epu16 (__m128i __A, __m128i __B)
916
+ ;
917
+
918
+ /* APPLE LOCAL begin 5814283 */
919
+ #define _mm_shufflehi_epi16(__A, __B) ((__m128i)__builtin_ia32_pshufhw ((__v8hi)(__A), __B))
920
+ #define _mm_shufflelo_epi16(__A, __B) ((__m128i)__builtin_ia32_pshuflw ((__v8hi)(__A), __B))
921
+ #define _mm_shuffle_epi32(__A, __B) ((__m128i)__builtin_ia32_pshufd ((__v4si)(__A), __B))
922
+ /* APPLE LOCAL end 5814283 */
923
+
924
+ static __inline void __attribute__((__always_inline__))
925
+ _mm_maskmoveu_si128 (__m128i __A, __m128i __B, char *__C)
926
+ ;
927
+
928
+ static __inline __m128i __attribute__((__always_inline__))
929
+ _mm_avg_epu8 (__m128i __A, __m128i __B)
930
+ ;
931
+
932
+ static __inline __m128i __attribute__((__always_inline__))
933
+ _mm_avg_epu16 (__m128i __A, __m128i __B)
934
+ ;
935
+
936
+ static __inline __m128i __attribute__((__always_inline__))
937
+ _mm_sad_epu8 (__m128i __A, __m128i __B)
938
+ ;
939
+
940
+ static __inline void __attribute__((__always_inline__))
941
+ _mm_stream_si32 (int *__A, int __B)
942
+ ;
943
+
944
+ static __inline void __attribute__((__always_inline__))
945
+ _mm_stream_si128 (__m128i *__A, __m128i __B)
946
+ ;
947
+
948
+ static __inline void __attribute__((__always_inline__))
949
+ _mm_stream_pd (double *__A, __m128d __B)
950
+ ;
951
+
952
+ static __inline void __attribute__((__always_inline__))
953
+ _mm_clflush (void const *__A)
954
+ ;
955
+
956
+ static __inline void __attribute__((__always_inline__))
957
+ _mm_lfence (void)
958
+ ;
959
+
960
+ static __inline void __attribute__((__always_inline__))
961
+ _mm_mfence (void)
962
+ ;
963
+
964
+ static __inline __m128i __attribute__((__always_inline__))
965
+ _mm_cvtsi32_si128 (int __A)
966
+ ;
967
+
968
+ #ifdef __x86_64__
969
+ /* Intel intrinsic. */
970
+ static __inline __m128i __attribute__((__always_inline__))
971
+ _mm_cvtsi64_si128 (long long __A)
972
+ ;
973
+
974
+ /* Microsoft intrinsic. */
975
+ static __inline __m128i __attribute__((__always_inline__))
976
+ _mm_cvtsi64x_si128 (long long __A)
977
+ ;
978
+ #endif
979
+
980
+ /* Casts between various SP, DP, INT vector types. Note that these do no
981
+ conversion of values, they just change the type. */
982
+ static __inline __m128 __attribute__((__always_inline__))
983
+ _mm_castpd_ps(__m128d __A)
984
+ ;
985
+
986
+ static __inline __m128i __attribute__((__always_inline__))
987
+ _mm_castpd_si128(__m128d __A)
988
+ ;
989
+
990
+ static __inline __m128d __attribute__((__always_inline__))
991
+ _mm_castps_pd(__m128 __A)
992
+ ;
993
+
994
+ static __inline __m128i __attribute__((__always_inline__))
995
+ _mm_castps_si128(__m128 __A)
996
+ ;
997
+
998
+ static __inline __m128 __attribute__((__always_inline__))
999
+ _mm_castsi128_ps(__m128i __A)
1000
+ ;
1001
+
1002
+ static __inline __m128d __attribute__((__always_inline__))
1003
+ _mm_castsi128_pd(__m128i __A)
1004
+ ;
1005
+ /* APPLE LOCAL end radar 4152603 */
1006
+
1007
+ /* APPLE LOCAL begin 4220129, 4286110 */
1008
+ /* Set all of the elements of the vector to A. */
1009
+
1010
+ static __inline __m128i __attribute__((__always_inline__))
1011
+ _mm_set1_epi64x (long long __A)
1012
+ ;
1013
+
1014
+ static __inline __m128i __attribute__((__always_inline__))
1015
+ _mm_set1_epi64 (__m64 __A)
1016
+ ;
1017
+
1018
+ static __inline __m128i __attribute__((__always_inline__))
1019
+ _mm_set1_epi32 (int __A)
1020
+ ;
1021
+
1022
+ static __inline __m128i __attribute__((__always_inline__))
1023
+ _mm_set1_epi16 (short __A)
1024
+ ;
1025
+
1026
+ static __inline __m128i __attribute__((__always_inline__))
1027
+ _mm_set1_epi8 (char __A)
1028
+ ;
1029
+ /* APPLE LOCAL end 4220129, 4286110 */
1030
+
1031
+ /* APPLE LOCAL begin nodebug inline 4152603 */
1032
+ #undef __always_inline__
1033
+ /* APPLE LOCAL end nodebug inline 4152603 */
1034
+
1035
+ #endif /* __SSE2__ */
1036
+
1037
+ #endif /* _EMMINTRIN_H_INCLUDED */